WO2012161318A1 - Image encoding device, image decoding device, image encoding method, image decoding method and program - Google Patents

Image encoding device, image decoding device, image encoding method, image decoding method and program Download PDF

Info

Publication number
WO2012161318A1
WO2012161318A1 PCT/JP2012/063503 JP2012063503W WO2012161318A1 WO 2012161318 A1 WO2012161318 A1 WO 2012161318A1 JP 2012063503 W JP2012063503 W JP 2012063503W WO 2012161318 A1 WO2012161318 A1 WO 2012161318A1
Authority
WO
WIPO (PCT)
Prior art keywords
prediction
unit
image
pixel
processing target
Prior art date
Application number
PCT/JP2012/063503
Other languages
French (fr)
Japanese (ja)
Inventor
大津 誠
Original Assignee
シャープ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by シャープ株式会社 filed Critical シャープ株式会社
Publication of WO2012161318A1 publication Critical patent/WO2012161318A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • the present invention relates to an image encoding device, an image decoding device, an image encoding method, an image decoding method, and a program.
  • H.264 H.264 Non-patent Reference 1
  • macroblocks a plurality of blocks
  • prediction methods sequentially selecting a prediction method having a high coding efficiency for each macroblock.
  • an intra prediction encoding method (intra prediction encoding method) that predicts and encodes an encoding target block using pixel information of a block that has already been encoded in the screen.
  • an inter-picture predictive coding method (inter-predictive coding method) for predicting and coding an encoding target block with reference to an image different from an image to be processed.
  • the intra prediction encoding method is a macroblock (16 pixels ⁇ 16 pixels) unit, or 4 pixels ⁇ 4 pixels and 8 pixels ⁇ 8 pixels (8 pixels ⁇ 8 pixels are further divided by H.264 FRExt).
  • Information for identifying the prediction mode and the code amount when the difference image (residual component) between the prediction image generated according to the prescribed prediction mode and the original image to be encoded is encoded in units)
  • An optimal prediction method is selected based on the amount of code necessary for encoding (Non-patent Document 1).
  • FIGS. 6 and 6 Four types of prediction modes (FIGS. 6 and 6 will be described later) can be applied to a block of 16 pixels ⁇ 16 pixels, and prediction using one type of DC component (average value prediction) and three types of prediction are possible. There are predictions using angles (vertical prediction, horizontal prediction, planar prediction).
  • FIGS. 4 and 4 Nine kinds of prediction modes (FIGS. 4 and 4 will be described later) can be applied to a block of 4 pixels ⁇ 4 pixels or 8 pixels ⁇ 8 pixels, and prediction (average) by one type of DC component is applicable.
  • Value prediction) and prediction using eight types of prediction angles prediction of non-uniform angles of 45 ° to 206.57 °).
  • prediction is performed using the prediction modes of the upper part and the left part of the processing target block, and when the prediction matches with the prediction mode A 1-bit flag is prepared, and a determination is made by setting a flag that matches. If the prediction does not match, the flag is not set, and encoding is performed by adding information for 3 bits for determining the remaining eight types of prediction modes excluding the prediction mode that does not match. If the prediction is correct, only one bit of information is required to encode the prediction mode, but if the prediction is not correct, information of 4 bits is required.
  • a prediction mode for example, an index value indicating a mode
  • the invention described in Japanese Patent Application Laid-Open No. H10-228707 is H.264 so that prediction can be performed at an arbitrary prediction angle for the purpose of improving the encoding efficiency of the intra prediction encoding scheme.
  • the number of prediction modes is increased compared to the H.264 method.
  • a technique is disclosed in which a theoretical reference pixel position is obtained from a predicted angle and the position of a pixel to be processed, and a pixel value corresponding to the pixel position is generated by interpolating surrounding reference pixels.
  • the invention described in Patent Document 2 discloses a method for accurately predicting a prediction mode based on the angle of surrounding prediction modes.
  • the present invention has been made in view of such circumstances, and an object of the present invention is to provide an encoding device, a decoding device, an encoding method, a decoding method, and a program that improve the accuracy of a predicted image while suppressing an increase in the amount of code. Is to provide.
  • the present invention has been made to solve the above-described problems, and according to one aspect of the present invention, when an input image is encoded, pixel values of peripheral pixels around the processing target pixel are used.
  • An image encoding apparatus that performs intra prediction to predict a pixel value of the processing target pixel, and includes information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels.
  • An image coding apparatus comprising an intra-screen prediction unit that determines a predicted value for each pixel based on the image.
  • the other aspect of this invention is the above-mentioned image coding apparatus, Comprising:
  • the said prediction part in a screen is for every pixel based on the distance between the pixels of the said process target pixel and the said surrounding pixel. A predicted value is determined.
  • the intra-screen prediction unit uses information indicating a boundary of the subject and a distance to the subject of the input image.
  • a subject boundary detection unit for detecting, and when the boundary is not detected between the processing target pixel and a pixel adjacent in a predetermined direction, the processing target is detected using the pixel adjacent in the predetermined direction;
  • a prediction value of a pixel is predicted.
  • a pixel value of a processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel.
  • An image decoding device that performs prediction, and when performing the intra prediction, information for each pixel based on information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels.
  • An image decoding apparatus comprising an in-screen prediction unit that determines a predicted value.
  • the other aspect of this invention is the above-mentioned image decoding apparatus, Comprising:
  • the said prediction part in a screen is the prediction for every pixel based on the distance between the pixels of the said process target pixel and the said surrounding pixel. It is characterized by determining a value.
  • the in-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the input image.
  • a subject boundary detection unit that detects the boundary between the processing target pixel and a pixel adjacent in a predetermined direction, and uses the pixel adjacent in the predetermined direction to detect the processing target pixel. The prediction value is predicted.
  • a pixel value of the processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel.
  • An image encoding method for performing prediction wherein when performing intra prediction, information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels. This is an image coding method characterized by having a process of determining a predicted value for each pixel based on the above.
  • intra-frame prediction when decoding an input image, intra-frame prediction that predicts a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel.
  • the intra prediction unit is based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. And a process for determining a predicted value for each pixel.
  • a pixel value of the processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel.
  • the image coding apparatus that performs prediction performs the intra prediction, prediction for each pixel is performed based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. It is a program for functioning as an in-screen prediction unit for determining a value.
  • intra-frame prediction when decoding an input image, intra-frame prediction that predicts the pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel.
  • the prediction value for each pixel is calculated based on the information indicating the distance to the subject in the processing target pixel and the information indicating the distance to the subject in the peripheral pixels. It is a program for functioning as an in-screen prediction unit to be determined.
  • FIG. 1 is a schematic block diagram showing a configuration of a moving image transmission system according to an embodiment of the present invention.
  • the moving image transmission system 10 in this embodiment includes an image encoding device 100, a communication network 500, an image decoding device 800, and a display device 600.
  • the image encoding device 100 encodes the image and the depth map from the image signal R of the image to be encoded and the depth map signal D of the depth map corresponding to the image, and uses the encoded data.
  • a certain encoded data E is generated and output.
  • the communication network 500 transmits the encoded data E output from the image encoding device 100 to the image decoding device 800.
  • the image decoding apparatus 800 decodes the transmitted encoded data E, and generates an image signal R ′ of the decoded image.
  • the display device 600 includes an image display device such as a liquid crystal display or a plasma display, and displays an image indicated by the image signal R ′ generated by the image decoding device 800.
  • the image encoding device 100 is provided in a television broadcasting station, for example, and encodes a broadcast program.
  • the communication network 500 is a communication network that transmits using broadcast waves
  • the image decoding device 800 and the display device 600 are provided in a television receiver.
  • the Internet or a mobile phone network may be used as the communication network 500.
  • the image encoding apparatus 100 is provided in a content holder that edits contents stored and sold on a DVD (Digital Versatile Disc) or a BD (Blu-ray Disc), and encodes these contents.
  • the encoded image E is stored in a DVD, a BD, or the like, and is delivered by a delivery network instead of the communication network 500.
  • the image decoding device 800 is provided in a DVD player, a BD player, or the like.
  • FIG. 2 is a schematic block diagram illustrating a configuration of the image encoding device 100 according to the present embodiment.
  • the image coding apparatus 100 includes an image input unit 101, a subtraction unit 102, an orthogonal transformation unit 103, a quantization unit 104, an entropy coding unit 105, an inverse quantization unit 106, an inverse orthogonal transformation unit 107, an addition unit 108, and a prediction method.
  • Control unit 109, selection unit 110, deblocking filter unit 111, frame memory unit 112, motion compensation unit 113, motion vector detection unit 114, depth information use intra prediction unit 115, depth map encoding unit 116, depth map decoding unit 117 and a depth input unit 118.
  • the de-blocking filter unit 111, the frame memory 112, the motion compensation unit 113, and the motion vector detection 114 constitute an inter prediction unit 120.
  • the depth information use intra prediction unit 115 and the depth map decoding unit 117 constitute an intra prediction unit 121.
  • the image input unit 101 for example, outputs an image signal R (input image signal) indicating an image to be encoded (input image) as an example every 5 frames (the types of 5 frames will be described later). 100 from outside.
  • the image input unit 101 divides the input image frame represented by the acquired input image signal into blocks having a predetermined size (for example, 16 pixels in the vertical direction ⁇ 16 pixels in the horizontal direction).
  • the image input unit 101 outputs an image block signal B representing each of the divided blocks to the subtraction unit 102, the motion vector detection unit 114, and the depth information use intra prediction unit 115.
  • the image input unit 101 repeats this process for each image frame until the output for all the blocks in the image frame is completed and the acquired image is completed while sequentially changing the block position.
  • the input image to the image encoding device 100 includes at least a reference image (base view).
  • the reference image is an image of one predetermined viewpoint included in a multi-view (multi-view) moving image for stereoscopic display, and is an image serving as a basis for calculating a depth map.
  • the depth map is distance information representing the depth or distance from the photographing device of the subject represented in the reference image, and includes a quantized value given for each pixel of the reference image. Each of the quantized values is called a depth value, for example, a value quantized with 8 bits.
  • the image signal R for every five frames input to the image input unit 101 is, for example, an image signal of an I picture (I0), a B picture (B3), a B picture (B2), a B picture (B4), and a P picture (P1). including.
  • the image signal R input to the image encoding device 100 is input in this order (hereinafter referred to as input order), for example.
  • input order hereinafter referred to as input order
  • the leading I, etc. indicates the type of image
  • 0, etc. indicates the order of encoding (hereinafter, encoding order) (therefore, the input order and the encoding order are different). .
  • An I picture is an intra-frame picture (Intra Frame Picture), which can be decoded using only a code obtained by encoding the picture.
  • the P picture is an inter-frame forward prediction image (Predictive Picture), which is an image that can be decoded using a code obtained by encoding the image and a code obtained by encoding an image signal of a past frame.
  • Predictive Picture inter-frame forward prediction image
  • a B picture is a bi-predictive coded picture (Bi-directional Predictive Picture), which is obtained by decoding the input picture and using a code obtained by coding a plurality of past or future frame image signals. It is an image that can be decoded.
  • the subtraction unit 102 subtracts the prediction image block signal output from the selection unit 110 from the image block signal output from the image input unit 101 to generate a difference image block signal.
  • the subtraction unit 102 outputs the generated difference image block signal to the orthogonal transformation unit 103.
  • the orthogonal transform unit 103 performs orthogonal transform on the difference image block signal output from the subtraction unit 102 to generate signals indicating the strengths of various frequency characteristics.
  • the orthogonal transform unit 103 orthogonally transforms the difference image block signal
  • the difference image block signal is subjected to, for example, DCT transform (discrete cosine transform) and a frequency domain signal (for example, DCT transform). , DCT coefficients).
  • DCT transform discrete cosine transform
  • a frequency domain signal for example, DCT transform
  • DCT coefficients DCT coefficients
  • other methods for example, FFT (Fast Fourier Transform)
  • the orthogonal transform unit 103 outputs the coefficient value included in the generated frequency domain signal to the quantization unit 104.
  • the quantization unit 104 quantizes the coefficient value indicating each frequency characteristic intensity output from the orthogonal transform unit 103, and generates the generated quantized signal ED (difference image block code) with the entropy encoding unit 105 and the inverse quantization unit 106. Output to.
  • ED difference image block code
  • the inverse quantization unit 106 performs inverse quantization on the quantized signal ED output from the quantization unit 104 to generate a decoded frequency domain signal, and outputs the decoded frequency domain signal to the inverse orthogonal transform unit 107.
  • the inverse orthogonal transform unit 107 performs, for example, inverse DCT transform on the input decoded frequency domain signal to generate a decoded differential image block signal that is a spatial domain signal.
  • the inverse orthogonal transform unit 107 is not limited to the inverse DCT transform, and other methods (eg, IFFT (Inverse Fast Fourier Transform)) are used. It may be used.
  • the inverse orthogonal transform unit 107 outputs the generated decoded difference image block signal to the addition unit 108.
  • the addition unit 108 acquires the predicted image block signal from the selection unit 110 and acquires the decoded difference image block signal from the inverse orthogonal transform unit 107.
  • the adder 108 adds the decoded differential image block signal to the predicted image block signal, and generates a reference image block signal RB obtained by encoding / decoding the input image (internal decoding).
  • the reference image block signal RB is output to the inter prediction unit 120 and the intra prediction unit 121.
  • the inter prediction unit 120 acquires the reference image block signal RB from the addition unit 108 and acquires the image block signal from the image input unit 101.
  • the inter prediction unit 120 performs inter prediction using these signals, and generates an inter prediction image block signal.
  • the inter prediction unit 120 outputs the generated inter prediction image block signal to the prediction method control unit 109 and the selection unit 110.
  • the inter prediction unit 120 outputs the generated inter prediction coding information IPE to the prediction scheme control unit 109.
  • the inter prediction unit 120 will be described later.
  • the intra prediction unit 121 acquires the reference image block signal RB from the addition unit 108, acquires the image block signal from the image input unit 101, and acquires depth map encoded data from the depth map encoding unit 116.
  • the intra prediction unit 121 performs intra prediction using these signals and data, and generates an intra predicted image block signal.
  • the intra prediction unit 121 outputs the generated intra prediction image block signal to the prediction scheme control unit 109 and the selection unit 110.
  • the intra prediction unit 121 outputs the generated intra prediction encoding information TPE to the prediction scheme control unit 109.
  • the intra prediction unit 121 will be described later.
  • the depth input unit 118 acquires the depth map signal D of the depth map corresponding to the input image input to the image input unit 101 from the outside of the image encoding device 100.
  • the depth input unit 118 divides (depth block signal) the acquired depth map so that the input image block divided by the image input unit 101 has the same position and the same block size, and the depth map encoding unit.
  • the depth map encoding unit 116 encodes the depth block signal output from the depth input unit 118 using, for example, variable length encoding (entropy encoding), and converts the depth map encoded data E ⁇ b> 2 whose data amount is further compressed. Generate.
  • the depth map encoding unit 116 outputs the generated depth map encoded data E2 to the intra prediction unit 121 and the outside of the image encoding device 100 (for example, the image decoding device 800 via the communication network 500).
  • the inter prediction unit 120 includes a deblocking filter unit 111, a frame memory 112, a motion compensation unit 113, and a motion vector detection unit 114.
  • the deblocking filter unit 111 acquires the reference image block signal RB from the adder unit 108, and reduces the block distortion generated when the image is encoded, for example, a known encoding method (for example, H.264 Reference Software). FIR (Finite Impulse Response) filter processing used in JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, 2008).
  • the deblocking filter unit 111 outputs the processing result (correction block signal) to the frame memory 112.
  • the frame memory 112 holds the correction block signal output from the deblocking filter unit 111 as a part of the image of the frame number together with information for identifying the frame number.
  • the motion vector detection unit 114 searches for a block similar to the image block signal input from the image input unit 101 from the image stored in the frame memory 112 (block matching), and vector information (motion vector) representing the searched block. Is generated.
  • the motion vector detection unit 114 calculates an index value between the divided blocks for each area, and searches for an area where the calculated index value is minimum.
  • the motion vector detection unit 114 has two areas: a block in the reference image area having the smallest index value, a block in the reference image area having the next smallest index value, and the two areas. Find out.
  • the motion vector detection unit 114 uses, for example, an absolute value sum (SAD: Sum of Absolute Difference) of a difference between a luminance value of a pixel included in a divided block and a luminance value in a certain region of the reference image.
  • SAD an absolute value sum
  • SAD between a block (for example, a size of N ⁇ N pixels) divided from the input image signal and the block of the reference image signal is expressed by the following equation (1).
  • Iin (i0 + i, j0 + j) represents the luminance value at the coordinates (i0 + i, j0 + j) of the input image
  • (i0, j0) represents the pixel coordinates of the upper left corner of the divided block
  • Iref (i0 + i + p, j0 + j + q) is a luminance value at the coordinates (i0 + i + p, j0 + j + q) of the reference image
  • (p, q) is a shift amount (motion vector) based on the coordinates of the upper left corner of the divided block.
  • the motion vector detection unit 114 calculates SAD (p, q) for each (p, q) in block matching, and finds (p, q) that minimizes SAD (p, q).
  • (P, q) represents a vector (motion vector) from the divided block in the input image to the position of the reference region in the reference image.
  • the motion compensation unit 113 acquires a motion vector from the motion vector detection unit 114, and outputs the corresponding reference block to the prediction scheme control unit 109 and the selection unit 110 as an inter prediction image block signal.
  • the motion compensator 113 outputs the corresponding image block when the motion vector output from the motion vector detector 114 is one, and applies when the motion vector output from the motion vector detector 114 is two. Two image blocks are averaged and output.
  • the motion compensation unit 113 outputs information necessary for prediction (hereinafter, inter prediction coding information IPE), for example, a motion vector, to the prediction scheme control unit 109.
  • inter prediction coding information IPE information necessary for prediction
  • the intra prediction unit 121 includes a depth map decoding unit 117 and a depth information use intra prediction unit 115.
  • the depth map decoding unit 117 decodes the depth block signal having a larger amount of information by using, for example, variable length decoding, the depth map encoded data output from the depth map encoding unit 116.
  • the depth map decoding unit 117 outputs the decoded depth map D ′ (depth block decoded signal) to the depth information use intra prediction unit 115.
  • FIG. 3 is a schematic block diagram illustrating a configuration of the depth information use intra prediction unit 115 according to the present embodiment.
  • the processing of the depth information use intra prediction unit 115 will be described with reference to FIG.
  • the depth information use intra prediction unit 115 includes a first prediction mode execution unit 200-1 to an nth prediction mode execution unit 200-n (n is a natural number of 1 or more, for example, 6), and a depth use prediction mode execution.
  • Unit 201 and a prediction mode selection unit 202 is a prediction mode selection unit 202.
  • the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n perform the first to nth prediction mode execution units 200-n according to the processing in each prediction mode (prediction image block generation method) from the reference image block signal RB output from the addition unit 108. An n-th predicted image block signal is generated.
  • the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n output the generated first to n-th prediction image block signals to the prediction mode selection unit 202.
  • Each of the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n is, for example, a conventional intra-screen prediction mode (for example, H.264 Reference Software JM ver. 13.2 Encoder, http: // iphome. hhi.de/suehring/tml/, 2008) is used to perform in-screen prediction (intra prediction).
  • H.264 there are nine types of intra prediction applied to a 4 ⁇ 4 pixel sub-block obtained by further dividing a macroblock, and four types of intra prediction applied to a macroblock unit (note that 8 ⁇ 8 pixels).
  • In-screen prediction using sub-blocks is formulated in H.264 FRExt, and the same intra-screen prediction method as 4 ⁇ 4 pixels is applied.
  • the first prediction mode execution unit 200-1 performs intra prediction (intra-screen prediction) using, for example, 4 ⁇ 4 sub-blocks.
  • the second prediction mode execution unit 200-2 performs intra prediction using, for example, 8 ⁇ 8 sub-blocks.
  • the third prediction mode execution unit 200-3 to the sixth prediction mode execution unit 200-6 perform four types of prediction methods, for example, in units of 16 ⁇ 16 macroblocks.
  • the first prediction mode execution generation unit 200-1 further divides the reference image block signal output from the addition unit 108 into 4 ⁇ 4 sub-block sizes, and performs prediction in units of 4 ⁇ 4 pixels in the order shown in FIG. Execute the method. That is, a 16 ⁇ 16 pixel block is divided into four 8 ⁇ 8 pixel blocks, which are processed in the order of upper left, upper right, lower left, and lower right. For each of these 8 ⁇ 8 pixel blocks, each is divided into four 4 ⁇ 4 pixel sub-blocks, and within each of these 8 ⁇ 8 pixel blocks, upper left, upper right, lower left, right Intra prediction is performed in the following order.
  • the first prediction mode execution unit 200-1 includes a 4 ⁇ 4 pixel predicted image block generated by each of nine types of prediction methods and a corresponding sub-block of the image block signal B output from the image input unit 101. An index indicating the degree of correlation is calculated, and a prediction method is selected for each sub-block based on the index.
  • the first prediction mode execution unit 200-1 calculates, for example, the absolute value sum (SAD) of the luminance value differences as the index, and sets the prediction method with the smallest SAD value to the corresponding 4 ⁇ 4 pixel sub A block prediction method is selected, and a first predicted image block signal at a corresponding position is generated. Also, the prediction method is retained. The first prediction mode execution unit 200-1 repeats the above process until a prediction method for 16 ⁇ 16 pixels and a first predicted image block signal are generated.
  • SAD absolute value sum
  • the second prediction mode execution unit 200-2 further divides the reference image block signal RB output from the addition unit 108 into four 8 ⁇ 8 pixel sub-blocks, and uses the prediction used in the first prediction mode execution unit 200-1. Nine kinds of prediction methods similar to those in the mode 0 to the prediction mode 8 are applied to each of the 8 ⁇ 8 pixel sub-blocks to generate a prediction image. At the same time, the prediction method is retained.
  • the second prediction mode execution unit 200-2 repeats the above processing, sequentially determines the prediction method in units of 8 ⁇ 8 pixel sub-blocks, and predicts all 16 ⁇ 16 pixel block prediction methods and predicted images based on the prediction method. Generate block signals.
  • the third prediction mode execution unit 200-3 to the sixth prediction mode execution unit 200-6 are 16 ⁇ 16 pixel intra predictions (intra-screen predictions), and use the reference image block signal output from the addition unit 108. Prediction image block signals corresponding to prediction modes 0 to 3 in FIG. 6 are generated.
  • the depth use prediction mode execution unit 201 obtains a reference image block signal from the addition unit 108 and a depth block decoded signal from the depth map decoding unit 117, and uses the depth map to suppress prediction across a subject boundary. I do. Details of the depth use prediction mode execution unit 201 will be described later.
  • the depth use prediction mode execution unit 201 outputs the prediction image block signal and the prediction method to the prediction mode selection unit 202.
  • the prediction mode selection unit 202 acquires the prediction image block signal generated by the first prediction mode execution unit 200-1 to the nth prediction mode execution unit 200-n and the depth use prediction mode execution unit 201 and information necessary for prediction. To do.
  • the information necessary for prediction is applied to each sub-block of the first prediction mode execution unit 200-1 and the second prediction mode execution unit 200-2 that perform processing by further dividing 16 ⁇ 16 pixels into sub-blocks, for example.
  • Information indicating the predicted mode, and information indicating the prediction mode indicating the direction of prediction of the depth use prediction mode execution unit 201 are examples of the prediction mode.
  • the prediction mode selection unit 202 selects one prediction image block signal having the smallest index value from the obtained prediction image block signals (including the prediction image block signal output by the depth use prediction mode execution unit 201). .
  • the prediction mode selection unit 202 for example, as shown by the following equation, the luminance value Iin (i0 + i, j0 + j) of the corresponding image block included in the input image input from the image input unit 101 and the candidate prediction image SAD with the luminance value Ip, m (i0 + i, j0 + j) of the block is used.
  • m is an index indicating which prediction mode of which prediction mode execution unit is. Accordingly, Ip, m (x, y) is a luminance value at the coordinates x, y of the predicted image in the prediction mode m. Further, i0 and j0 are the coordinates of the upper left vertex of the block, and N is the size of the block (the number of pixels on one side).
  • the effectiveness of processing for each prediction mode such as the correlation between the image block included in the input image as an index value and the candidate predicted image block, the similarity, or the amount of information after encoding. Any variable that represents can be used.
  • the prediction mode selection unit 202 generates prediction mode information including an index representing the prediction mode. Alternatively, the prediction mode selection unit 202 selects a prediction mode in which information necessary for prediction exists (specifically, the first prediction mode execution unit 200-1, the second prediction mode execution unit 200-2, and the depth). In the prediction mode of the use prediction mode execution unit 201, the index and information necessary for this prediction are collected to generate prediction mode information.
  • the prediction mode selection unit 202 sends the selected prediction image block signal (hereinafter referred to as an intra prediction image block signal) to the selection unit 110 and the prediction scheme control unit 109, and the prediction mode information (hereinafter referred to as intra prediction coding information TPE). Output to the prediction method control unit 109.
  • the prediction scheme control unit 109 receives the picture type of the input image, the inter prediction image block signal input from the inter prediction unit 120, its inter prediction encoding information IPE, and the intra prediction unit 121.
  • a prediction scheme is determined based on the intra-predicted image block signal and the intra-encoded information, and information on the prediction scheme is output to the selection unit 110 and the entropy encoding unit 105.
  • the prediction method control unit 109 monitors the picture type of the input image, and selects an intra prediction method when the input image is an I picture.
  • the prediction scheme control unit 109 uses, for example, a conventional technique (from the residual between the number of bits generated by the encoding performed by the entropy encoding unit 105 and the original image of the subtraction unit 102 (for example, the Lagrange cost is calculated using H.264 Reference Software JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, ⁇ 2008), and either inter prediction method or intra prediction method is selected. select.
  • the prediction method control unit 109 adds information that can specify the prediction method to the coding information corresponding to the selected prediction method, among the inter prediction coding information IPE or the intra prediction coding information TPE, as prediction coding information.
  • the data is output to the entropy encoding unit 105.
  • the selection unit 110 selects the inter prediction image block signal input from the inter prediction unit 120 or the intra prediction image block signal input from the intra prediction unit 121 according to the prediction method information input from the prediction method control unit 109. Then, the predicted image block signal is output to the subtracting unit 102 and the adding unit 108.
  • the selection unit 110 selects and outputs the inter prediction image block signal input from the inter prediction unit 120, and the prediction method control unit 109.
  • the input prediction method is intra prediction
  • the intra prediction image block signal input from the intra prediction unit 121 is selected and output.
  • the entropy encoding unit 105 packs the differential image code input from the quantization unit 104 and the prediction encoding information input from the prediction scheme control unit 109, for example, variable length encoding (entropy).
  • the encoded data E1 is generated by using the encoding) to compress the amount of information.
  • the entropy encoding 105 outputs the generated encoded data E1 to the outside of the image encoding device 100 (for example, the image decoding device 800 via the communication network 500).
  • In-screen prediction is performed by predicting pixels of a processing target block using surrounding pixels as described above.
  • a predicted image block signal is created by sequentially copying neighboring pixels that have been processed in the prediction direction.
  • the pixel of the processing target block can be accurately predicted by this intra prediction, and the difference (residual) between the pixel of the processing target block and the pixel of the prediction block As a result, the code amount can be reduced (or the error during decoding can be reduced).
  • the prediction directions of the intra prediction using the depth performed by the depth use prediction mode execution unit 201 in the present embodiment are the prediction in the vertical direction (prediction mode 0) and the prediction in the horizontal direction (prediction mode 1) shown in FIG. is there.
  • prediction mode 0 the prediction in the vertical direction
  • prediction mode 1 the prediction in the horizontal direction
  • the processing described below can be applied (however, except for the prediction mode 2 in FIG. 6), other prediction directions can be used.
  • the processing described below can be applied to the prediction method in units of sub-blocks in FIG. 4 (except for prediction mode 2). That is, as in the present embodiment, a new prediction mode may be added while leaving the conventional prediction mode, or the prediction method performed by the depth use prediction mode execution unit 201 may be replaced with the conventional method. Thus, the number of modes may not be increased.
  • the following describes an example in which a depth usage prediction mode is newly added.
  • FIG. 7 and 8 are diagrams for explaining the processing concept of the depth use prediction mode execution unit 201.
  • the graphic indicated by a circle indicates a pixel for which processing has been completed, and can be referred to when a predicted pixel block is generated.
  • a graphic indicated by a square indicates a pixel to be processed, and is a target that is predicted using pixels that can be referred to in the vicinity.
  • the arrow indicates the direction of prediction, and pixels that can be referred to are sequentially predicted (specifically, simply copied) in the direction of the arrow. That is, in the prediction mode of FIG. 7, the pixel value is copied in the vertical direction, and in the prediction mode of FIG. 8, the pixel value is copied in the horizontal direction. 7 and 8, a thick broken line indicates the boundary of the subject.
  • FIG. 9 is a schematic block diagram illustrating a configuration of the depth use prediction mode execution unit 201 according to the present embodiment.
  • the depth use prediction mode execution unit 201 includes a boundary control prediction image generation unit 300, a boundary prediction control unit 301, and a subject boundary detection unit 302.
  • the subject boundary detection unit 302 acquires a depth block signal representing a depth value of a pixel corresponding to the image block signal B to be processed from the depth map decoding unit 117, and detects a depth edge.
  • Depth edge detection is performed by thresholding the difference between adjacent pixels in the depth map. Whether or not the depth edge exists in the horizontal direction is determined by whether or not the absolute value of the difference between pixels adjacent in the vertical direction is larger than the threshold value TV, as shown in Expression (3). Similarly, whether or not the depth edge exists in the vertical direction is determined by whether or not the absolute value of the difference between pixels adjacent in the horizontal direction is larger than the threshold value TH.
  • D (i, j) represents a depth map value at the pixel position (i, j).
  • TV and TH are threshold values used when determining whether or not edges exist in the horizontal direction and the vertical direction, respectively.
  • the threshold is 10, for example.
  • FIGS. 7 and 8 As an example of the depth edge detection result obtained by the above method, a case where a depth edge is detected as shown by a thick dotted line in FIGS. 7 and 8 will be described.
  • the position of the depth edge is the same.
  • the prediction accuracy is remarkably lowered at the pixel straddling this edge and the subsequent pixels in the prediction direction.
  • the boundary prediction control unit 301 controls the prediction performed by the boundary control predicted image generation unit 300 using the boundary information (depth edge) of the subject in the horizontal direction and the vertical direction input from the subject boundary detection unit 302. Specifically, when there is a depth edge perpendicular to the prediction direction, the boundary prediction control unit 301 performs control to suppress copying from pixels adjacent to the prediction direction.
  • the control for suppressing copying of pixels in the prediction direction is, for example, controlling the processing in the boundary control predicted image generation unit 300 as follows.
  • the boundary control predicted image generation unit 300 acquires the reference image block signal RB from the addition unit 108 and generates a predicted image block signal as follows.
  • the prediction mode of the boundary control prediction image generation unit 300 includes a prediction mode in which the prediction direction is vertical as shown in FIG. 7 and a prediction mode in which the prediction direction is horizontal as shown in FIG. Yes (two types of predicted image block signals are generated).
  • the boundary prediction control unit 301 The boundary control predicted image generation unit 300 is processed in the same manner as in the conventional prediction method. That is, the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 so as to copy the pixel value immediately before the processing target pixel in the prediction direction.
  • the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 so as to copy the pixel value of the pixel Pv1 to the pixel value of the pixel Qv1.
  • the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 to perform the following processing.
  • the boundary control prediction image generation unit 300 generates a prediction pixel according to the following formula when there is a depth edge perpendicular to the prediction direction.
  • Expression (5) is an expression for generating a prediction pixel when a depth edge exists in the horizontal direction in the prediction mode in the vertical direction.
  • Expression (6) is an expression for generating a prediction pixel when there is a depth edge in the vertical direction in the horizontal prediction mode. Since the basic processing is the same in the horizontal direction and the vertical direction, Expression (5) in the horizontal direction will be described below.
  • G [x] on the left side is a predicted pixel value of the pixel x.
  • the evaluation formula is formed from two terms.
  • the first term (
  • the second term (Dis (Qvi, pre)) represents the distance between the position of the pixel to be processed and the pixel position of each pixel in the pre.
  • the meaning of each term is that the first term is controlled so as to select a pixel having a depth value so that a pixel showing the subject considered to be the same as the subject appearing in the processing target pixel can be referred to as much as possible.
  • Term Term.
  • the term also suppresses the use of a pixel that has a subject boundary between the target pixel and the processing target pixel, as in the control by the boundary prediction control unit 301.
  • the second term is a term for selecting pixels as close as possible to the pixel to be processed.
  • ⁇ and ⁇ accumulated in each term are constants for changing the weighting between the first term and the second term. Specifically, for example, ⁇ is 0.1 and ⁇ is 1.0.
  • the sum of the first term and the second term is used in the evaluation formula, but a ratio may be used. Further, only the first term may be used.
  • the above formulas (5) and (6) are used only when there is a depth edge between the pixel to be processed and the previous pixel in the prediction direction. The above formulas (5) and (6) may be used.
  • the boundary prediction control unit 301 when there is a depth edge (subject boundary) between the pixel to be processed and the previous pixel (adjacent pixel) in the prediction direction (predetermined direction), By using the above-described equations (5) and (6) by the boundary control predicted image generation unit 300, the use of the pixel value of the previous pixel in the prediction direction is suppressed.
  • the first term is the difference in depth value between the processing target pixel and the pixel (peripheral pixel) in the previous column (or row) in the prediction direction. Therefore, it is possible to suppress the use of peripheral pixels that have a subject boundary between the processing target pixels and have a large difference in depth value.
  • the boundary control predicted image generation unit 300 generates a predicted image block predicted in the horizontal direction and the vertical direction.
  • the boundary control predicted image generation unit 300 determines the correlation between the image block input from the image input unit 101 and the predicted image block predicted in each of the two types of prediction modes using, for example, the SAD value. As a result of this determination, the boundary control prediction image generation unit 300 selects a prediction image block having a higher correlation (similar) and outputs the prediction image block to the prediction mode selection unit 202.
  • the boundary control prediction image generation unit 300 also outputs prediction encoding information indicating the prediction mode of the selected prediction image block to the prediction mode selection unit 202. In this way, since control is performed to suppress continuous pixel prediction within the depth map boundary (subject boundary) indicating the distance to the subject, the prediction accuracy can be improved.
  • FIG. 10 is a flowchart showing an image encoding process performed by the image encoding apparatus 100 according to the present embodiment.
  • Step S201 The image encoding apparatus 100 acquires an image for each frame and a depth map corresponding to the image from the outside. Thereafter, the process proceeds to step S202.
  • Step S202 The image input unit 101 divides an input image signal for each frame acquired from the outside of the image encoding device 100 into blocks of a predetermined size (for example, 16 pixels in the vertical direction ⁇ 16 pixels in the horizontal direction).
  • the depth input unit 118 divides the depth map synchronized with the image input to the image input unit 101 in the same manner as the image division performed by the image input unit 101, and sends the depth map to the depth map encoding unit 116. Output.
  • the image coding apparatus 100 repeats the processing from step S203 to step S211 for each image block in the frame.
  • the depth map encoding unit 116 encodes the depth map input from the depth input unit 118, and converts the depth map encoded data whose data amount is further compressed into the intra prediction unit 121 and the image encoding device. 100 (for example, the image decoding device 800). Thereafter, the process of step S204 and the process of step S205 are performed in parallel.
  • Step S ⁇ b> 204 The inter prediction unit 120 acquires an image block signal from the image input unit 101, and acquires a reference image block signal decoded by the addition unit 108. The inter prediction unit 120 performs inter prediction using these acquired signals. The inter prediction unit 120 outputs the inter prediction image block signal generated by the inter prediction to the prediction method control unit 109 and the selection unit 110, and outputs the inter prediction coding information IPE to the prediction method control unit 109.
  • a reset image block an image block signal with all pixel values being 0
  • it will progress to step S206.
  • Step S205 The intra prediction unit 121 acquires an image block signal from the image input unit 101, acquires depth map encoded data from the depth map encoding unit 116, and receives the reference image block signal decoded by the addition unit 108. get.
  • the intra prediction unit 121 performs intra prediction using these acquired signals.
  • the intra prediction unit 121 outputs the intra prediction image block signal generated by the intra prediction to the prediction scheme control unit 109 and the selection unit 110, and outputs the intra prediction coding information TPE to the prediction scheme control unit 109.
  • a reset image block an image block in which all pixel values are 0
  • the process of the intra estimation part 121 is completed, it will progress to step S206.
  • the prediction scheme control unit 109 receives the inter prediction image block signal and the inter prediction encoding information IPE from the inter prediction unit 120, and receives the intra prediction image block signal and the intra prediction encoding information TPE from the intra prediction unit 121. .
  • the prediction scheme control unit 109 selects a prediction mode with good coding efficiency based on the Lagrangian cost.
  • the prediction method control unit 109 outputs information indicating the selected prediction mode to the selection unit 110.
  • the prediction scheme control unit 109 outputs prediction encoding information corresponding to the selected prediction mode to the entropy encoding unit 105.
  • the selection unit 101 selects an inter prediction image block signal input from the inter prediction unit 120 or an intra prediction image block signal input from the intra prediction unit 121 according to the prediction mode information input from the prediction method control unit 109. To the subtraction unit 102 and the addition unit 108. Thereafter, the process proceeds to step S207.
  • Step S207 The subtraction unit 102 subtracts the predicted image block signal output from the selection unit 110 from the image block signal output from the image input unit 101 to generate a difference image block signal.
  • the subtraction unit 102 outputs the difference image block signal to the orthogonal transformation unit 103. Thereafter, the process proceeds to step S208.
  • the orthogonal transform unit 103 acquires the difference image block signal from the subtraction unit 102, and performs the orthogonal transform.
  • the orthogonal transform unit 103 outputs the signal after the orthogonal transform to the quantization unit 104.
  • the quantization unit 104 performs the above quantization process on the signal input from the orthogonal transform unit 103 to generate a difference image code.
  • the quantization unit 104 outputs the difference image code to the entropy coding unit 105 and the inverse quantization unit 106.
  • the entropy encoding unit 105 packs the differential image code input from the quantization unit 104 and the prediction encoding information input from the prediction scheme control unit 109, and performs variable length encoding (entropy encoding).
  • the entropy encoding unit 105 To generate encoded data E1 in which the amount of information is further compressed.
  • the entropy encoding unit 105 outputs the encoded data E1 to the outside of the image encoding device 100 (for example, the image decoding device 800). Thereafter, the process proceeds to step S209.
  • Step S209 The inverse quantization unit 106 acquires the difference image code ED from the quantization unit 104, and performs the inverse process of the quantization performed by the quantization unit 104.
  • the inverse quantization unit 106 outputs the signal generated by this processing to the inverse orthogonal transform unit 107.
  • the inverse orthogonal transform unit 107 acquires the inversely quantized signal from the inverse quantization unit 106, performs the inverse orthogonal transform process of the orthogonal transform process performed by the orthogonal transform unit 103, and obtains a difference image (decoded difference image block signal). ).
  • the inverse orthogonal transform unit 107 outputs the decoded difference image block signal to the addition unit 108. Thereafter, the process proceeds to step S210.
  • Step S210 The addition unit 108 adds the predicted image block signal output from the selection unit 110 to the decoded difference image block signal output from the inverse orthogonal transform unit 107, and decodes the input image (reference image block). signal).
  • the adding unit 108 outputs the reference image block signal to the inter prediction unit 120 and the intra prediction unit 121. Then, it progresses to step S211.
  • Step S211 When the image coding apparatus 100 has not completed the processes of Steps S203 to S210 for all the blocks in the frame, the block to be processed is changed and the process returns to Step S202. When all the processes are completed, the process ends.
  • FIG. 11 is a flowchart for explaining the processing of the inter prediction unit 120.
  • Step S ⁇ b> 301 The deblocking filter unit 111 acquires a reference image block signal from the addition unit 108 that is outside the inter prediction unit 120, and performs the FIR filter process.
  • the deblocking filter unit 111 outputs the corrected block signal after the filtering process to the frame memory 112. Thereafter, the process proceeds to step S302.
  • Step S302 The frame memory 112 acquires the correction block signal of the deblocking filter unit 111, and holds the correction block signal as a part of the image together with information that can identify the frame number. Thereafter, the process proceeds to step S303.
  • Step S303 Upon receiving the image block signal from the image input unit 101, the motion vector detection unit 114 searches the image stored in the frame memory 112 for a block similar to the image block output by the image input unit 101 (block). Matching) and generating vector information (motion vector) representing the found block. The motion vector detection unit 114 outputs information necessary for encoding including the detected vector information to the motion compensation unit 113. Thereafter, the process proceeds to step S304.
  • Step S304 The motion compensation unit 113 acquires information necessary for encoding from the motion vector detection 114, and extracts a corresponding prediction block from the frame memory.
  • the motion compensation unit 113 outputs the prediction image block signal extracted from the frame memory to the prediction method control unit 109 and the selection unit 110 as an inter prediction image block signal.
  • the motion compensation unit 113 outputs information necessary for prediction acquired from the motion vector detection unit 114 to the prediction method control unit 109. Thereafter, the inter prediction is terminated.
  • FIG. 12 is a flowchart for explaining processing of the intra prediction unit 121.
  • the depth map decoding unit 117 acquires depth map encoded data E2 from the depth map encoding unit 116, and decodes a depth map having a larger amount of information by using, for example, variable length decoding.
  • the depth map decoding unit 117 outputs the decoded depth map (depth block decoded signal) to the depth information use intra prediction unit 115. Thereafter, the process proceeds to step S402.
  • Step S402 The first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n perform processing in each prediction mode (prediction image block generation method) from the reference image block signal acquired from the addition unit 108. In response, first to n-th predicted image block signals are generated.
  • the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n output the generated first to n-th prediction image block signals to the prediction mode selection unit 202.
  • the depth use prediction mode execution unit 201 generates a prediction image block signal using depth from the reference image block signal acquired from the addition unit 108 and the depth block decoded signal acquired from the depth map decoding unit 117, and a prediction mode selection unit To 202. Thereafter, the process proceeds to step S403. Note that the predicted image generation processing performed by the depth use prediction mode execution unit 201 is as described above.
  • the prediction mode selection unit 202 receives the prediction image block signal from the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n and the depth-use prediction mode execution unit 201 and information necessary for prediction. Enter.
  • the prediction mode selection unit 202 selects a prediction mode with high coding efficiency by the above method from the input prediction image block signals (including the prediction image block signal input from the depth-based prediction mode execution unit). , Corresponding prediction mode information is generated.
  • the prediction mode selection unit 202 sends the selected prediction image block signal (hereinafter referred to as an intra prediction image block signal) to the selection unit 110 and the prediction scheme control unit 109, and the prediction mode information (hereinafter referred to as intra prediction coding information TPE). Output to the prediction method control unit 109. Then, intra prediction is complete
  • FIG. 13 is a schematic block diagram showing the configuration of the image decoding device 800 according to this embodiment.
  • the image decoding apparatus 800 includes an encoded data input unit 813, an entropy decoding unit 801, an inverse quantization unit 802, an inverse orthogonal transform unit 803, an addition unit 804, a prediction scheme control unit 805, a selection unit 806, and a deblocking filter unit 807.
  • the deblocking filter unit 807, the frame memory 808, and the motion compensation unit 809 constitute an inter processing unit 820.
  • the depth information utilization intra prediction unit 810 and the depth map decoding unit 811 constitute an intra processing unit 821.
  • the encoded data input unit 813 divides the encoded data E1 acquired from the outside (for example, the image encoding device 100) into processing block units and outputs the result to the entropy decoding unit 801.
  • the encoded data input unit 813 repeatedly outputs the blocks until the blocks are sequentially changed until all the blocks in the frame are completed and the acquired encoded data is completed.
  • the entropy decoding unit 801 performs processing reverse to the encoding method (for example, variable-length encoding) performed by the entropy encoding unit 105 on the encoded data divided into processing units acquired from the encoded data input unit 813 ( For example, entropy decoding that is variable length decoding) is performed to generate a difference image block code and predictive coding information PE.
  • the entropy decoding unit 801 outputs the difference image block code to the inverse quantization unit 802 and the prediction coding information PE to the prediction scheme control unit 805.
  • the inverse quantization unit 802 performs inverse quantization on the difference image block code input from the entropy decoding unit 801 to generate a decoded frequency domain signal, and outputs the decoded frequency domain signal to the inverse orthogonal transform unit 803.
  • the inverse orthogonal transform unit 803 generates a decoded difference image block signal that is a spatial domain signal by, for example, inverse DCT transforming the decoded frequency domain signal output from the inverse quantization unit 802.
  • the inverse orthogonal transform unit 803 can generate a spatial domain signal based on the decoded frequency domain signal
  • the inverse orthogonal transform unit 803 is not limited to the inverse DCT transform, and other methods (for example, IFFT (Inverse Fast Fourier Transform)) are used. It may be used.
  • the inverse orthogonal transform unit 803 outputs the generated decoded difference image block signal to the addition unit 804.
  • the prediction method control unit 805 extracts the prediction method PM in units of macroblocks adopted by the image coding device 100 from the prediction coding information PE input from the entropy decoding unit 801.
  • the prediction method PM is inter prediction or intra prediction.
  • the prediction method control unit 805 outputs information regarding the extracted prediction method PM to the selection unit 806.
  • the prediction scheme control unit 805 takes out the prediction coding information corresponding to the prediction scheme PM extracted from the prediction coding information PE output from the entropy decoding unit 801, and stores it in the processing unit corresponding to the extracted prediction scheme PM.
  • Predictive coding information is output.
  • the prediction method control unit 805 outputs the inter prediction coding information IPE to the inter processing unit 820 when the prediction method PM is inter prediction.
  • the prediction method control unit 805 outputs the intra prediction encoding information TPE to the intra processing unit 821 when the prediction method PM is intra prediction.
  • an inter prediction image block signal is selected.
  • an intra prediction image block signal is selected.
  • the selection unit 806 outputs the selected predicted image block signal to the addition unit 804.
  • the addition unit 804 adds the predicted image block signal output from the selection unit 806 to the decoded difference image block signal output from the inverse orthogonal transform unit 803 to generate a decoded image block signal DB.
  • the adding unit 804 outputs the decoded decoded image block signal DB to the inter processing unit 820, the intra processing unit 821, and the image output unit 812.
  • the inter processing unit 820 includes a deblocking filter unit 807, a frame memory 808, and a motion compensation unit 809.
  • the deblocking filter unit 807 performs the same processing as the FIR filter performed by the deblocking filter unit 111 on the decoded image block signal DB input from the addition unit 804, and the processing result (correction block signal) is framed. Output to the memory 808.
  • the frame memory 808 acquires the correction block signal from the deblocking filter unit 807, and holds the correction block signal as a part of the image together with information that can identify the frame number.
  • the motion compensation unit 809 acquires inter prediction coding information IPE from the prediction method control unit 805, and extracts reference image information and prediction vector information (motion vector) from the inter prediction coding information IPE.
  • the motion compensation unit 809 extracts a target image block signal (predicted image block signal) from the images stored in the frame memory 808 based on the extracted reference image information and predicted vector information.
  • the motion compensation unit 809 extracts one corresponding image block from the frame memory 808 and outputs it to the selection unit 806.
  • two prediction vectors motion vectors
  • two corresponding image blocks are taken out from the frame memory 808, averaged, and output to the selection unit 806.
  • This signal output from the inter processing unit 820 (motion compensation unit 809) to the selection unit 806 is an inter prediction image block signal.
  • the intra processing unit 821 includes a depth information use intra prediction unit 810 and a depth map decoding unit 811.
  • the depth map encoded data input unit 814 divides the depth map encoded data E2 input from the outside (for example, the image encoding device 100) into processing blocks, and outputs them to the intra processing unit 821.
  • the depth map decoding unit 811 reverses the block unit depth map encoded data output from the depth map encoded data input unit 814 to the encoding method (for example, variable length encoding) performed by the depth map encoding unit 116.
  • the depth block decoded signal is generated by performing entropy decoding which is the above process (for example, variable length decoding).
  • the depth map decoding unit 811 outputs the depth block decoded signal to the depth information use intra prediction unit 810.
  • FIG. 14 is a schematic block diagram illustrating a configuration of the depth information use intra prediction unit 810.
  • the depth information use intra prediction unit 810 includes a first prediction mode execution unit 900-1, a second prediction mode execution unit 900-2, an nth prediction mode execution unit 900-n, a depth use prediction mode execution unit 901, and a prediction mode selection. A portion 902 is included.
  • the prediction mode selection unit 902 includes an index (prediction mode) indicating the prediction mode created by the prediction mode selection unit 202 of the image encoding device 100 from the intra prediction encoding information TPE output by the prediction method control unit 805, Each piece of information necessary for prediction is extracted.
  • the information necessary for prediction is extracted because the prediction mode indicated by the index is a prediction mode in which information necessary for prediction exists (specifically, the sub-blocks of the first prediction mode and the second prediction mode). This is a case where a prediction image is generated in units and a depth use prediction mode).
  • the prediction mode selection unit 902 extracts information necessary for prediction, the prediction mode selection unit 902 outputs the information to the corresponding prediction mode execution units 900-1 to 900-n and 901.
  • the prediction mode selection unit 902 selects a prediction image block signal of the prediction mode indicated by the index (prediction mode) from the prediction image block signals generated by each prediction mode execution unit, and selects the prediction image block signal as an intra prediction image block signal. Output to the unit 806.
  • the first prediction mode execution unit 900-1, the second prediction mode execution unit 900-2, and the nth prediction mode execution unit 900-n are provided in the depth information use intra prediction unit 115 of the image encoding device 100.
  • the same processing as that of the execution unit 200-1, the second prediction mode execution unit 200-2, and the nth prediction mode execution unit 200-n is performed.
  • the prediction mode (necessary for prediction) in each sub-block is used.
  • Information is input from the prediction mode selection unit 902, and the corresponding prediction mode is executed in units of sub-blocks.
  • the prediction mode it is the content shown in FIG.
  • the depth use prediction mode execution unit 901 acquires information necessary for prediction (specifically, information indicating the direction of prediction) from the prediction mode selection unit 902, and acquires a depth block decoded signal from the depth map decoding unit 811.
  • the depth use prediction mode execution unit 901 uses the acquired information and signal to generate a predicted image block signal as performed by the depth use prediction mode execution unit 201 of the image encoding device 100.
  • the information necessary for prediction is information regarding the direction of prediction selected by the depth use prediction mode execution unit 201.
  • the configuration of the depth usage prediction mode execution unit 901 is basically the same as the configuration of the depth usage prediction mode execution unit 201.
  • the depth control prediction mode is implemented as the final process in which the boundary control prediction image generation unit 300 of the image encoding device 100 selects the prediction block in the horizontal direction and the prediction block in the vertical direction based on the correlation between the input image and the input image.
  • the boundary control predicted image generation unit 300 of the unit 901 is different in that selection is performed using information necessary for prediction.
  • the depth use prediction mode execution unit 901 generates the same predicted image block signal as the depth use prediction mode execution unit 201 at the time of encoding.
  • FIG. 15 is a flowchart showing an image decoding process performed by the image decoding apparatus 800 according to this embodiment.
  • the image decoding apparatus 800 acquires the encoded data E including the encoded data E1 of the image and the encoded data E2 of the depth map from the image encoding apparatus 100 via the communication network 500. Thereafter, the process proceeds to step S602.
  • Step S602 The encoded data input unit 813 divides the acquired encoded data E1 of the image into processing blocks corresponding to a predetermined size (for example, 16 pixels in the vertical direction ⁇ 16 pixels in the horizontal direction) to generate entropy.
  • the data is output to the decoding unit 801.
  • the depth map encoded data input unit 814 inputs depth map encoded data synchronized with the encoded data input to the encoded data input unit 813 from the outside of the image decoding apparatus 800, and the encoded data input unit 813.
  • the data is divided into the same processing units as the division performed in the above and output to the intra processing unit 821.
  • the image decoding apparatus 800 repeats the processing in steps S603 to S608 for each image block in the frame.
  • the entropy decoding unit 801 performs entropy decoding on the encoded data output from the encoded data input unit 813, and generates a difference image block code and predictive encoding information.
  • the entropy decoding unit 801 outputs the difference image block code to the inverse quantization unit 802, and outputs the prediction coding information to the prediction scheme control unit 805.
  • the prediction scheme control unit 805 acquires the prediction coding information from the entropy decoding unit 801, and extracts information regarding the prediction scheme PM and the prediction coding information corresponding to the prediction scheme PM.
  • each block may be processed in parallel, or only one of the processes may be performed in accordance with the prediction method PM.
  • Step S604 The inter processing unit 820 acquires the inter prediction coding information IPE output from the prediction scheme control unit 805 and the decoded image block signal DB output from the adding unit 804, and performs inter processing.
  • the inter processing unit 820 outputs the generated inter predicted image block signal to the selection unit 806. The contents of the inter processing will be described later.
  • a reset image block signal an image block signal in which all pixel values are 0
  • the process proceeds to step S606.
  • Step S605 The intra processing unit 821 acquires the intra prediction encoded information TPE output from the prediction scheme control unit 805 and the decoded image block signal DB output from the adding unit 804, and performs intra prediction.
  • the intra processing unit 821 outputs the generated intra predicted image block signal to the selection unit 806.
  • the intra prediction process will be described later. In the first process, when the process of the adding unit 804 is not completed, a reset image block signal (an image block signal in which all pixel values are 0) is input. If the process of the intra estimation part 821 is completed, it will progress to step S606.
  • Step S606 The selection unit 806 acquires information on the prediction method PM output from the prediction method control unit 805, and outputs the inter prediction image block signal output from the inter processing unit 820 or the intra prediction output from the intra processing unit 821. An image signal is selected and output to the adder 804. Thereafter, the process proceeds to step S607.
  • Step S607 The inverse quantization unit 802 performs the inverse process of the quantization performed by the quantization unit 104 of the image coding device 100 on the difference image block code input from the entropy decoding unit 801.
  • the inverse quantization unit 802 outputs the generated decoded frequency domain signal to the inverse orthogonal transform unit 803.
  • the inverse orthogonal transform unit 803 obtains the inversely quantized decoded frequency domain signal from the inverse quantization unit 802, and performs the inverse orthogonal transform process of the orthogonal transform process performed by the orthogonal transform unit 103 of the image encoding device 100. Then, the difference image (decoded difference image block signal) is decoded.
  • the inverse orthogonal transform unit 803 outputs the decoded decoded difference image block signal to the adding unit 804.
  • the adding unit 804 adds the predicted image block signal output from the selection unit 806 to the decoded difference image block signal output from the inverse orthogonal transform unit 803 to generate a decoded image block signal DB.
  • the adding unit 804 outputs the decoded decoded image block signal DB to the image output unit 812, the inter processing unit 820, and the intra processing unit 821. Thereafter, the process proceeds to step S608.
  • Step S608 The image output unit 812 generates the output image signal R ′ by arranging the decoded image block signal DB output by the adding unit 804 at a corresponding position in the image. If the processes in steps S603 to S607 have not been completed for all the blocks in the frame, the block to be processed is changed and the process returns to step S602.
  • the image output unit 812 When outputting the generated output image signal R ′ to the outside of the image decoding device 800 (display device 600), the image output unit 812, for example, has 5 frames (I picture (I0), B picture ( B3), B picture (B2), B picture (B4) and P picture (P1)).
  • FIG. 16 is a flowchart for explaining the inter processing in step S604.
  • the deblocking / filtering unit 807 acquires the decoded image block signal DB from the adding unit 804 that is external to the inter processing unit 820, and performs the FIR filter processing performed at the time of encoding.
  • the deblocking filter unit 807 outputs the corrected corrected block signal to the frame memory 808. Thereafter, the process proceeds to step S702.
  • Step S ⁇ b> 702 The frame memory 808 holds the correction block signal output from the deblocking filter unit 807 as part of the image together with information that can identify the frame number. Thereafter, the process proceeds to step S703. (Step S703)
  • the motion compensation unit 809 acquires the inter prediction coding information IPE from the prediction scheme control unit 805, and extracts a corresponding prediction block signal from the frame memory.
  • the motion compensation unit 809 outputs the prediction image block signal extracted from the frame memory to the selection unit 806 as an inter prediction image block signal. Thereafter, the inter processing is terminated.
  • FIG. 17 is a flowchart illustrating the intra processing in step S605.
  • the depth map decoding unit 811 acquires depth map encoded data divided into processing units from the depth map encoded data input unit 814, and decodes a depth map having a larger amount of information using variable length decoding, for example. To do.
  • the depth map decoding unit 811 outputs the decoded depth map (depth block decoded signal) to the depth information use intra prediction unit 810. Thereafter, the process proceeds to step S802.
  • Step S802 The first prediction mode execution unit 900-1 to the n-th prediction mode execution unit 900-n generate a prediction image block signal using the decoded image block signal DB output from the addition unit 804.
  • the prediction mode execution unit that performs processing in units of sub-blocks specifically, the first prediction mode execution unit 900-1 and the second prediction mode execution unit 900-2 are employed in the image coding apparatus 100.
  • Information indicating the prediction mode of each sub-block is acquired from the prediction mode selection unit 902, and a prediction image block signal is generated.
  • First prediction mode execution unit 900-1 to n-th prediction mode execution unit 900-n output the generated first to n-th prediction image block signals to prediction mode selection unit 902.
  • the depth use prediction mode execution unit 901 includes a decoded image block signal DB output from the addition unit 108, a depth block decoded signal output from the depth depth map decoding unit 811, and information necessary for prediction output from the prediction mode selection unit 902. (Specifically, information indicating the direction of prediction) is used to perform the same processing as the processing performed by the depth use prediction mode execution unit 201 in FIG. 3 to generate a depth use prediction image.
  • the depth use prediction mode execution unit 901 outputs the generated prediction image signal to the prediction mode selection unit 902. Thereafter, the process proceeds to step S803.
  • the prediction mode selection unit 902 extracts information indicating the prediction mode employed by the image encoding device 100 from the intra prediction encoding information TPE input from the prediction method control unit 805, and predicts the corresponding prediction mode.
  • the image block signal is output to the selection unit 806 as an intra prediction image block signal.
  • the prediction mode selection unit 902 further extracts the prediction mode of each subblock and outputs the information to the corresponding prediction mode execution unit. Then, intra prediction is complete
  • the prediction mode selection unit 902 extracts information regarding the prediction direction and outputs the information to the depth use prediction mode execution unit 901.
  • the image encoding device 100 described above includes the depth input unit 118 and the depth map encoding unit 116, and the image decoding device 800 includes the depth map encoded data input unit 814 and the depth map decoding unit 811. It is not limited to this.
  • information regarding the depth map corresponding to the input image may be made available in the image decoding apparatus 800 by a separate means.
  • the image encoding device 100 and the image decoding device 800 are configured to receive the depth map via a communication line from a server device that stores the depth map in correspondence with video information installed outside or offline. May be. Therefore, a video title indicating video information can be searched through a communication line, and when the video information is selected, a corresponding depth map can be received.
  • the image encoding device 100 also acquires an image of a viewpoint different from the input image, and between the pixels included in the input image and the pixels included in the image of the viewpoint different from the input image.
  • a depth map generation unit that generates a depth map having a pixel value as a value representing the parallax of the image may be provided. In that case, the depth map generation unit outputs the generated depth map to the depth input unit 118.
  • the image decoding apparatus 800 generates a second output image having a viewpoint different from the output image based on the output image and the depth map of the same frame as the output image, and outputs the second output image to the outside. Also good.
  • the image encoding apparatus 100 inputs the input image signal every 5 frames.
  • the image encoding apparatus 100 is not limited to this and may input every arbitrary number of frames.
  • the image decoding apparatus 800 outputs the output image signal every 5 frames.
  • the image decoding apparatus 800 is not limited to this and may output every arbitrary number of frames. .
  • the image to be encoded is a moving image, but it may be a still image.
  • the image to be encoded is a multi-viewpoint image, and the depth-use prediction mode is used only in the viewpoint image with the corresponding depth map.
  • the conventional prediction mode is used for the viewpoint image without the corresponding depth map. May be used.
  • the present embodiment has two prediction modes for performing control to suppress continuous pixel prediction at the boundary of the depth map indicating the distance to the subject when performing intra prediction. Since only two prediction modes are added as compared with the prior art, it is possible to improve the accuracy of the predicted image while suppressing an increase in code amount due to an increase in the number of prediction modes. And the accuracy of the prediction image is improved while suppressing the increase in the amount of code due to the increase in the number of prediction modes, so the residual between the prediction image and the input image is minimized, realizing highly efficient image encoding and decoding can do. If the depth-based prediction mode is used instead of the conventional prediction mode, the number of prediction modes does not increase, and therefore the increase in code amount can be further suppressed.
  • a part of the image coding apparatus 100 and the image decoding apparatus 800 in the above-described embodiment for example, the subtraction unit 102, the orthogonal transformation unit 103, the quantization unit 104, the entropy coding unit 105, the inverse quantization unit 106, and the inverse Orthogonal transformation unit 107, addition unit 108, prediction scheme control unit 109, selection unit 110, deblocking filter unit 111, motion compensation unit 113, motion vector detection unit 114, depth information use intra prediction unit 115, depth map encoding unit 116, a depth map decoding unit 117, an entropy decoding unit 801, an inverse quantization unit 802, an inverse orthogonal transform unit 803, an addition unit 804, a prediction scheme control unit 805, a selection unit 806, a deblocking filter unit 807, a motion compensation unit 809, a depth information use intra prediction unit 810 and a depth map decoding unit 811 are connected. It may be realized by Yuta.
  • the program for realizing the control function may be recorded on a computer-readable recording medium, and the program recorded on the recording medium may be read into a computer system and executed.
  • the “computer system” here is a computer system built in the image encoding device 100 or the image decoding device 800, and includes an OS and hardware such as peripheral devices.
  • the “computer-readable recording medium” refers to a storage device such as a flexible medium, a magneto-optical disk, a portable medium such as a ROM or a CD-ROM, and a hard disk incorporated in a computer system.
  • the “computer-readable recording medium” is a medium that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line,
  • a volatile memory inside a computer system serving as a server or a client may be included and a program that holds a program for a certain period of time.
  • the program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system.
  • part or all of the image encoding device 100 and the image decoding device 800 in the above-described embodiment may be realized as an integrated circuit such as an LSI (Large Scale Integration).
  • LSI Large Scale Integration
  • Each functional block of the image encoding device 100 and the image decoding device 800 may be individually made into a processor, or a part or all of them may be integrated into a processor.
  • the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. Further, in the case where an integrated circuit technology that replaces LSI appears due to progress in semiconductor technology, an integrated circuit based on the technology may be used.
  • an image for performing intra prediction when an input image is encoded, an image for performing intra prediction that predicts a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel.
  • the encoding device wherein when performing the intra prediction, a screen that suppresses using the peripheral pixels having a subject boundary represented by the input image between the processing target pixels and the peripheral pixels.
  • An inner prediction unit is provided.
  • Another aspect of the present invention is the image encoding device described above, wherein the intra-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the input image.
  • a boundary detection unit is provided.
  • the intra prediction unit includes a pixel adjacent to the processing target pixel in a predetermined direction among the peripheral pixels, and the processing target.
  • the pixel value of the processing target pixel is predicted using the pixel adjacent in the predetermined direction, and the pixel adjacent in the predetermined direction and the processing target.
  • a predicted image generation unit is provided that suppresses predicting the pixel value of the processing target pixel using a pixel adjacent in the predetermined direction.
  • Another aspect of the present invention is the above-described image encoding device, in which the in-screen prediction unit represents a peripheral pixel used when predicting a pixel value of the processing target pixel.
  • a prediction image generation unit that determines at least based on the difference between the information indicating the distance to the subject and the information indicating the distance to the subject represented by the processing target pixel.
  • the intra-screen prediction unit uses the peripheral pixels and the processing target pixels as peripheral pixels used when predicting the pixel value of the processing target pixel.
  • a predicted image generation unit that determines at least based on the distance.
  • intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel.
  • the image decoding device when performing the intra prediction, it is possible to suppress use of the peripheral pixels having a subject boundary represented by the encoded image between the processing target pixels and the peripheral pixels.
  • An in-screen prediction unit is provided.
  • Another aspect of the present invention is the above-described image decoding device, wherein the in-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the encoded image.
  • a boundary detection unit is provided.
  • the intra prediction unit includes a pixel adjacent to the processing target pixel in a predetermined direction among the peripheral pixels, and the processing target pixel. If there is no boundary between the subject and the pixel, the pixel value of the processing target pixel is predicted using pixels adjacent in the predetermined direction, and the pixel adjacent to the predetermined direction and the processing target pixel A prediction image generation unit that suppresses prediction of a pixel value of the processing target pixel using pixels adjacent in the predetermined direction when the boundary of the subject is between the two.
  • Another aspect of the present invention is the above-described image decoding device, wherein the intra-screen prediction unit determines a peripheral pixel used when predicting a pixel value of the processing target pixel to a subject represented by the peripheral pixel.
  • a predicted image generation unit that determines at least based on the difference between the information indicating the distance to the subject and the information indicating the distance to the subject represented by the processing target pixel.
  • Another aspect of the present invention is the above-described image decoding device, wherein the intra-screen prediction unit uses the peripheral pixels and the processing target as peripheral pixels used when predicting a pixel value of the processing target pixel.
  • a predicted image generation unit that determines at least based on the distance to the pixel;
  • intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel.
  • the intra prediction when the intra prediction is performed, use of the peripheral pixels having a boundary of a subject represented by the input image between the processing target pixels and the peripheral pixels is suppressed. Have a process.
  • intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel.
  • the intra prediction when the intra prediction is performed, use of the peripheral pixels having a boundary of a subject represented by the encoded image between the processing target pixels and the peripheral pixels is suppressed. Have a process.
  • intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel.
  • the computer of the image encoding device performs the intra prediction, the use of the peripheral pixels having the boundary of the subject represented by the input image between the processing target pixels and the peripheral pixels is suppressed. It is a program for functioning as an in-screen prediction unit.
  • intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel.
  • the computer of the image decoding apparatus performs the intra prediction, the use of the peripheral pixels having a boundary of the subject represented by the encoded image between the peripheral pixels and the processing target pixel is suppressed. It is a program for functioning as an in-screen prediction unit.
  • DESCRIPTION OF SYMBOLS 10 Moving image transmission system 100 ... Image coding apparatus 101 ... Image input part 102 ... Subtraction part 103 ... Orthogonal transformation part 104 ... Quantization part 105 ... Entropy coding part 106 ... Inverse quantization part 107 ... Inverse orthogonal transformation part 108 DESCRIPTION OF SYMBOLS ... Addition unit 109 ... Prediction method control unit 110 ... Selection unit 111 ... Deblocking filter unit 112 ... Frame memory unit 113 ... Motion compensation unit 114 ... Motion vector detection unit 115 ... Depth information utilization intra prediction unit 116 ... Depth map encoding 117: Depth map decoding unit 118 ... Depth input unit 120 ...
  • Inter prediction unit 121 ... Intra prediction unit 200-1 ... First prediction mode execution unit 200-2 ... Second prediction mode execution unit 200-n ... nth prediction mode Execution unit 201 ... depth use prediction mode execution unit 202 ... prediction mode selection unit 300 ... boundary Control predicted image generation unit 301 ... Boundary prediction control unit 302 ... Subject boundary detection unit 500 ... Communication network 600 ... Display device 800 ... Image decoding device 801 ... Entropy decoding unit 802 ... Inverse quantization unit 803 ... Inverse orthogonal transformation unit 804 ... Addition 805 ... Prediction method control unit 806 ... Selection unit 807 ... Deblocking filter unit 808 ... Frame memory 809 ... Motion compensation unit 810 ...
  • Depth information use intra prediction unit 811 Depth map decoding unit 812 ... Image output unit 813 ... Encoding Data input unit 814 ; Depth map encoded data input unit 820 ... Inter processing unit 821 ... Intra processing unit 900-1 ... First prediction mode execution unit 900-2 ... Second prediction mode execution unit 900-n ... nth prediction mode Implementation unit 901 ... Depth use prediction mode implementation unit 902 ... Prediction mode selection unit

Abstract

The present invention is an image encoding device which, when encoding an input image, uses pixel values of surrounding pixels in a periphery of a pixel to be processed in order to perform intra-screen prediction for predicting the pixel value of the pixel to be processed, and is characterized in being provided with an intra-screen prediction unit which determines the predictive value of each pixel on the basis of information indicating the distance to a subject for the pixel to be processed and information indicating the distance to the subject for the surrounding pixels.

Description

画像符号化装置、画像復号装置、画像符号化方法、画像復号方法およびプログラムImage encoding device, image decoding device, image encoding method, image decoding method, and program
 本発明は、画像符号化装置、画像復号装置、画像符号化方法、画像復号方法およびプログラムに関する。
 本願は、2011年5月25日に、日本に出願された特願2011-117425号に基づき優先権を主張し、その内容をここに援用する。
The present invention relates to an image encoding device, an image decoding device, an image encoding method, an image decoding method, and a program.
This application claims priority on May 25, 2011 based on Japanese Patent Application No. 2011-117425 filed in Japan, the contents of which are incorporated herein by reference.
 近年、ブロードバンドなどの通信インフラの普及や、携帯電話やビデオカメラなどによって手軽に映像を撮影できる映像取得機器の普及、あるいはテレビ放送をHDD(Hard Disk Drive;ハードディスクドライブ)やBD(Blu-ray Disc;ブルーレイディスク)に録画保存できるハードディスクBDレコーダなどの普及によって、一般消費者が大容量の映像を取り扱う機会が増えてきている。また、映像表示方式の高解像度化に伴って取り扱う映像の容量は益々大きくなり、高性能な映像圧縮技術が必要とされている。 In recent years, communication infrastructure such as broadband has become widespread, video acquisition devices that can easily shoot video with mobile phones, video cameras, etc., or television broadcasting via HDD (Hard Disk Drive) or BD (Blu-ray Disc) With the widespread use of hard disk BD recorders that can record and store on a Blu-ray Disc), the opportunity for general consumers to handle large-capacity video is increasing. In addition, with the increase in resolution of the video display system, the volume of video to be handled increases more and more, and a high-performance video compression technique is required.
 そうした中、国際標準動画圧縮規格H.264(非特許参考文献1)が、画質向上および符号化効率向上を目的として、2003年に規格化された。H.264は、画像を複数のブロック(以下、マクロブロック)に分割を行い、複数の予測方式を駆使し、そのマクロブロック単位で符号化効率の高い予測方式を逐次選択することで符号化効率を高めている。 In such circumstances, the international standard video compression standard H.264 H.264 (Non-patent Reference 1) was standardized in 2003 for the purpose of improving image quality and encoding efficiency. H. H.264 increases the coding efficiency by dividing an image into a plurality of blocks (hereinafter referred to as macroblocks), making full use of a plurality of prediction methods, and sequentially selecting a prediction method having a high coding efficiency for each macroblock. ing.
 H.264で行われる符号化方法については、画面内で符号化が既に完了したブロックの画素情報を用いて、符号化対象ブロックを予測して符号化する画面内予測符号化方式(イントラ予測符号化方式)と、処理を行う画像とは異なる画像を参照して符号化対象ブロックを予測して符号化する画像間予測符号化方式(インター予測符号化方式)がある。 H. As for the encoding method performed in H.264, an intra prediction encoding method (intra prediction encoding method) that predicts and encodes an encoding target block using pixel information of a block that has already been encoded in the screen. ) And an inter-picture predictive coding method (inter-predictive coding method) for predicting and coding an encoding target block with reference to an image different from an image to be processed.
 イントラ予測符号化方式は、マクロブロック(16画素×16画素)単位、あるいはマクロブロックをさらに分割した4画素×4画素、8画素×8画素(8画素×8画素は、H.264 FRExtにて策定)単位で、規定の予測モードに従って生成される予測画像と符号化を行う元の画像との差分画像(残差成分)を符号化した際の符号量と、予測モードを特定するための情報を符号化する際に必要な符号量に基づいて、最適な予測方式が選択される(非特許文献1)。 The intra prediction encoding method is a macroblock (16 pixels × 16 pixels) unit, or 4 pixels × 4 pixels and 8 pixels × 8 pixels (8 pixels × 8 pixels are further divided by H.264 FRExt). Information for identifying the prediction mode and the code amount when the difference image (residual component) between the prediction image generated according to the prescribed prediction mode and the original image to be encoded is encoded in units) An optimal prediction method is selected based on the amount of code necessary for encoding (Non-patent Document 1).
 16画素×16画素のブロックに対しては、4種類の予測モード(図6、図6については後述する)が適用可能で、1種類のDC成分による予測(平均値予測)と3種類の予測角度を用いた予測(垂直予測、水平予測、平面予測)がある。4画素×4画素、あるいは8画素×8画素のブロックに対しては、9種類の予測モード(図4、図4については後述する)が適用可能であり、1種類のDC成分による予測(平均値予測)と8種類の予測角度を用いた予測(45°~206.57°の不均等角度の予測)がある。 Four types of prediction modes (FIGS. 6 and 6 will be described later) can be applied to a block of 16 pixels × 16 pixels, and prediction using one type of DC component (average value prediction) and three types of prediction are possible. There are predictions using angles (vertical prediction, horizontal prediction, planar prediction). Nine kinds of prediction modes (FIGS. 4 and 4 will be described later) can be applied to a block of 4 pixels × 4 pixels or 8 pixels × 8 pixels, and prediction (average) by one type of DC component is applicable. Value prediction) and prediction using eight types of prediction angles (prediction of non-uniform angles of 45 ° to 206.57 °).
 また、予測モードを特定するための情報(例えばモードを示すインデックス値)の符号化については、処理対象ブロックの上部と左部の予測モードを用いて予測を行い、その予測に合致する場合には、1ビット分のフラグを用意し、合致したというフラグを立てて判断する。予測に合致しない場合には、そのフラグを立てずに、合致しない予測モードを除いた残りの8種類の予測モードを判断するための3ビット分の情報を付加して符号化する。予測が合えば、予測モードを符号化するのに1ビットの情報で済むが、予測が合わなければ4ビット分の情報を必要とする。 In addition, regarding encoding of information for specifying a prediction mode (for example, an index value indicating a mode), prediction is performed using the prediction modes of the upper part and the left part of the processing target block, and when the prediction matches with the prediction mode A 1-bit flag is prepared, and a determination is made by setting a flag that matches. If the prediction does not match, the flag is not set, and encoding is performed by adding information for 3 bits for determining the remaining eight types of prediction modes excluding the prediction mode that does not match. If the prediction is correct, only one bit of information is required to encode the prediction mode, but if the prediction is not correct, information of 4 bits is required.
 例えば、特許文献1に記載の発明は、イントラ予測符号化方式の符号化効率を向上させることを目的に、任意の予測角度で予測ができるように、H.264の方式に比べて予測モードの数を増やしている。予測角度と処理を行う画素の位置から、理論上の参照画素位置を求め、その画素位置に該当する画素値を、周辺の参照画素を補間することにより生成する技術について開示している。
 特許文献2に記載の発明は、周囲の予測モードの角度に基づいて精度よく予測モードを予測する方式について開示している。
For example, the invention described in Japanese Patent Application Laid-Open No. H10-228707 is H.264 so that prediction can be performed at an arbitrary prediction angle for the purpose of improving the encoding efficiency of the intra prediction encoding scheme. The number of prediction modes is increased compared to the H.264 method. A technique is disclosed in which a theoretical reference pixel position is obtained from a predicted angle and the position of a pixel to be processed, and a pixel value corresponding to the pixel position is generated by interpolating surrounding reference pixels.
The invention described in Patent Document 2 discloses a method for accurately predicting a prediction mode based on the angle of surrounding prediction modes.
特開2009-284275号公報JP 2009-284275 A 特開2010-056701号公報JP 2010-056701 A
 しかしながら、特許文献1に記載の画面内予測方式では、予測精度を向上させるために予測の角度を細かくする必要がある。予測の角度を細かくすると、その角度に1対1に対応する予測モード数が増える。その結果として予測モードを表現するビット数が増えるため符号化する際の符号量が増加する問題がある。
 また、特許文献2に記載の画面内予測方式は、予測モードを従来に比べて精度良く予測できたとしても、予測画像の精度を向上させることはできない。
However, in the intra prediction method described in Patent Document 1, it is necessary to make the angle of prediction fine in order to improve the prediction accuracy. If the angle of prediction is made fine, the number of prediction modes corresponding to the angle one-to-one increases. As a result, since the number of bits representing the prediction mode increases, there is a problem that the amount of codes at the time of encoding increases.
Moreover, even if the prediction method in a screen described in patent document 2 can predict a prediction mode more accurately compared with the past, it cannot improve the precision of a prediction image.
 本発明は、このような事情に鑑みてなされたもので、その目的は、符号量の増加を抑えつつ、予測画像の精度を向上させる符号化装置、復号装置、符号化方法、復号方法およびプログラムを提供することにある。 The present invention has been made in view of such circumstances, and an object of the present invention is to provide an encoding device, a decoding device, an encoding method, a decoding method, and a program that improve the accuracy of a predicted image while suppressing an increase in the amount of code. Is to provide.
 (1)この発明は上述した課題を解決するためになされたもので、本発明の一態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置であって、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部を具備することを特徴とする画像符号化装置である。 (1) The present invention has been made to solve the above-described problems, and according to one aspect of the present invention, when an input image is encoded, pixel values of peripheral pixels around the processing target pixel are used. An image encoding apparatus that performs intra prediction to predict a pixel value of the processing target pixel, and includes information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels. An image coding apparatus comprising an intra-screen prediction unit that determines a predicted value for each pixel based on the image.
 (2)また、本発明の他の態様は、上述の画像符号化装置であって、前記画面内予測部は、前記処理対象画素と前記周辺画素との画素間の距離に基づいて画素ごとの予測値を決定することを特徴とする。 (2) Moreover, the other aspect of this invention is the above-mentioned image coding apparatus, Comprising: The said prediction part in a screen is for every pixel based on the distance between the pixels of the said process target pixel and the said surrounding pixel. A predicted value is determined.
 (3)また、本発明の他の態様は、上述の画像符号化装置であって、前記画面内予測部は、前記被写体の境界を、前記入力画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を備え、前記処理対象画素と所定の方向に隣接する画素との間に前記境界が検出されなかった場合には、前記所定の方向に隣接する画素を用いて前記処理対象画素の予測値を予測することを特徴とする。 (3) According to another aspect of the present invention, there is provided the above-described image encoding device, wherein the intra-screen prediction unit uses information indicating a boundary of the subject and a distance to the subject of the input image. A subject boundary detection unit for detecting, and when the boundary is not detected between the processing target pixel and a pixel adjacent in a predetermined direction, the processing target is detected using the pixel adjacent in the predetermined direction; A prediction value of a pixel is predicted.
 (4)また、本発明の他の態様は、符号化画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置であって、前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部
を具備することを特徴とする画像復号装置である。
(4) According to another aspect of the present invention, when a coded image is decoded, a pixel value of a processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel. An image decoding device that performs prediction, and when performing the intra prediction, information for each pixel based on information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels. An image decoding apparatus comprising an in-screen prediction unit that determines a predicted value.
 (5)また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記処理対象画素と前記周辺画素との画素間の距離に基づいて画素ごとの予測値を決定することを特徴とする。 (5) Moreover, the other aspect of this invention is the above-mentioned image decoding apparatus, Comprising: The said prediction part in a screen is the prediction for every pixel based on the distance between the pixels of the said process target pixel and the said surrounding pixel. It is characterized by determining a value.
 (6)また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記被写体の境界を、前記入力画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を備え、前記処理対象画素と所定の方向に隣接する画素との間に前記境界が検出されなかった場合には、前記所定の方向に隣接する画素を用いて前記処理対象画素の予測値を予測することを特徴とする。 (6) According to another aspect of the present invention, there is provided the above-described image decoding device, wherein the in-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the input image. A subject boundary detection unit that detects the boundary between the processing target pixel and a pixel adjacent in a predetermined direction, and uses the pixel adjacent in the predetermined direction to detect the processing target pixel. The prediction value is predicted.
 (7)また、本発明の他の態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化方法であって、前記画面内予測を行う際に、画面内予測部が前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する過程を有することを特徴とする画像符号化方法である。 (7) According to another aspect of the present invention, when an input image is encoded, a pixel value of the processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel. An image encoding method for performing prediction, wherein when performing intra prediction, information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels This is an image coding method characterized by having a process of determining a predicted value for each pixel based on the above.
 (8)また、本発明の他の態様は、入力画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号方法であって、前記画面内予測を行う際に、画面内予測部が前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する過程を有することを特徴とする画像復号方法である。 (8) Further, according to another aspect of the present invention, when decoding an input image, intra-frame prediction that predicts a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel. When the intra prediction is performed, the intra prediction unit is based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. And a process for determining a predicted value for each pixel.
 (9)また、本発明の他の態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置を、前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部として機能させるためのプログラムである。 (9) According to another aspect of the present invention, when an input image is encoded, a pixel value of the processing target pixel is predicted using pixel values of peripheral pixels around the processing target pixel. When the image coding apparatus that performs prediction performs the intra prediction, prediction for each pixel is performed based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. It is a program for functioning as an in-screen prediction unit for determining a value.
 (10)また、本発明の他の態様は、入力画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置を、前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部として機能させるためのプログラムである。 (10) According to another aspect of the present invention, when decoding an input image, intra-frame prediction that predicts the pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel. When the image decoding apparatus performing the intra prediction, the prediction value for each pixel is calculated based on the information indicating the distance to the subject in the processing target pixel and the information indicating the distance to the subject in the peripheral pixels. It is a program for functioning as an in-screen prediction unit to be determined.
 この発明によれば、符号量の増加を抑えつつ、画面内予測符号化における、予測画像の精度を向上させることができる。 According to the present invention, it is possible to improve the accuracy of a predicted image in intra prediction encoding while suppressing an increase in code amount.
本発明の一実施形態による画像伝送システムの構成を示すブロック図である。It is a block diagram which shows the structure of the image transmission system by one Embodiment of this invention. 同実施形態における画像符号化装置の構成を示す概略ブロック図である。It is a schematic block diagram which shows the structure of the image coding apparatus in the embodiment. 同実施形態における奥行き情報利用イントラ予測部の構成を示す概略ブロック図である。It is a schematic block diagram which shows the structure of the depth information utilization intra estimation part in the embodiment. 同実施形態における4×4画素のサブブロック単位の画面内予測の予測モードを示す図である。It is a figure which shows the prediction mode of the prediction in a screen of the subblock unit of 4x4 pixel in the embodiment. 同実施形態における4×4画素のサブブロック単位の画面内予測の符号化順序を示す図である。It is a figure which shows the encoding order of the prediction in a screen of the subblock unit of 4x4 pixel in the embodiment. 同実施形態における16×16画素単位の画面内予測の予測モードを示す図である。It is a figure which shows the prediction mode of the prediction in a screen of a 16x16 pixel unit in the embodiment. 同実施形態におけるデプス利用予測モード実施部の処理概念を説明する図(垂直方向)である。It is a figure (vertical direction) explaining the processing concept of the depth utilization prediction mode implementation part in the embodiment. 同実施形態におけるデプス利用予測モード実施部の処理概念を説明する図(水平方向)である。It is a figure (horizontal direction) explaining the processing concept of the depth utilization prediction mode implementation part in the embodiment. 同実施形態におけるデプス利用予測モード実施部の構成を示す概略ブロック図である。It is a schematic block diagram which shows the structure of the depth utilization prediction mode implementation part in the embodiment. 同実施形態における画像符号化装置が行う画像符号化処理を示すフローチャートである。It is a flowchart which shows the image coding process which the image coding apparatus in the embodiment performs. 同実施形態における画像符号化装置が実行するインター予測処理を示すフローチャートである。It is a flowchart which shows the inter prediction process which the image coding apparatus in the embodiment performs. 同実施形態における画像符号化装置が実行するイントラ予測処理を示すフローチャートである。It is a flowchart which shows the intra prediction process which the image coding apparatus in the embodiment performs. 同実施形態における画像復号装置の構成を示す概略ブロック図である。It is a schematic block diagram which shows the structure of the image decoding apparatus in the embodiment. 同実施形態における奥行き情報利用イントラ予測部の構成を示す概略ブロック図である。It is a schematic block diagram which shows the structure of the depth information utilization intra estimation part in the embodiment. 同実施形態における画像復号装置が実行する画像復号処理を示すフローチャートである。It is a flowchart which shows the image decoding process which the image decoding apparatus in the embodiment performs. 同実施形態における画像復号装置が実行するインター処理を示すフローチャートである。It is a flowchart which shows the inter process which the image decoding apparatus in the embodiment performs. 同実施形態における画像復号装置が実行するイントラ処理を示すフローチャートである。It is a flowchart which shows the intra process which the image decoding apparatus in the embodiment performs.
 以下、図面を参照して、本発明の実施の形態について説明する。図1は、この発明の一実施形態による動画像伝送システムの構成を示す概略ブロック図である。図1に示すように、本実施形態における動画像伝送システム10は、画像符号化装置100、通信網500、画像復号装置800、表示装置600を含んで構成される。画像符号化装置100は、符号化の対象となる画像の画像信号Rと該画像に対応したデプスマップのデプスマップ信号Dとから、該画像およびデプスマップを符号化し、これらを符号化したデータである符号化データEを生成して出力する。通信網500は、画像符号化装置100が出力した符号化データEを、画像復号装置800に伝送する。画像復号装置800は、伝送されてきた符号化データEを復号し、復号結果の画像の画像信号R’を生成する。表示装置600は、液晶ディスプレイ、プラズマディスプレイなどの画像表示デバイスを有し、画像復号装置800が生成した画像信号R’が示す画像を表示する。 Hereinafter, embodiments of the present invention will be described with reference to the drawings. FIG. 1 is a schematic block diagram showing a configuration of a moving image transmission system according to an embodiment of the present invention. As shown in FIG. 1, the moving image transmission system 10 in this embodiment includes an image encoding device 100, a communication network 500, an image decoding device 800, and a display device 600. The image encoding device 100 encodes the image and the depth map from the image signal R of the image to be encoded and the depth map signal D of the depth map corresponding to the image, and uses the encoded data. A certain encoded data E is generated and output. The communication network 500 transmits the encoded data E output from the image encoding device 100 to the image decoding device 800. The image decoding apparatus 800 decodes the transmitted encoded data E, and generates an image signal R ′ of the decoded image. The display device 600 includes an image display device such as a liquid crystal display or a plasma display, and displays an image indicated by the image signal R ′ generated by the image decoding device 800.
 画像符号化装置100は、例えば、テレビジョン放送局に設けられ、放送番組を符号化する。この場合、通信網500は、放送波を用いて送信する通信網であり、画像復号装置800および表示装置600は、テレビジョン受像機に備えられる。あるいは、通信網500としてインターネットや携帯電話網などを用いてもよい。また、画像符号化装置100は、DVD(Digital Versatile Disc)やBD(Blu-ray Disc)に格納して販売されるコンテンツを編集するコンテンツホルダに設けられ、これらのコンテンツを符号化する。
この場合、符号化画像EはDVD、BDなどに格納され、通信網500に変えて、配送網により、配送される。そして、画像復号装置800は、DVDプレーヤ、BDプレーヤなどに備えられる。
The image encoding device 100 is provided in a television broadcasting station, for example, and encodes a broadcast program. In this case, the communication network 500 is a communication network that transmits using broadcast waves, and the image decoding device 800 and the display device 600 are provided in a television receiver. Alternatively, the Internet or a mobile phone network may be used as the communication network 500. The image encoding apparatus 100 is provided in a content holder that edits contents stored and sold on a DVD (Digital Versatile Disc) or a BD (Blu-ray Disc), and encodes these contents.
In this case, the encoded image E is stored in a DVD, a BD, or the like, and is delivered by a delivery network instead of the communication network 500. The image decoding device 800 is provided in a DVD player, a BD player, or the like.
 図2は、本実施形態に係る画像符号化装置100の構成を示す概略ブロック図である。
画像符号化装置100は、画像入力部101、減算部102、直交変換部103、量子化部104、エントロピー符号化部105、逆量子化部106、逆直交変換部107、加算部108、予測方式制御部109、選択部110、デブロッキング・フィルタ部111、フレームメモリ部112、動き補償部113、動きベクトル検出部114、奥行き情報利用イントラ予測部115、デプスマップ符号化部116、デプスマップ復号部117及びデプス入力部118を含んで構成される。なお、デブロッキング・フィルタ部111と、フレームメモリ112と、動き補償部113と、動きベクトル検出114とで、インター予測部120を構成する。また、奥行き情報利用イントラ予測部115と、デプスマップ復号部117とで、イントラ予測部121を構成する。
FIG. 2 is a schematic block diagram illustrating a configuration of the image encoding device 100 according to the present embodiment.
The image coding apparatus 100 includes an image input unit 101, a subtraction unit 102, an orthogonal transformation unit 103, a quantization unit 104, an entropy coding unit 105, an inverse quantization unit 106, an inverse orthogonal transformation unit 107, an addition unit 108, and a prediction method. Control unit 109, selection unit 110, deblocking filter unit 111, frame memory unit 112, motion compensation unit 113, motion vector detection unit 114, depth information use intra prediction unit 115, depth map encoding unit 116, depth map decoding unit 117 and a depth input unit 118. The de-blocking filter unit 111, the frame memory 112, the motion compensation unit 113, and the motion vector detection 114 constitute an inter prediction unit 120. In addition, the depth information use intra prediction unit 115 and the depth map decoding unit 117 constitute an intra prediction unit 121.
 画像入力部101は、符号化の対象となる画像(入力画像)を示す画像信号R(入力画像信号)を、一例として5フレーム毎(5フレームの種類については後述する)に、画像符号化装置100の外部から取得する。画像入力部101は、取得した入力画像信号が表す入力画像フレームを予め定めた大きさ(例えば、垂直方向16画素×水平方向16画素)のブロックに分割する。画像入力部101は、分割したブロックの各々を表す画像ブロック信号Bを、減算部102、動きベクトル検出部114及び奥行き情報利用イントラ予測部115に出力する。画像入力部101は、ブロック位置を順次変えながら、画像フレーム内のすべてのブロックについての出力が完了し、そして取得した画像が終了するまで、各画像フレームに対して、この処理を繰り返す。 The image input unit 101, for example, outputs an image signal R (input image signal) indicating an image to be encoded (input image) as an example every 5 frames (the types of 5 frames will be described later). 100 from outside. The image input unit 101 divides the input image frame represented by the acquired input image signal into blocks having a predetermined size (for example, 16 pixels in the vertical direction × 16 pixels in the horizontal direction). The image input unit 101 outputs an image block signal B representing each of the divided blocks to the subtraction unit 102, the motion vector detection unit 114, and the depth information use intra prediction unit 115. The image input unit 101 repeats this process for each image frame until the output for all the blocks in the image frame is completed and the acquired image is completed while sequentially changing the block position.
 本実施形態において、画像符号化装置100への入力画像は少なくとも基準画像(ベースビュー;base view)を含む。基準画像とは、立体表示用の多視点(マルチビュー;multi-view)の動画像に含まれる予め定めた一つの視点の画像であって、デプスマップを算出する基礎となる画像である。デプスマップ(depth map)は、基準画像に表された被写体の撮影装置からの奥行又は距離を表す距離情報であって、基準画像の画素毎に与えられた量子化値を含んで構成される。その量子化値の各々をデプス(depth)値といい、例えば8ビットで量子化された値をとる。 In this embodiment, the input image to the image encoding device 100 includes at least a reference image (base view). The reference image is an image of one predetermined viewpoint included in a multi-view (multi-view) moving image for stereoscopic display, and is an image serving as a basis for calculating a depth map. The depth map is distance information representing the depth or distance from the photographing device of the subject represented in the reference image, and includes a quantized value given for each pixel of the reference image. Each of the quantized values is called a depth value, for example, a value quantized with 8 bits.
 画像入力部101に入力される5フレーム毎の画像信号Rは、例えばIピクチャ(I0)、Bピクチャ(B3)、Bピクチャ(B2)、Bピクチャ(B4)及びPピクチャ(P1)の画像信号を含む。画像符号化装置100に入力される画像信号Rは、例えば、この順序(以下、入力順序)で入力される。ここで、符号(I0等)において、先頭のI等は画像の種類を示し、0等は符号化される順序(以下、符号化順序)を示す(従って、入力順序と符号化順序は異なる)。Iピクチャとは、フレーム内画像(Intra Frame Picture)であって、これを符号化した符号だけを用いて入力画像を復号できる画像である。Pピクチャとは、フレーム間順方向予測画像(Predictive Picture)であって、これを符号化した符号と、過去のフレームの画像信号を符号化した符号を用いて入力画像を復号できる画像である。Bピクチャとは、双方向予測符号化画像(Bi-directional Predictive Picture)であって、これを復号した符号と、過去もしくは未来の複数のフレームの画像信号を符号化した符号を用いて入力画像を復号できる画像である。 The image signal R for every five frames input to the image input unit 101 is, for example, an image signal of an I picture (I0), a B picture (B3), a B picture (B2), a B picture (B4), and a P picture (P1). including. The image signal R input to the image encoding device 100 is input in this order (hereinafter referred to as input order), for example. Here, in the code (I0, etc.), the leading I, etc., indicates the type of image, and 0, etc., indicates the order of encoding (hereinafter, encoding order) (therefore, the input order and the encoding order are different). . An I picture is an intra-frame picture (Intra Frame Picture), which can be decoded using only a code obtained by encoding the picture. The P picture is an inter-frame forward prediction image (Predictive Picture), which is an image that can be decoded using a code obtained by encoding the image and a code obtained by encoding an image signal of a past frame. A B picture is a bi-predictive coded picture (Bi-directional Predictive Picture), which is obtained by decoding the input picture and using a code obtained by coding a plurality of past or future frame image signals. It is an image that can be decoded.
 減算部102は、画像入力部101が出力した画像ブロック信号から、選択部110が出力した予測画像ブロック信号を減算して、差分画像ブロック信号を生成する。減算部102は、生成した差分画像ブロック信号を直交変換部103に出力する。
 直交変換部103は、減算部102が出力した差分画像ブロック信号を直交変換し、種々の周波数特性の強度を示す信号を生成する。
The subtraction unit 102 subtracts the prediction image block signal output from the selection unit 110 from the image block signal output from the image input unit 101 to generate a difference image block signal. The subtraction unit 102 outputs the generated difference image block signal to the orthogonal transformation unit 103.
The orthogonal transform unit 103 performs orthogonal transform on the difference image block signal output from the subtraction unit 102 to generate signals indicating the strengths of various frequency characteristics.
 直交変換部103は、差分画像ブロック信号を直交変換する際、その差分画像ブロック信号を、例えばDCT変換(離散コサイン変換;Discrete Cosine Transform)して周波数領域信号(例えば、DCT変換を行った場合は、DCT係数)を生成する。直交変換部103は、差分画像ブロック信号に基づき周波数領域信号を生成することができれば、DCT変換に限らず、他の方法(例えば、FFT(高速フーリエ変換;Fast Fourier Transform))を用いてもよい。直交変換部103は、生成した周波数領域信号に含まれる係数値を、量子化部104に出力する。
 量子化部104は、直交変換部103が出力した各周波数特性強度を示す係数値を量子化し、生成した量子化信号ED(差分画像ブロック符号)を、エントロピー符号化部105と逆量子化部106に出力する。
When the orthogonal transform unit 103 orthogonally transforms the difference image block signal, the difference image block signal is subjected to, for example, DCT transform (discrete cosine transform) and a frequency domain signal (for example, DCT transform). , DCT coefficients). As long as the orthogonal transform unit 103 can generate a frequency domain signal based on the difference image block signal, other methods (for example, FFT (Fast Fourier Transform)) may be used instead of the DCT transform. . The orthogonal transform unit 103 outputs the coefficient value included in the generated frequency domain signal to the quantization unit 104.
The quantization unit 104 quantizes the coefficient value indicating each frequency characteristic intensity output from the orthogonal transform unit 103, and generates the generated quantized signal ED (difference image block code) with the entropy encoding unit 105 and the inverse quantization unit 106. Output to.
 逆量子化部106は、量子化部104が出力した量子化信号EDを、逆量子化して復号周波数領域信号を生成し、逆直交変換部107に出力する。
 逆直交変換部107は、入力された復号周波数領域信号を、例えば逆DCT変換して空間領域信号である復号差分画像ブロック信号を生成する。逆直交変換部107は、復号周波数領域信号に基づき空間領域信号を生成することができれば、逆DCT変換に限らず、他の方法(例えば、IFFT(高速フーリエ逆変換;Inverse Fast Fourier Transform))を用いてもよい。
 逆直交変換部107は、生成した復号差分画像ブロック信号を加算部108に出力する。
The inverse quantization unit 106 performs inverse quantization on the quantized signal ED output from the quantization unit 104 to generate a decoded frequency domain signal, and outputs the decoded frequency domain signal to the inverse orthogonal transform unit 107.
The inverse orthogonal transform unit 107 performs, for example, inverse DCT transform on the input decoded frequency domain signal to generate a decoded differential image block signal that is a spatial domain signal. As long as the inverse orthogonal transform unit 107 can generate a spatial domain signal based on the decoded frequency domain signal, the inverse orthogonal transform unit 107 is not limited to the inverse DCT transform, and other methods (eg, IFFT (Inverse Fast Fourier Transform)) are used. It may be used.
The inverse orthogonal transform unit 107 outputs the generated decoded difference image block signal to the addition unit 108.
 加算部108は、選択部110から予測画像ブロック信号を取得し、逆直交変換部107から復号差分画像ブロック信号を取得する。加算部108は、予測画像ブロック信号に復号差分画像ブロック信号を加算し、入力画像を符号化・復号した参照画像ブロック信号RBを生成する(内部デコード)。この参照画像ブロック信号RBは、インター予測部120及びイントラ予測部121に出力される。 The addition unit 108 acquires the predicted image block signal from the selection unit 110 and acquires the decoded difference image block signal from the inverse orthogonal transform unit 107. The adder 108 adds the decoded differential image block signal to the predicted image block signal, and generates a reference image block signal RB obtained by encoding / decoding the input image (internal decoding). The reference image block signal RB is output to the inter prediction unit 120 and the intra prediction unit 121.
 インター予測部120は、加算部108より参照画像ブロック信号RBを取得し、画像入力部101より画像ブロック信号を取得する。インター予測部120は、これらの信号を用いてインター予測し、インター予測画像ブロック信号を生成する。インター予測部120は、生成したインター予測画像ブロック信号を予測方式制御部109と選択部110に出力する。同時に、インター予測部120は、生成したインター予測符号化情報IPEを、予測方式制御部109に出力する。インター予測部120については後述する。 The inter prediction unit 120 acquires the reference image block signal RB from the addition unit 108 and acquires the image block signal from the image input unit 101. The inter prediction unit 120 performs inter prediction using these signals, and generates an inter prediction image block signal. The inter prediction unit 120 outputs the generated inter prediction image block signal to the prediction method control unit 109 and the selection unit 110. At the same time, the inter prediction unit 120 outputs the generated inter prediction coding information IPE to the prediction scheme control unit 109. The inter prediction unit 120 will be described later.
 イントラ予測部121は、加算部108より参照画像ブロック信号RBを取得し、画像入力部101より画像ブロック信号を取得し、デプスマップ符号化部116よりデプスマップ符号化データを取得する。イントラ予測部121は、これらの信号およびデータを用いてイントラ予測し、イントラ予測画像ブロック信号を生成する。イントラ予測部121は、生成したイントラ予測画像ブロック信号を、予測方式制御部109と選択部110に出力する。同時に、イントラ予測部121は、生成したイントラ予測符号化情報TPEを、予測方式制御部109に出力する。イントラ予測部121については後述する。 The intra prediction unit 121 acquires the reference image block signal RB from the addition unit 108, acquires the image block signal from the image input unit 101, and acquires depth map encoded data from the depth map encoding unit 116. The intra prediction unit 121 performs intra prediction using these signals and data, and generates an intra predicted image block signal. The intra prediction unit 121 outputs the generated intra prediction image block signal to the prediction scheme control unit 109 and the selection unit 110. At the same time, the intra prediction unit 121 outputs the generated intra prediction encoding information TPE to the prediction scheme control unit 109. The intra prediction unit 121 will be described later.
 デプス入力部118は、画像入力部101に入力される入力画像に対応するデプスマップのデプスマップ信号Dを、画像符号化装置100の外部より取得する。デプス入力部118は、取得したデプスマップを、画像入力部101において分割された入力画像ブロックと同じ位置でかつ、同じブロックサイズとなるように、分割(デプスブロック信号)し、デプスマップ符号化部116に出力する。デプスマップ符号化部116は、デプス入力部118が出力したデプスブロック信号を、例えば可変長符号化(エントロピー符号化)を用いて符号化し、データ量がより圧縮されたデプスマップ符号化データE2を生成する。デプスマップ符号化部116は、生成したデプスマップ符号化データE2をイントラ予測部121と画像符号化装置100の外部(例えば、通信網500を介して画像復号装置800)に出力する。 The depth input unit 118 acquires the depth map signal D of the depth map corresponding to the input image input to the image input unit 101 from the outside of the image encoding device 100. The depth input unit 118 divides (depth block signal) the acquired depth map so that the input image block divided by the image input unit 101 has the same position and the same block size, and the depth map encoding unit. To 116. The depth map encoding unit 116 encodes the depth block signal output from the depth input unit 118 using, for example, variable length encoding (entropy encoding), and converts the depth map encoded data E <b> 2 whose data amount is further compressed. Generate. The depth map encoding unit 116 outputs the generated depth map encoded data E2 to the intra prediction unit 121 and the outside of the image encoding device 100 (for example, the image decoding device 800 via the communication network 500).
 続いて、インター予測部120について説明する。インター予測部120は、デブロッキング・フィルタ部111、フレームメモリ112、動き補償部113、及び動きベクトル検出部114より構成される。
 デブロッキング・フィルタ部111は、加算部108より参照画像ブロック信号RBを取得し、画像の符号化時に発生するブロック歪みを減少させるための、例えば公知の符号化方法(例えば、H.264 Reference Software JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, 2008)で用いられたFIR(Finite Impulse Response;有限インパルス応答)フィルタ処理を行う。デブロッキング・フィルタ部111は、処理結果(補正ブロック信号)をフレームメモリ112に出力する。
Next, the inter prediction unit 120 will be described. The inter prediction unit 120 includes a deblocking filter unit 111, a frame memory 112, a motion compensation unit 113, and a motion vector detection unit 114.
The deblocking filter unit 111 acquires the reference image block signal RB from the adder unit 108, and reduces the block distortion generated when the image is encoded, for example, a known encoding method (for example, H.264 Reference Software). FIR (Finite Impulse Response) filter processing used in JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, 2008). The deblocking filter unit 111 outputs the processing result (correction block signal) to the frame memory 112.
 フレームメモリ112は、デブロッキング・フィルタ部111が出力した補正ブロック信号を、フレーム番号を同定できる情報と共に、該フレーム番号の画像の一部として保持する。
 動きベクトル検出部114は、画像入力部101より入力される画像ブロック信号に類似するブロックを、フレームメモリ112に蓄積された画像より探し出し(ブロックマッチング)、探し出したブロックを表すベクトル情報(動きベクトル)を生成する。動きベクトル検出部114は、ブロックマッチングを行う際、当該分割されたブロックとの間の指標値を領域毎に算出し、算出した指標値が最小となる領域を探し出す。入力画像信号がBピクチャである場合は、動きベクトル検出部114は、その指標値が最小となる参照画像の領域におけるブロックと、次に指標値が小さい参照画像の領域におけるブロックと、2つの領域を探し出す。
The frame memory 112 holds the correction block signal output from the deblocking filter unit 111 as a part of the image of the frame number together with information for identifying the frame number.
The motion vector detection unit 114 searches for a block similar to the image block signal input from the image input unit 101 from the image stored in the frame memory 112 (block matching), and vector information (motion vector) representing the searched block. Is generated. When performing block matching, the motion vector detection unit 114 calculates an index value between the divided blocks for each area, and searches for an area where the calculated index value is minimum. When the input image signal is a B picture, the motion vector detection unit 114 has two areas: a block in the reference image area having the smallest index value, a block in the reference image area having the next smallest index value, and the two areas. Find out.
 指標値は、画像信号間の相関性や類似性を示すものであればよい。動きベクトル検出部114は、例えば、分割されたブロックに含まれる画素の輝度値と参照画像のある領域における輝度値の差の絶対値総和(SAD;Sum of Absolute Difference)を用いる。入力画像信号から分割されたブロック(例えば、大きさがN×N画素)と参照画像信号のブロックとの間のSADは次の式(1)で表される。 The index value only needs to indicate the correlation or similarity between image signals. The motion vector detection unit 114 uses, for example, an absolute value sum (SAD: Sum of Absolute Difference) of a difference between a luminance value of a pixel included in a divided block and a luminance value in a certain region of the reference image. SAD between a block (for example, a size of N × N pixels) divided from the input image signal and the block of the reference image signal is expressed by the following equation (1).
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 式(1)において、Iin(i0+i,j0+j)は入力画像の座標(i0+i,j0+j)における輝度値、(i0,j0)は当該分割されたブロックの左上端の画素座標を示す。Iref(i0+i+p,j0+j+q)は参照画像の座標(i0+i+p,j0+j+q)における輝度値、(p,q)は当該分割されたブロックの左上端の座標を基準にしたシフト量(動きベクトル)である。 In Expression (1), Iin (i0 + i, j0 + j) represents the luminance value at the coordinates (i0 + i, j0 + j) of the input image, and (i0, j0) represents the pixel coordinates of the upper left corner of the divided block. Iref (i0 + i + p, j0 + j + q) is a luminance value at the coordinates (i0 + i + p, j0 + j + q) of the reference image, and (p, q) is a shift amount (motion vector) based on the coordinates of the upper left corner of the divided block.
 即ち、動きベクトル検出部114は、ブロックマッチングにおいて、(p,q)毎にSAD(p,q)を算出し、SAD(p,q)を最小とする(p,q)を探し出す。(p,q)は、入力画像中の当該分割されたブロックから参照画像中の当該参照領域の位置までのベクトル(動きベクトル)を表す。
 動き補償部113は、動きベクトル検出部114より動きベクトルを取得し、該当する参照ブロックをインター予測画像ブロック信号として、予測方式制御部109と選択部110に出力する。動き補償部113は、動きベクトル検出部114が出力した動きベクトルが1つの場合には、該当する画像ブロックを出力し、動きベクトル検出部114が出力した動きベクトルが2つの場合には、該当する2つの画像ブロックを平均化して出力する。動き補償部113は、予測に必要な情報(以下、インター予測符号化情報IPE)、例えば動きベクトルを予測方式制御部109へ出力する。
That is, the motion vector detection unit 114 calculates SAD (p, q) for each (p, q) in block matching, and finds (p, q) that minimizes SAD (p, q). (P, q) represents a vector (motion vector) from the divided block in the input image to the position of the reference region in the reference image.
The motion compensation unit 113 acquires a motion vector from the motion vector detection unit 114, and outputs the corresponding reference block to the prediction scheme control unit 109 and the selection unit 110 as an inter prediction image block signal. The motion compensator 113 outputs the corresponding image block when the motion vector output from the motion vector detector 114 is one, and applies when the motion vector output from the motion vector detector 114 is two. Two image blocks are averaged and output. The motion compensation unit 113 outputs information necessary for prediction (hereinafter, inter prediction coding information IPE), for example, a motion vector, to the prediction scheme control unit 109.
 続いて、イントラ予測部121について説明する。イントラ予測部121は、デプスマップ復号部117、及び奥行き情報利用イントラ予測部115より構成される。
 デプスマップ復号部117は、デプスマップ符号化部116が出力したデプスマップ符号化データを、例えば可変長復号を用いて情報量がより多いデプスブロック信号を復号する。デプスマップ復号部117は、復号したデプスマップD’(デプスブロック復号信号)を奥行き情報利用イントラ予測部115に出力する。
Next, the intra prediction unit 121 will be described. The intra prediction unit 121 includes a depth map decoding unit 117 and a depth information use intra prediction unit 115.
The depth map decoding unit 117 decodes the depth block signal having a larger amount of information by using, for example, variable length decoding, the depth map encoded data output from the depth map encoding unit 116. The depth map decoding unit 117 outputs the decoded depth map D ′ (depth block decoded signal) to the depth information use intra prediction unit 115.
 図3は、本実施形態に係る奥行き情報利用イントラ予測部115の構成を示す概略ブロック図である。図3を用いて、奥行き情報利用イントラ予測部115の処理について説明する。奥行き情報利用イントラ予測部115は、具体的には、第1予測モード実施部200-1~第n予測モード実施部200-n(nは1以上の自然数、例えば6)、デプス利用予測モード実施部201、及び予測モード選択部202を含んで構成される。 FIG. 3 is a schematic block diagram illustrating a configuration of the depth information use intra prediction unit 115 according to the present embodiment. The processing of the depth information use intra prediction unit 115 will be described with reference to FIG. Specifically, the depth information use intra prediction unit 115 includes a first prediction mode execution unit 200-1 to an nth prediction mode execution unit 200-n (n is a natural number of 1 or more, for example, 6), and a depth use prediction mode execution. Unit 201 and a prediction mode selection unit 202.
 第1予測モード実施部200-1~第n予測モード実施部200-nは、加算部108が出力した参照画像ブロック信号RBから各予測モード(予測画像ブロックの生成方式)による処理に従って第1~第n予測画像ブロック信号をそれぞれ生成する。第1予測モード実施部200-1~第n予測モード実施部200-nは、生成した第1~第n予測画像ブロック信号を予測モード選択部202に出力する。 The first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n perform the first to nth prediction mode execution units 200-n according to the processing in each prediction mode (prediction image block generation method) from the reference image block signal RB output from the addition unit 108. An n-th predicted image block signal is generated. The first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n output the generated first to n-th prediction image block signals to the prediction mode selection unit 202.
 第1予測モード実施部200-1~第n予測モード実施部200-nの各々は、例えば従来の画面内予測モード(例えば、H.264 Reference Software JM ver. 13.2 Encoder,http://iphome.hhi.de/suehring/tml/, 2008)のいずれかを用いて、画面内予測(イントラ予測)を実施する。H.264では、マクロブロックをさらに分割した4×4画素のサブブロックに適用する9種類の画面内予測と、マクロブロック単位に適用する4種類の画面内予測方式がある(なお、8×8画素のサブブロックを用いた画面内予測については、H.264 FRExtにて策定されていて、4×4画素と同じ画面内予測方式が適用される。)。 Each of the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n is, for example, a conventional intra-screen prediction mode (for example, H.264 Reference Software JM ver. 13.2 Encoder, http: // iphome. hhi.de/suehring/tml/, 2008) is used to perform in-screen prediction (intra prediction). H. In H.264, there are nine types of intra prediction applied to a 4 × 4 pixel sub-block obtained by further dividing a macroblock, and four types of intra prediction applied to a macroblock unit (note that 8 × 8 pixels). (In-screen prediction using sub-blocks is formulated in H.264 FRExt, and the same intra-screen prediction method as 4 × 4 pixels is applied.)
 具体的には、第1予測モード実施部200-1は、例えば4×4のサブブロックを用いたイントラ予測(画面内予測)を行う。第2予測モード実施部200-2は、例えば8×8のサブブロックを用いたイントラ予測を行う。第3予測モード実施部200-3~第6予測モード実施部200-6は、例えば16×16のマクロブロック単位の4種類の予測方式を行う。 Specifically, the first prediction mode execution unit 200-1 performs intra prediction (intra-screen prediction) using, for example, 4 × 4 sub-blocks. The second prediction mode execution unit 200-2 performs intra prediction using, for example, 8 × 8 sub-blocks. The third prediction mode execution unit 200-3 to the sixth prediction mode execution unit 200-6 perform four types of prediction methods, for example, in units of 16 × 16 macroblocks.
 第1予測モード実施生成部200-1は、加算部108が出力した参照画像ブロック信号をさらに4×4のサブブロックサイズに分割をし、図5に示した順番で4×4画素単位の予測方式を実行する。すなわち、16×16画素のブロックを、4つの8×8画素のブロックに分割し、これらを左上、右上、左下、右下の順に処理する。そして、これらの8×8画素のブロックの各々については、それぞれを、4つの4×4画素のサブブロックに分割し、これら各々の8×8画素のブロック内で、左上、右上、左下、右下の順にイントラ予測を行う。 The first prediction mode execution generation unit 200-1 further divides the reference image block signal output from the addition unit 108 into 4 × 4 sub-block sizes, and performs prediction in units of 4 × 4 pixels in the order shown in FIG. Execute the method. That is, a 16 × 16 pixel block is divided into four 8 × 8 pixel blocks, which are processed in the order of upper left, upper right, lower left, and lower right. For each of these 8 × 8 pixel blocks, each is divided into four 4 × 4 pixel sub-blocks, and within each of these 8 × 8 pixel blocks, upper left, upper right, lower left, right Intra prediction is performed in the following order.
 4×4画素のサブブロックの予測方式は、図4に示したように予測モード0から予測モード8までの9種類の予測方式がある。第1予測モード実施部200-1は、9種類の予測方式の各々で生成した4×4画素の予測画像ブロックと、画像入力部101が出力した画像ブロック信号Bの対応するサブブロックとの間の相関度を示す指標を算出し、その指標に基づき、サブブロック毎に予測方式を選択する。第1予測モード実施部200-1は、その指標として例えば、輝度値の差の絶対値総和(SAD)を算出し、SADの値が一番小さくなる予測方式を該当する4×4画素のサブブロック予測方式として選択し、対応する位置の第1予測画像ブロック信号を生成する。また、その予測方式を保持しておく。
 第1予測モード実施部200-1は、16×16画素分の予測方式と第1予測画像ブロック信号が生成されるまで上記処理を繰り返す。
As shown in FIG. 4, there are nine types of prediction methods from prediction mode 0 to prediction mode 8 as the prediction method for a 4 × 4 pixel sub-block. The first prediction mode execution unit 200-1 includes a 4 × 4 pixel predicted image block generated by each of nine types of prediction methods and a corresponding sub-block of the image block signal B output from the image input unit 101. An index indicating the degree of correlation is calculated, and a prediction method is selected for each sub-block based on the index. The first prediction mode execution unit 200-1 calculates, for example, the absolute value sum (SAD) of the luminance value differences as the index, and sets the prediction method with the smallest SAD value to the corresponding 4 × 4 pixel sub A block prediction method is selected, and a first predicted image block signal at a corresponding position is generated. Also, the prediction method is retained.
The first prediction mode execution unit 200-1 repeats the above process until a prediction method for 16 × 16 pixels and a first predicted image block signal are generated.
 第2予測モード実施部200-2は、加算部108が出力した参照画像ブロック信号RBをさらに4つの8×8画素のサブブロックに分割し、第1予測モード実施部200-1で用いた予測モード0から予測モード8と同様の9種類の予測方式を8×8画素のサブブロックの各々に適用し、予測画像を生成する。同時に、その予測方式を保持しておく。
 第2予測モード実施部200-2は、上記処理を繰り返し、順次8×8画素のサブブロック単位で予測方式を確定させ、16×16画素ブロックの全ての予測方式と、その予測方式による予測画像ブロック信号を生成する。
The second prediction mode execution unit 200-2 further divides the reference image block signal RB output from the addition unit 108 into four 8 × 8 pixel sub-blocks, and uses the prediction used in the first prediction mode execution unit 200-1. Nine kinds of prediction methods similar to those in the mode 0 to the prediction mode 8 are applied to each of the 8 × 8 pixel sub-blocks to generate a prediction image. At the same time, the prediction method is retained.
The second prediction mode execution unit 200-2 repeats the above processing, sequentially determines the prediction method in units of 8 × 8 pixel sub-blocks, and predicts all 16 × 16 pixel block prediction methods and predicted images based on the prediction method. Generate block signals.
 第3予測モード実施部200-3から第6予測モード実施部200-6は、16×16画素単位のイントラ予測(画面内予測)で、加算部108が出力した参照画像ブロック信号を用いて、図6の予測モード0から予測モード3に対応する予測画像ブロック信号をそれぞれ生成する。 The third prediction mode execution unit 200-3 to the sixth prediction mode execution unit 200-6 are 16 × 16 pixel intra predictions (intra-screen predictions), and use the reference image block signal output from the addition unit 108. Prediction image block signals corresponding to prediction modes 0 to 3 in FIG. 6 are generated.
 デプス利用予測モード実施部201は、加算部108より参照画像ブロック信号とデプスマップ復号部117よりデプスブロック復号信号を取得し、デプスマップを用いて、被写体の境界を跨ぐ予測を抑制した画面内予測を行う。デプス利用予測モード実施部201の詳細については後述する。デプス利用予測モード実施部201は、予測画像ブロック信号と予測方式を予測モード選択部202に出力する。 The depth use prediction mode execution unit 201 obtains a reference image block signal from the addition unit 108 and a depth block decoded signal from the depth map decoding unit 117, and uses the depth map to suppress prediction across a subject boundary. I do. Details of the depth use prediction mode execution unit 201 will be described later. The depth use prediction mode execution unit 201 outputs the prediction image block signal and the prediction method to the prediction mode selection unit 202.
 予測モード選択部202は、第1予測モード実施部200-1から第n予測モード実施部200-n、およびデプス利用予測モード実施部201が生成した予測画像ブロック信号と予測に必要な情報を取得する。予測に必要な情報とは、例えば16×16画素をサブブロックにさらに分割して処理を行う第1予測モード実施部200-1、及び第2予測モード実施部200-2の各サブブロックに適用された予測モードを示す情報や、デプス利用予測モード実施部201の予測の方向を示す予測モードを示す情報である。 The prediction mode selection unit 202 acquires the prediction image block signal generated by the first prediction mode execution unit 200-1 to the nth prediction mode execution unit 200-n and the depth use prediction mode execution unit 201 and information necessary for prediction. To do. The information necessary for prediction is applied to each sub-block of the first prediction mode execution unit 200-1 and the second prediction mode execution unit 200-2 that perform processing by further dividing 16 × 16 pixels into sub-blocks, for example. Information indicating the predicted mode, and information indicating the prediction mode indicating the direction of prediction of the depth use prediction mode execution unit 201.
 予測モード選択部202は、取得した予測画像ブロック信号の中から(デプス利用予測モード実施部201が出力する予測画像ブロック信号も含む)、最も指標値が小さくなる1つの予測画像ブロック信号を選択する。予測モード選択部202は、その指標値として、例えば次式で示すように、画像入力部101から入力された入力画像に含まれる該当する画像ブロックの輝度値Iin(i0+i,j0+j)と候補予測画像ブロックの輝度値Ip,m(i0+i,j0+j)とのSADを用いる。 The prediction mode selection unit 202 selects one prediction image block signal having the smallest index value from the obtained prediction image block signals (including the prediction image block signal output by the depth use prediction mode execution unit 201). . As the index value, the prediction mode selection unit 202, for example, as shown by the following equation, the luminance value Iin (i0 + i, j0 + j) of the corresponding image block included in the input image input from the image input unit 101 and the candidate prediction image SAD with the luminance value Ip, m (i0 + i, j0 + j) of the block is used.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 式(2)において、mは、どの予測モード実施部のどの予測モードであるかを表すインデックスである。したがって、Ip,m(x,y)は、予測モードmによる予測画像の座標x,yにおける輝度値である。また、i0、j0は、当該ブロックの左上頂点の座標であり、Nはブロックの大きさ(一辺の画素数)である。なお、本実施形態には、SAD以外にも指標値として入力画像に含まれる画像ブロックと候補予測画像ブロックの相関、類似度、又は符号化後の情報量など、予測モード毎の処理の有効性を表す変数であれば利用することができる。 In Expression (2), m is an index indicating which prediction mode of which prediction mode execution unit is. Accordingly, Ip, m (x, y) is a luminance value at the coordinates x, y of the predicted image in the prediction mode m. Further, i0 and j0 are the coordinates of the upper left vertex of the block, and N is the size of the block (the number of pixels on one side). In the present embodiment, in addition to SAD, the effectiveness of processing for each prediction mode, such as the correlation between the image block included in the input image as an index value and the candidate predicted image block, the similarity, or the amount of information after encoding. Any variable that represents can be used.
 予測モード選択部202は、この予測モードを表すインデックスを含む予測モード情報を生成する。あるいは、予測モード選択部202は、予測に必要な情報が存在する予測モードを選択した場合(具体的には、第1予測モード実施部200-1、第2予測モード実施部200-2及びデプス利用予測モード実施部201の予測モード)には、インデックスとこの予測に必要な情報をまとめて、予測モード情報を生成する。
 予測モード選択部202は、選択した予測画像ブロック信号(以下、イントラ予測画像ブロック信号)を選択部110と予測方式制御部109へ、また前記予測モード情報(以下、イントラ予測符号化情報TPE)を予測方式制御部109へ出力する。
The prediction mode selection unit 202 generates prediction mode information including an index representing the prediction mode. Alternatively, the prediction mode selection unit 202 selects a prediction mode in which information necessary for prediction exists (specifically, the first prediction mode execution unit 200-1, the second prediction mode execution unit 200-2, and the depth). In the prediction mode of the use prediction mode execution unit 201, the index and information necessary for this prediction are collected to generate prediction mode information.
The prediction mode selection unit 202 sends the selected prediction image block signal (hereinafter referred to as an intra prediction image block signal) to the selection unit 110 and the prediction scheme control unit 109, and the prediction mode information (hereinafter referred to as intra prediction coding information TPE). Output to the prediction method control unit 109.
 図2に戻り、予測方式制御部109は、入力画像のピクチャの種類と、インター予測部120より入力されるインター予測画像ブロック信号とそのインター予測符号化情報IPE、及びイントラ予測部121より入力されるイントラ予測画像ブロック信号とそのイントラ符号化情報とに基づいて予測方式を決め、その予測方式の情報を選択部110とエントロピー符号化部105に出力する。予測方式制御部109は、入力画像のピクチャの種類を監視し、入力画像がIピクチャの場合はイントラ予測方式を選択する。PピクチャやBピクチャの場合には、予測方式制御部109は、エントロピー符号化部105で行う符号化により生成されるビット数と減算部102の原画像との残差から、例えば従来の手法(例えば、H.264 Reference Software JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, 2008)を用いてラグランジュコストを算出し、インター予測方式とイントラ予測方式のいずれかを選択する。 Returning to FIG. 2, the prediction scheme control unit 109 receives the picture type of the input image, the inter prediction image block signal input from the inter prediction unit 120, its inter prediction encoding information IPE, and the intra prediction unit 121. A prediction scheme is determined based on the intra-predicted image block signal and the intra-encoded information, and information on the prediction scheme is output to the selection unit 110 and the entropy encoding unit 105. The prediction method control unit 109 monitors the picture type of the input image, and selects an intra prediction method when the input image is an I picture. In the case of a P picture or a B picture, the prediction scheme control unit 109 uses, for example, a conventional technique (from the residual between the number of bits generated by the encoding performed by the entropy encoding unit 105 and the original image of the subtraction unit 102 ( For example, the Lagrange cost is calculated using H.264 Reference Software JM ver. 13.2 Encoder, http://iphome.hhi.de/suehring/tml/, ラ 2008), and either inter prediction method or intra prediction method is selected. select.
 予測方式制御部109は、インター予測符号化情報IPEもしくはイントラ予測符号化情報TPEのうち、前記選択した予測方式に対応する符号化情報に予測方式を特定できる情報を付加し予測符号化情報として、エントロピー符号化部105に出力する。
 選択部110は、予測方式制御部109より入力される予測方式の情報に従って、インター予測部120より入力されるインター予測画像ブロック信号、あるいはイントラ予測部121より入力されるイントラ予測画像ブロック信号を選択して、減算部102及び加算部108に予測画像ブロック信号を出力する。選択部110は、予測方式制御部109より入力される予測方式がインター予測である場合には、インター予測部120より入力されるインター予測画像ブロック信号を選択して出力し、予測方式制御部109より入力される予測方式がイントラ予測である場合は、イントラ予測部121より入力されるイントラ予測画像ブロック信号を選択して出力する。
The prediction method control unit 109 adds information that can specify the prediction method to the coding information corresponding to the selected prediction method, among the inter prediction coding information IPE or the intra prediction coding information TPE, as prediction coding information. The data is output to the entropy encoding unit 105.
The selection unit 110 selects the inter prediction image block signal input from the inter prediction unit 120 or the intra prediction image block signal input from the intra prediction unit 121 according to the prediction method information input from the prediction method control unit 109. Then, the predicted image block signal is output to the subtracting unit 102 and the adding unit 108. When the prediction method input from the prediction method control unit 109 is inter prediction, the selection unit 110 selects and outputs the inter prediction image block signal input from the inter prediction unit 120, and the prediction method control unit 109. When the input prediction method is intra prediction, the intra prediction image block signal input from the intra prediction unit 121 is selected and output.
 エントロピー符号化部105は、量子化部104より入力される差分画像符号と、予測方式制御部109より入力される予測符号化情報をパッキング(packing;詰込)し、例えば可変長符号化(エントロピー符号化)を用いて符号化し、情報量がより圧縮された符号化データE1を生成する。エントロピー符号化105は、生成した符号化データE1を画像符号化装置100の外部(例えば、通信網500を介して、画像復号装置800)に出力する。 The entropy encoding unit 105 packs the differential image code input from the quantization unit 104 and the prediction encoding information input from the prediction scheme control unit 109, for example, variable length encoding (entropy). The encoded data E1 is generated by using the encoding) to compress the amount of information. The entropy encoding 105 outputs the generated encoded data E1 to the outside of the image encoding device 100 (for example, the image decoding device 800 via the communication network 500).
 <デプス利用予測モード実施部 処理概要>
 次に、デプス利用予測モード実施部201によって、予測画像ブロックを生成する方法について説明する。
 画面内予測は、前述のように周囲の画素を用いて処理対象ブロックの画素の予測を行っている。具体的には、画面内予測は、処理の完了した周辺の画素を予測の方向に順次コピーしていくことで予測画像ブロック信号を作成する。
<Depth usage prediction mode implementation process overview>
Next, a method for generating a predicted image block by the depth use prediction mode execution unit 201 will be described.
In-screen prediction is performed by predicting pixels of a processing target block using surrounding pixels as described above. Specifically, in the intra prediction, a predicted image block signal is created by sequentially copying neighboring pixels that have been processed in the prediction direction.
 従って、テクスチャの特性が大きく変わらない平坦な領域は、この画面内予測によって処理対象ブロックの画素を精度よく予測することが可能で、処理対象ブロックの画素と予測ブロックの画素の差(残差)を小さくすることができ、その結果、符号量を小さく(あるいは、復号時の誤差を小さく)することができる。 Therefore, in a flat region where the texture characteristics do not change significantly, the pixel of the processing target block can be accurately predicted by this intra prediction, and the difference (residual) between the pixel of the processing target block and the pixel of the prediction block As a result, the code amount can be reduced (or the error during decoding can be reduced).
 一方、異なる被写体は、一般的にテクスチャの特性が大きく異なる。しかしながら、画面内予測は、異なる被写体間をまたがる領域においても、一様に処理がなされるために、予測精度が落ち、符号量を増加させる問題がある。
 ここで、異なる被写体は、同一距離でお互いに接している場合を除いて、異なるデプス値を持っていることが多い。つまり、デプス値の違いを利用(例えば、デプス値のエッジを検出)することで、異なる被写体を分離する(異なる被写体の境界を検知)ことが可能である。
 この情報を用いて、上記画面内予測の画素の予測を制御することで予測精度を向上させることができる。具体的には、以下のとおりである。
On the other hand, different subjects generally have significantly different texture characteristics. However, in-screen prediction is performed uniformly even in regions that cross between different subjects, so that there is a problem that prediction accuracy is reduced and code amount is increased.
Here, different subjects often have different depth values unless they are in contact with each other at the same distance. That is, by using the difference in depth value (for example, detecting an edge of the depth value), it is possible to separate different subjects (detect boundaries between different subjects).
By using this information, the prediction accuracy can be improved by controlling the prediction of pixels in the intra prediction. Specifically, it is as follows.
 本実施形態におけるデプス利用予測モード実施部201が実施するデプスを利用した画面内予測の予測方向は、図6に示す垂直方向の予測(予測モード0)と水平方向の予測(予測モード1)である。但し、以下に説明する処理を適用できる場合(但し、図6の予測モード2を除く)には、その他の予測の方向を用いることも可能である。また、図4におけるサブブロック単位の予測方式(但し、予測モード2を除く)に対しても、以下に説明する処理を適用することも可能である。つまり、本実施形態のように、従来の予測モードを残したまま新たな予測モードとして追加するようにしてもよいし、デプス利用予測モード実施部201による予測方式を従来の方式に置き換えて用いることで、モード数を増やさないようにしてもよい。以下は、デプス利用予測モードを新たに追加する例について記載する。 The prediction directions of the intra prediction using the depth performed by the depth use prediction mode execution unit 201 in the present embodiment are the prediction in the vertical direction (prediction mode 0) and the prediction in the horizontal direction (prediction mode 1) shown in FIG. is there. However, when the processing described below can be applied (however, except for the prediction mode 2 in FIG. 6), other prediction directions can be used. Also, the processing described below can be applied to the prediction method in units of sub-blocks in FIG. 4 (except for prediction mode 2). That is, as in the present embodiment, a new prediction mode may be added while leaving the conventional prediction mode, or the prediction method performed by the depth use prediction mode execution unit 201 may be replaced with the conventional method. Thus, the number of modes may not be increased. The following describes an example in which a depth usage prediction mode is newly added.
 図7、図8は、デプス利用予測モード実施部201の処理概念を説明するための図である。図7、図8において、円で示した図形は、処理の完了した画素を示しており、予測画素ブロックを生成する際に参照することができる。また、四角で示した図形は、処理対象画素を示しており、周辺の参照できる画素を用いて予測される対象である。さらに、矢印は、予測の方向を示しており、参照できる画素を矢印の方向に順次予測(具体的には、単純にコピー)していく。すなわち図7の予測モードでは、垂直方向に画素値をコピーし、図8の予測モードでは、水平方向に画素値をコピーする。図7、図8において、太い破線は、被写体の境界を示す。 7 and 8 are diagrams for explaining the processing concept of the depth use prediction mode execution unit 201. FIG. 7 and 8, the graphic indicated by a circle indicates a pixel for which processing has been completed, and can be referred to when a predicted pixel block is generated. Further, a graphic indicated by a square indicates a pixel to be processed, and is a target that is predicted using pixels that can be referred to in the vicinity. Furthermore, the arrow indicates the direction of prediction, and pixels that can be referred to are sequentially predicted (specifically, simply copied) in the direction of the arrow. That is, in the prediction mode of FIG. 7, the pixel value is copied in the vertical direction, and in the prediction mode of FIG. 8, the pixel value is copied in the horizontal direction. 7 and 8, a thick broken line indicates the boundary of the subject.
 図9は、本実施形態に係るデプス利用予測モード実施部201の構成を示す概略ブロック図である。図9に示すように、デプス利用予測モード実施部201は、境界制御予測画像生成部300、境界予測制御部301、及び被写体境界検出部302を含んで構成される。 FIG. 9 is a schematic block diagram illustrating a configuration of the depth use prediction mode execution unit 201 according to the present embodiment. As illustrated in FIG. 9, the depth use prediction mode execution unit 201 includes a boundary control prediction image generation unit 300, a boundary prediction control unit 301, and a subject boundary detection unit 302.
 被写体境界検出部302は、処理対象の画像ブロック信号Bに対応する画素のデプス値を表すデプスブロック信号を、デプスマップ復号部117より取得し、デプスのエッジを検出する。デプスのエッジの検出は、デプスマップの隣接画素の差分を閾値処理することによって行う。デプスのエッジが水平方向に存在するか否かの判定は、式(3)に示すように、垂直方向に隣接する画素の差の絶対値が閾値TVよりも大きいか否かにより行う。
同様に、デプスのエッジが垂直方向に存在するか否かの判定は、水平方向に隣接する画素の差の絶対値が閾値THよりも大きいか否かにより行う。
The subject boundary detection unit 302 acquires a depth block signal representing a depth value of a pixel corresponding to the image block signal B to be processed from the depth map decoding unit 117, and detects a depth edge. Depth edge detection is performed by thresholding the difference between adjacent pixels in the depth map. Whether or not the depth edge exists in the horizontal direction is determined by whether or not the absolute value of the difference between pixels adjacent in the vertical direction is larger than the threshold value TV, as shown in Expression (3).
Similarly, whether or not the depth edge exists in the vertical direction is determined by whether or not the absolute value of the difference between pixels adjacent in the horizontal direction is larger than the threshold value TH.
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
 ここで、D(i,j)は、画素位置(i,j)におけるデプスマップ値を表す。TV、THは、それぞれ水平方向と垂直方向にエッジが存在するか否かを判定する際に用いる閾値である。具体的には、閾値は、例えば10である。 Here, D (i, j) represents a depth map value at the pixel position (i, j). TV and TH are threshold values used when determining whether or not edges exist in the horizontal direction and the vertical direction, respectively. Specifically, the threshold is 10, for example.
 上記方法によるデプスエッジの検出結果の一例として、図7、図8の太点線の様にデプスエッジが検出された場合について説明する。図7の垂直方向予測モードと、図8の水平方向予測モードにおいて、デプスエッジの位置は、同じである。図7、図8の例では、予測対象ブロックを左右に分割するように、デプスのエッジが存在している。この場合、処理対象ブロック内の左側と右側において異なる被写体が映っている可能性が高い。従来の垂直方向の予測、あるいは、水平方向の予測では、このエッジを跨るところ、及び予測方向におけるその先の画素において、予測精度が著しく低下する。 As an example of the depth edge detection result obtained by the above method, a case where a depth edge is detected as shown by a thick dotted line in FIGS. 7 and 8 will be described. In the vertical direction prediction mode of FIG. 7 and the horizontal direction prediction mode of FIG. 8, the position of the depth edge is the same. In the examples of FIGS. 7 and 8, there are depth edges so that the prediction target block is divided into left and right. In this case, there is a high possibility that different subjects appear on the left and right sides in the processing target block. In the prediction in the conventional vertical direction or the prediction in the horizontal direction, the prediction accuracy is remarkably lowered at the pixel straddling this edge and the subsequent pixels in the prediction direction.
 境界予測制御部301は、被写体境界検出部302から入力される水平方向と垂直方向の被写体の境界情報(デプスエッジ)を用いて、境界制御予測画像生成部300で行う予測を制御する。具体的には、境界予測制御部301は、予測方向に対して垂直なデプスエッジが存在する場合、予測方向に隣接する画素からコピーすることを抑制する制御を行う。予測方向に画素をコピーすることを抑制する制御とは、例えば、境界制御予測画像生成部300における処理を、以下の様に制御することである。 The boundary prediction control unit 301 controls the prediction performed by the boundary control predicted image generation unit 300 using the boundary information (depth edge) of the subject in the horizontal direction and the vertical direction input from the subject boundary detection unit 302. Specifically, when there is a depth edge perpendicular to the prediction direction, the boundary prediction control unit 301 performs control to suppress copying from pixels adjacent to the prediction direction. The control for suppressing copying of pixels in the prediction direction is, for example, controlling the processing in the boundary control predicted image generation unit 300 as follows.
 境界制御予測画像生成部300は、加算部108より参照画像ブロック信号RBを取得し、以下のように予測画像ブロック信号を生成する。境界制御予測画像生成部300の予測モードには、本実施形態では、図7に示すように予測方向が垂直である予測モードと、図8に示すように予測方向が水平である予測モードとがある(2種類の予測画像ブロック信号を生成する)。境界予測制御部301は、処理対象画素と、その予測方向の一つ前の画素との間に被写体の境界が存在しないとき、すなわち予測方向に垂直な方向にデプスのエッジが存在しないときは、境界制御予測画像生成部300に従来の予測方法と同様に処理させる。すなわち、境界予測制御部301は、予測方向における処理対象画素の一つ前の画素値をコピーするように、境界制御予測画像生成部300を制御する。 The boundary control predicted image generation unit 300 acquires the reference image block signal RB from the addition unit 108 and generates a predicted image block signal as follows. In this embodiment, the prediction mode of the boundary control prediction image generation unit 300 includes a prediction mode in which the prediction direction is vertical as shown in FIG. 7 and a prediction mode in which the prediction direction is horizontal as shown in FIG. Yes (two types of predicted image block signals are generated). When the boundary of the subject does not exist between the processing target pixel and the pixel immediately before the prediction direction, that is, when there is no depth edge in the direction perpendicular to the prediction direction, the boundary prediction control unit 301 The boundary control predicted image generation unit 300 is processed in the same manner as in the conventional prediction method. That is, the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 so as to copy the pixel value immediately before the processing target pixel in the prediction direction.
 例えば、図7において、処理対象画素がQv1のとき、予測方向に一つ前の画素であるPv1との間には、デプスエッジが存在しない。このため、境界予測制御部301は、画素Pv1の画素値をコピーして、画素Qv1の画素値とするように、境界制御予測画像生成部300を制御する。水平方向についても同様(図8の画素Ph2から画素Qh2への処理など)である。一方、予測方向に垂直な方向にエッジが存在する場合は、境界予測制御部301は、境界制御予測画像生成部300に対して、以下の処理を実施するように制御する。 For example, in FIG. 7, when the pixel to be processed is Qv1, there is no depth edge between Pv1 which is the previous pixel in the prediction direction. For this reason, the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 so as to copy the pixel value of the pixel Pv1 to the pixel value of the pixel Qv1. The same applies to the horizontal direction (processing from the pixel Ph2 to the pixel Qh2 in FIG. 8). On the other hand, when an edge exists in a direction perpendicular to the prediction direction, the boundary prediction control unit 301 controls the boundary control prediction image generation unit 300 to perform the following processing.
 境界制御予測画像生成部300は、予測方向に垂直なデプスエッジが存在する場合には、以下の式によって、予測画素を生成する。 The boundary control prediction image generation unit 300 generates a prediction pixel according to the following formula when there is a depth edge perpendicular to the prediction direction.
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
 式(5)は、垂直方向の予測モードにおいて、水平方向にデプスのエッジが存在する場合に、予測画素を生成するための式である。式(6)は、水平方向の予測モードにおいて、垂直方向にデプスのエッジが存在する場合に、予測画素を生成するための式である。基本的な処理は、水平方向と垂直方向で同じであるため、以下は水平方向の式(5)について説明する。
 式(5)において、左辺のG[x]は、画素xの予測画素値である。右辺のpre={Pv1,…,Pv16}が付されたargmin{}は、preを、画素Pv1~Pv16のいずれかとしたときに、argmin{}内の評価式が最小となるpreを示す関数である。従って、評価式を最小にする画素値が、pre={Pv1,…,Pv16}(処理を行っている画素の一つ前のラインに存在する画素)の中から選択され、左辺の処理対象画素の画素値としてコピーされる。
Expression (5) is an expression for generating a prediction pixel when a depth edge exists in the horizontal direction in the prediction mode in the vertical direction. Expression (6) is an expression for generating a prediction pixel when there is a depth edge in the vertical direction in the horizontal prediction mode. Since the basic processing is the same in the horizontal direction and the vertical direction, Expression (5) in the horizontal direction will be described below.
In Expression (5), G [x] on the left side is a predicted pixel value of the pixel x. Argmin {} with pre = {Pv1,..., Pv16} on the right side is a function indicating pre that the evaluation expression in argmin {} is minimum when pre is any of the pixels Pv1 to Pv16. is there. Therefore, the pixel value that minimizes the evaluation formula is selected from pre = {Pv1,..., Pv16} (pixels existing in the line immediately before the pixel being processed), and the processing target pixel on the left side Is copied as a pixel value.
 評価式は、2つの項から形成されている。第1項(|D(Qvi)-D(pre)|)は、処理対象画素のデプス値と、前記preの中の各画素に対応したデプス値の差の絶対値を表す。第2項(Dis(Qvi,pre))は、処理対象画素の位置と、前記preの中の各画素の画素位置の間の距離を表す。それぞれの項の意味は、第1項は、処理対象画素に映っている被写体と同一と考えられる被写体を映している画素をできるだけ参照できるように、デプス値の近い画素を選択するように制御する項である。すなわち、当該項も、境界予測制御部301による制御と同様に、処理対象画素との間に被写体の境界がある画素を用いることを抑制するようになっている。第2項は、処理対象画素と近い画素をできるだけ選択するようにする項である。また、各項に積算されている、αとβは、第1項と第2項との間の重みづけを変えるための定数である。具体的には、例えばαは0.1、βは1.0である。 The evaluation formula is formed from two terms. The first term (| D (Qvi) −D (pre) |) represents the absolute value of the difference between the depth value of the pixel to be processed and the depth value corresponding to each pixel in the pre. The second term (Dis (Qvi, pre)) represents the distance between the position of the pixel to be processed and the pixel position of each pixel in the pre. The meaning of each term is that the first term is controlled so as to select a pixel having a depth value so that a pixel showing the subject considered to be the same as the subject appearing in the processing target pixel can be referred to as much as possible. Term. That is, the term also suppresses the use of a pixel that has a subject boundary between the target pixel and the processing target pixel, as in the control by the boundary prediction control unit 301. The second term is a term for selecting pixels as close as possible to the pixel to be processed. Further, α and β accumulated in each term are constants for changing the weighting between the first term and the second term. Specifically, for example, α is 0.1 and β is 1.0.
 なお、上述の式(5)(6)では、第1項と第2項との和を評価式に用いているが、比を用いるようにしてもよい。また、第1項のみを用いるようにしてもよい。
 また、本実施形態では、処理対象の画素と、予測方向に一つ前の画素との間にデプスエッジがあるときにのみ、上述の式(5)(6)を用いるようにしているが、常に上述の式(5)(6)を用いるようにしてもよい。
In the above formulas (5) and (6), the sum of the first term and the second term is used in the evaluation formula, but a ratio may be used. Further, only the first term may be used.
In the present embodiment, the above formulas (5) and (6) are used only when there is a depth edge between the pixel to be processed and the previous pixel in the prediction direction. The above formulas (5) and (6) may be used.
 このように、境界予測制御部301は、処理対象の画素と、予測方向(所定の方向)に一つ前の画素(隣接する画素)との間にデプスエッジ(被写体の境界)があるときは、境界制御予測画像生成部300が上述の式(5)(6)を用いるようにすることで、予測方向に一つ前の画素の画素値を用いることを抑制している。
 また、上述の式(5)(6)は、第1項が、処理対象画素と、予測方向に一つ前の列(または行)の画素(周辺画素)とのデプス値の差となっているので、処理対象画素との間に被写体の境界があり、デプス値の差が大きくなっている周辺画素を用いることを抑制することができる。
In this way, the boundary prediction control unit 301, when there is a depth edge (subject boundary) between the pixel to be processed and the previous pixel (adjacent pixel) in the prediction direction (predetermined direction), By using the above-described equations (5) and (6) by the boundary control predicted image generation unit 300, the use of the pixel value of the previous pixel in the prediction direction is suppressed.
In the above formulas (5) and (6), the first term is the difference in depth value between the processing target pixel and the pixel (peripheral pixel) in the previous column (or row) in the prediction direction. Therefore, it is possible to suppress the use of peripheral pixels that have a subject boundary between the processing target pixels and have a large difference in depth value.
 境界制御予測画像生成部300は、水平方向と垂直方向に予測した予測画像ブロックを生成する。境界制御予測画像生成部300は、画像入力部101より入力された画像ブロックと前記2種類の予測モード各々で予測した予測画像ブロックの相関性を、例えばSAD値を用いて判定する。境界制御予測画像生成部300は、この判定の結果、より相関性の高い(類似した)予測画像ブロックを選択して、予測モード選択部202に出力する。
境界制御予測画像生成部300は、選択した予測画像ブロックの予測モードを示す予測符号化情報も、予測モード選択部202に出力する。
 このように、被写体までの距離を示すデプスマップの境界(被写体境界)において、連続するような画面内画素予測を抑制する制御を行うので、予測精度を向上させることが可能である。
The boundary control predicted image generation unit 300 generates a predicted image block predicted in the horizontal direction and the vertical direction. The boundary control predicted image generation unit 300 determines the correlation between the image block input from the image input unit 101 and the predicted image block predicted in each of the two types of prediction modes using, for example, the SAD value. As a result of this determination, the boundary control prediction image generation unit 300 selects a prediction image block having a higher correlation (similar) and outputs the prediction image block to the prediction mode selection unit 202.
The boundary control prediction image generation unit 300 also outputs prediction encoding information indicating the prediction mode of the selected prediction image block to the prediction mode selection unit 202.
In this way, since control is performed to suppress continuous pixel prediction within the depth map boundary (subject boundary) indicating the distance to the subject, the prediction accuracy can be improved.
 <画像符号化装置100フローチャート>
 次に、本実施形態に係る画像符号化装置100が行う画像符号化処理について説明する。図10は、本実施形態に係る画像符号化装置100が行う画像符号化処理を示すフローチャートである。
(ステップS201)画像符号化装置100は、外部からフレームごとの画像とそれに対応するデプスマップを取得する。その後、ステップS202に進む。
(ステップS202)画像入力部101は、画像符号化装置100の外部から取得したフレームごとの入力画像信号を予め定めた大きさ(例えば、垂直方向16画素×水平方向16画素)のブロックに分割して、減算部102とインター予測部120及びイントラ予測部121に出力する。また、デプス入力部118は、画像入力部101に入力された画像と同期されたデプスマップを、画像入力部101で実施された画像の分割と同様に分割して、デプスマップ符号化部116に出力する。
<Image Encoding Device 100 Flowchart>
Next, an image encoding process performed by the image encoding device 100 according to the present embodiment will be described. FIG. 10 is a flowchart showing an image encoding process performed by the image encoding apparatus 100 according to the present embodiment.
(Step S201) The image encoding apparatus 100 acquires an image for each frame and a depth map corresponding to the image from the outside. Thereafter, the process proceeds to step S202.
(Step S202) The image input unit 101 divides an input image signal for each frame acquired from the outside of the image encoding device 100 into blocks of a predetermined size (for example, 16 pixels in the vertical direction × 16 pixels in the horizontal direction). Output to the subtraction unit 102, the inter prediction unit 120, and the intra prediction unit 121. Further, the depth input unit 118 divides the depth map synchronized with the image input to the image input unit 101 in the same manner as the image division performed by the image input unit 101, and sends the depth map to the depth map encoding unit 116. Output.
 画像符号化装置100は、ステップS203-ステップS211の処理をフレーム内の画像ブロック毎に繰り返す。
(ステップS203)デプスマップ符号化部116は、デプス入力部118から入力されるデプスマップを符号化して、データ量がより圧縮されたデプスマップ符号化データを、イントラ予測部121と画像符号化装置100の外部に(例えば、画像復号装置800)に出力する。その後、ステップS204の処理とステップS205の処理とを並行して行う。
The image coding apparatus 100 repeats the processing from step S203 to step S211 for each image block in the frame.
(Step S203) The depth map encoding unit 116 encodes the depth map input from the depth input unit 118, and converts the depth map encoded data whose data amount is further compressed into the intra prediction unit 121 and the image encoding device. 100 (for example, the image decoding device 800). Thereafter, the process of step S204 and the process of step S205 are performed in parallel.
 (ステップS204)インター予測部120は、画像入力部101から画像ブロック信号を取得し、加算部108より復号された参照画像ブロック信号を取得する。インター予測部120は、取得したこれらの信号を用いて、インター予測を実施する。インター予測部120は、インター予測により生成したインター予測画像ブロック信号を予測方式制御部109と選択部110に、インター予測符号化情報IPEを予測方式制御部109に出力する。
なお、最初の処理において、加算部108の処理が完了していない場合には、加算部108からは、リセットされた画像ブロック(全ての画素値が0の画像ブロック信号)を取得する。インター予測部120の処理が完了すると、ステップS206に進む。
(Step S <b> 204) The inter prediction unit 120 acquires an image block signal from the image input unit 101, and acquires a reference image block signal decoded by the addition unit 108. The inter prediction unit 120 performs inter prediction using these acquired signals. The inter prediction unit 120 outputs the inter prediction image block signal generated by the inter prediction to the prediction method control unit 109 and the selection unit 110, and outputs the inter prediction coding information IPE to the prediction method control unit 109.
In the first process, when the processing of the adding unit 108 is not completed, a reset image block (an image block signal with all pixel values being 0) is acquired from the adding unit 108. If the process of the inter estimation part 120 is completed, it will progress to step S206.
 (ステップS205)イントラ予測部121は、画像入力部101から画像ブロック信号を取得し、デプスマップ符号化部116からデプスマップ符号化データを取得し、加算部108より復号された参照画像ブロック信号を取得する。イントラ予測部121は、これらの取得した信号を用いて、イントラ予測を実施する。イントラ予測部121は、イントラ予測により生成したイントラ予測画像ブロック信号を予測方式制御部109と選択部110に、イントラ予測符号化情報TPEを予測方式制御部109に出力する。なお、最初の処理において、加算部108の処理が完了していない場合には、リセットされた画像ブロック(全ての画素値が0の画像ブロック)を取得する。イントラ予測部121の処理が完了すると、ステップS206に進む。 (Step S205) The intra prediction unit 121 acquires an image block signal from the image input unit 101, acquires depth map encoded data from the depth map encoding unit 116, and receives the reference image block signal decoded by the addition unit 108. get. The intra prediction unit 121 performs intra prediction using these acquired signals. The intra prediction unit 121 outputs the intra prediction image block signal generated by the intra prediction to the prediction scheme control unit 109 and the selection unit 110, and outputs the intra prediction coding information TPE to the prediction scheme control unit 109. In the first process, when the process of the adding unit 108 is not completed, a reset image block (an image block in which all pixel values are 0) is acquired. If the process of the intra estimation part 121 is completed, it will progress to step S206.
 (ステップS206)予測方式制御部109は、インター予測部120よりインター予測画像ブロック信号とインター予測符号化情報IPEを受け取り、イントラ予測部121よりイントラ予測画像ブロック信号とイントラ予測符号化情報TPEを受取る。予測方式制御部109は、上記ラグランジュコストに基づいて、符号化効率の良い予測モードを選択する。予測方式制御部109は、選択した予測モードを示す情報を選択部110に出力する。予測方式制御部109は、選択した予測モードに対応する予測符号化情報をエントロピー符号化部105に出力する。
 選択部101は、予測方式制御部109から入力される予測モード情報に従って、インター予測部120から入力されるインター予測画像ブロック信号、あるいはイントラ予測部121から入力されるイントラ予測画像ブロック信号を選択して、減算部102と加算部108に出力する。その後、ステップS207に進む。
(Step S206) The prediction scheme control unit 109 receives the inter prediction image block signal and the inter prediction encoding information IPE from the inter prediction unit 120, and receives the intra prediction image block signal and the intra prediction encoding information TPE from the intra prediction unit 121. . The prediction scheme control unit 109 selects a prediction mode with good coding efficiency based on the Lagrangian cost. The prediction method control unit 109 outputs information indicating the selected prediction mode to the selection unit 110. The prediction scheme control unit 109 outputs prediction encoding information corresponding to the selected prediction mode to the entropy encoding unit 105.
The selection unit 101 selects an inter prediction image block signal input from the inter prediction unit 120 or an intra prediction image block signal input from the intra prediction unit 121 according to the prediction mode information input from the prediction method control unit 109. To the subtraction unit 102 and the addition unit 108. Thereafter, the process proceeds to step S207.
 (ステップS207)減算部102は、画像入力部101が出力した画像ブロック信号から選択部110が出力した予測画像ブロック信号を減算し、差分画像ブロック信号を生成する。減算部102は、差分画像ブロック信号を直交変換部103に出力する。その後、ステップS208に進む。 (Step S207) The subtraction unit 102 subtracts the predicted image block signal output from the selection unit 110 from the image block signal output from the image input unit 101 to generate a difference image block signal. The subtraction unit 102 outputs the difference image block signal to the orthogonal transformation unit 103. Thereafter, the process proceeds to step S208.
 (ステップS208)直交変換部103は、減算部102から差分画像ブロック信号を取得し、上記直交変換を実施する。直交変換部103は、直交変換後の信号を量子化部104に出力する。量子化部104は、直交変換部103から入力された信号を、上記量子化処理を実施し、差分画像符号を生成する。量子化部104は、差分画像符号を、エントロピー符号化部105と逆量子化部106に出力する。
 エントロピー符号化部105は、量子化部104から入力される差分画像符号と予測方式制御部109から入力される予測符号化情報をパッキング(packing;詰込)し、可変長符号化(エントロピー符号化)を行い、情報量がより圧縮された符号化データE1を生成する。エントロピー符号化部105は、符号化データE1を画像符号化装置100の外部に(例えば、画像復号装置800)に出力する。その後、ステップS209に進む。
(Step S208) The orthogonal transform unit 103 acquires the difference image block signal from the subtraction unit 102, and performs the orthogonal transform. The orthogonal transform unit 103 outputs the signal after the orthogonal transform to the quantization unit 104. The quantization unit 104 performs the above quantization process on the signal input from the orthogonal transform unit 103 to generate a difference image code. The quantization unit 104 outputs the difference image code to the entropy coding unit 105 and the inverse quantization unit 106.
The entropy encoding unit 105 packs the differential image code input from the quantization unit 104 and the prediction encoding information input from the prediction scheme control unit 109, and performs variable length encoding (entropy encoding). ) To generate encoded data E1 in which the amount of information is further compressed. The entropy encoding unit 105 outputs the encoded data E1 to the outside of the image encoding device 100 (for example, the image decoding device 800). Thereafter, the process proceeds to step S209.
 (ステップS209)逆量子化部106は、量子化部104から差分画像符号EDを取得し、量子化部104で実施した量子化の逆の処理を行う。逆量子化部106は、この処理により生成された信号を逆直交変換部107に出力する。逆直交変換部107は、逆量子化部106から逆量子化された信号を取得し、直交変換部103で実施した直交変換処理の逆直交変換処理を実施し、差分画像(復号差分画像ブロック信号)を復号する。逆直交変換部107は、復号された差分画像ブロック信号を加算部108に出力する。その後、ステップS210に進む。 (Step S209) The inverse quantization unit 106 acquires the difference image code ED from the quantization unit 104, and performs the inverse process of the quantization performed by the quantization unit 104. The inverse quantization unit 106 outputs the signal generated by this processing to the inverse orthogonal transform unit 107. The inverse orthogonal transform unit 107 acquires the inversely quantized signal from the inverse quantization unit 106, performs the inverse orthogonal transform process of the orthogonal transform process performed by the orthogonal transform unit 103, and obtains a difference image (decoded difference image block signal). ). The inverse orthogonal transform unit 107 outputs the decoded difference image block signal to the addition unit 108. Thereafter, the process proceeds to step S210.
 (ステップS210)加算部108は、逆直交変換部107が出力した復号された差分画像ブロック信号に、選択部110が出力した予測画像ブロック信号を加算して、入力画像を復号する(参照画像ブロック信号)。加算部108は、参照画像ブロック信号を、インター予測部120とイントラ予測部121に出力する。その後、ステップS211に進む。
(ステップS211)画像符号化装置100が、フレーム内の全ブロックについてステップS203-S210の処理が完了していない場合、処理対象となるブロックを変更してステップS202に戻る。全ての処理が完了している場合、終了する。
(Step S210) The addition unit 108 adds the predicted image block signal output from the selection unit 110 to the decoded difference image block signal output from the inverse orthogonal transform unit 107, and decodes the input image (reference image block). signal). The adding unit 108 outputs the reference image block signal to the inter prediction unit 120 and the intra prediction unit 121. Then, it progresses to step S211.
(Step S211) When the image coding apparatus 100 has not completed the processes of Steps S203 to S210 for all the blocks in the frame, the block to be processed is changed and the process returns to Step S202. When all the processes are completed, the process ends.
 図11は、インター予測部120の処理を説明するフローチャートである。
(ステップS301)デブロッキング・フィルタ部111は、インター予測部120の外部である加算部108から参照画像ブロック信号を取得し、上記FIRフィルタ処理を実施する。デブロッキング・フィルタ部111は、フィルタ処理後の補正ブロック信号をフレームメモリ112に出力する。その後、ステップS302に進む。
FIG. 11 is a flowchart for explaining the processing of the inter prediction unit 120.
(Step S <b> 301) The deblocking filter unit 111 acquires a reference image block signal from the addition unit 108 that is outside the inter prediction unit 120, and performs the FIR filter process. The deblocking filter unit 111 outputs the corrected block signal after the filtering process to the frame memory 112. Thereafter, the process proceeds to step S302.
 (ステップS302)フレームメモリ112は、デブロッキング・フィルタ部111の補正ブロック信号を取得し、フレーム番号を同定できる情報と共に画像の一部として補正ブロック信号を保持しておく。その後、ステップS303に進む。
(ステップS303)動きベクトル検出部114は、画像入力部101から画像ブロック信号を受取ると、画像入力部101が出力した画像ブロックに類似するブロックを、フレームメモリ112に蓄積された画像より探し出し(ブロックマッチング)、探し出したブロックを表すベクトル情報(動きベクトル)を生成する。動きベクトル検出部114は、検出したベクトル情報を含めた符号化のために必要な情報を動き補償部113に出力する。その後、ステップS304に進む。
(Step S302) The frame memory 112 acquires the correction block signal of the deblocking filter unit 111, and holds the correction block signal as a part of the image together with information that can identify the frame number. Thereafter, the process proceeds to step S303.
(Step S303) Upon receiving the image block signal from the image input unit 101, the motion vector detection unit 114 searches the image stored in the frame memory 112 for a block similar to the image block output by the image input unit 101 (block). Matching) and generating vector information (motion vector) representing the found block. The motion vector detection unit 114 outputs information necessary for encoding including the detected vector information to the motion compensation unit 113. Thereafter, the process proceeds to step S304.
 (ステップS304)動き補償部113は、動きベクトル検出114から符号化のために必要な情報を取得し、該当する予測ブロックをフレームメモリより抽出する。動き補償部113は、フレームメモリより抽出した予測画像ブロック信号をインター予測画像ブロック信号として予測方式制御部109と選択部110に出力する。同時に、動き補償部113は、動きベクトル検出部114より取得した予測に必要な情報を予測方式制御部109に出力する。その後、インター予測を終了する。 (Step S304) The motion compensation unit 113 acquires information necessary for encoding from the motion vector detection 114, and extracts a corresponding prediction block from the frame memory. The motion compensation unit 113 outputs the prediction image block signal extracted from the frame memory to the prediction method control unit 109 and the selection unit 110 as an inter prediction image block signal. At the same time, the motion compensation unit 113 outputs information necessary for prediction acquired from the motion vector detection unit 114 to the prediction method control unit 109. Thereafter, the inter prediction is terminated.
 図12は、イントラ予測部121の処理を説明するフローチャートである。
(ステップS401)デプスマップ復号部117は、デプスマップ符号化部116からデプスマップ符号化データE2を取得し、例えば可変長復号を用いて情報量がより多いデプスマップを復号する。デプスマップ復号部117は、復号したデプスマップ(デプスブロック復号信号)を奥行き情報利用イントラ予測部115に出力する。その後、ステップS402に進む。
FIG. 12 is a flowchart for explaining processing of the intra prediction unit 121.
(Step S401) The depth map decoding unit 117 acquires depth map encoded data E2 from the depth map encoding unit 116, and decodes a depth map having a larger amount of information by using, for example, variable length decoding. The depth map decoding unit 117 outputs the decoded depth map (depth block decoded signal) to the depth information use intra prediction unit 115. Thereafter, the process proceeds to step S402.
 (ステップS402)第1予測モード実施部200-1~第n予測モード実施部200-nは、加算部108から取得した参照画像ブロック信号から各予測モード(予測画像ブロックの生成方式)による処理に応じて第1~第n予測画像ブロック信号をそれぞれ生成する。第1予測モード実施部200-1~第n予測モード実施部200-nは、生成した第1~第n予測画像ブロック信号を予測モード選択部202に出力する。
 デプス利用予測モード実施部201は、加算部108から取得した参照画像ブロック信号とデプスマップ復号部117から取得したデプスブロック復号信号より、デプスを利用した予測画像ブロック信号を生成し、予測モード選択部202に出力する。その後、ステップS403に進む。なお、デプス利用予測モード実施部201が行う予測画像生成処理については前述のとおりである。
(Step S402) The first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n perform processing in each prediction mode (prediction image block generation method) from the reference image block signal acquired from the addition unit 108. In response, first to n-th predicted image block signals are generated. The first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n output the generated first to n-th prediction image block signals to the prediction mode selection unit 202.
The depth use prediction mode execution unit 201 generates a prediction image block signal using depth from the reference image block signal acquired from the addition unit 108 and the depth block decoded signal acquired from the depth map decoding unit 117, and a prediction mode selection unit To 202. Thereafter, the process proceeds to step S403. Note that the predicted image generation processing performed by the depth use prediction mode execution unit 201 is as described above.
 (ステップS403)予測モード選択部202は、第1予測モード実施部200-1から第n予測モード実施部200-n、およびデプス利用予測モード実施部201から予測画像ブロック信号と予測に必要な情報を入力する。
 予測モード選択部202は、前記入力した予測画像ブロック信号の中から(デプス利用予測モード実施部から入力される予測画像ブロック信号も含む)、上記方法により符号化の効率の良い予測モードを選択し、対応する予測モード情報を生成する。
 予測モード選択部202は、選択した予測画像ブロック信号(以下、イントラ予測画像ブロック信号)を選択部110と予測方式制御部109へ、また前記予測モード情報(以下、イントラ予測符号化情報TPE)を予測方式制御部109へ出力する。その後、イントラ予測を終了する。
(Step S403) The prediction mode selection unit 202 receives the prediction image block signal from the first prediction mode execution unit 200-1 to the n-th prediction mode execution unit 200-n and the depth-use prediction mode execution unit 201 and information necessary for prediction. Enter.
The prediction mode selection unit 202 selects a prediction mode with high coding efficiency by the above method from the input prediction image block signals (including the prediction image block signal input from the depth-based prediction mode execution unit). , Corresponding prediction mode information is generated.
The prediction mode selection unit 202 sends the selected prediction image block signal (hereinafter referred to as an intra prediction image block signal) to the selection unit 110 and the prediction scheme control unit 109, and the prediction mode information (hereinafter referred to as intra prediction coding information TPE). Output to the prediction method control unit 109. Then, intra prediction is complete | finished.
 次に、本実施形態に係る画像復号装置800について説明する。図13は、本実施形態に係る画像復号装置800の構成を示す概略ブロック図である。画像復号装置800は、符号化データ入力部813、エントロピー復号部801、逆量子化部802、逆直交変換部803、加算部804、予測方式制御部805、選択部806、デブロッキング・フィルタ部807、フレームメモリ808、動き補償部809、奥行き情報利用イントラ予測部810、デプスマップ復号部811、画像出力部812、及びデプスマップ符号化データ入力部814を含んで構成される。なお、デブロッキング・フィルタ部807と、フレームメモリ808と、動き補償部809とでインター処理部820を構成する。また、奥行き情報利用イントラ予測部810と、デプスマップ復号部811とでイントラ処理部821を構成する。 Next, the image decoding device 800 according to this embodiment will be described. FIG. 13 is a schematic block diagram showing the configuration of the image decoding device 800 according to this embodiment. The image decoding apparatus 800 includes an encoded data input unit 813, an entropy decoding unit 801, an inverse quantization unit 802, an inverse orthogonal transform unit 803, an addition unit 804, a prediction scheme control unit 805, a selection unit 806, and a deblocking filter unit 807. A frame memory 808, a motion compensation unit 809, a depth information use intra prediction unit 810, a depth map decoding unit 811, an image output unit 812, and a depth map encoded data input unit 814. The deblocking filter unit 807, the frame memory 808, and the motion compensation unit 809 constitute an inter processing unit 820. Also, the depth information utilization intra prediction unit 810 and the depth map decoding unit 811 constitute an intra processing unit 821.
 符号化データ入力部813は、外部(例えば、画像符号化装置100)から取得した符号化データE1を、処理ブロック単位に分割して、エントロピー復号部801に出力する。符号化データ入力部813は、ブロック位置を順次変えながら、フレーム内のすべてのブロックが完了し、そして取得した符号データが終了するまで繰り返して出力する。 The encoded data input unit 813 divides the encoded data E1 acquired from the outside (for example, the image encoding device 100) into processing block units and outputs the result to the entropy decoding unit 801. The encoded data input unit 813 repeatedly outputs the blocks until the blocks are sequentially changed until all the blocks in the frame are completed and the acquired encoded data is completed.
 エントロピー復号部801は、符号化データ入力部813から取得した処理単位に分割された符号化データを、エントロピー符号化部105が行った符号化方法(例えば、可変長符号化)と逆の処理(例えば、可変長復号)であるエントロピー復号して、差分画像ブロック符号と予測符号化情報PEを生成する。エントロピー復号部801は、差分画像ブロック符号を逆量子化部802に、予測符号化情報PEを予測方式制御部805に出力する。 The entropy decoding unit 801 performs processing reverse to the encoding method (for example, variable-length encoding) performed by the entropy encoding unit 105 on the encoded data divided into processing units acquired from the encoded data input unit 813 ( For example, entropy decoding that is variable length decoding) is performed to generate a difference image block code and predictive coding information PE. The entropy decoding unit 801 outputs the difference image block code to the inverse quantization unit 802 and the prediction coding information PE to the prediction scheme control unit 805.
 逆量子化部802は、エントロピー復号部801から入力された差分画像ブロック符号を、逆量子化して復号周波数領域信号を生成し、逆直交変換部803に出力する。
 逆直交変換部803は、逆量子化部802が出力した復号周波数領域信号を、例えば逆DCT変換して空間領域信号である復号差分画像ブロック信号を生成する。逆直交変換部803は、復号周波数領域信号に基づき空間領域信号を生成することができれば、逆DCT変換に限らず、他の方法(例えば、IFFT(高速フーリエ逆変換;Inverse Fast Fourier Transform))を用いてもよい。
 逆直交変換部803は、生成した復号差分画像ブロック信号を加算部804に出力する。
The inverse quantization unit 802 performs inverse quantization on the difference image block code input from the entropy decoding unit 801 to generate a decoded frequency domain signal, and outputs the decoded frequency domain signal to the inverse orthogonal transform unit 803.
The inverse orthogonal transform unit 803 generates a decoded difference image block signal that is a spatial domain signal by, for example, inverse DCT transforming the decoded frequency domain signal output from the inverse quantization unit 802. As long as the inverse orthogonal transform unit 803 can generate a spatial domain signal based on the decoded frequency domain signal, the inverse orthogonal transform unit 803 is not limited to the inverse DCT transform, and other methods (for example, IFFT (Inverse Fast Fourier Transform)) are used. It may be used.
The inverse orthogonal transform unit 803 outputs the generated decoded difference image block signal to the addition unit 804.
 予測方式制御部805は、エントロピー復号部801から入力される予測符号化情報PEの中から、画像符号化装置100で採用されたマクロブロック単位の予測方式PMを取り出す。ここで予測方式PMは、インター予測あるいはイントラ予測である。予測方式制御部805は、抽出した予測方式PMに関する情報を選択部806に出力する。また、予測方式制御部805は、エントロピー復号部801が出力した予測符号化情報PEの中から抽出した予測方式PMに対応する予測符号化情報を取り出し、抽出した予測方式PMに対応する処理部に予測符号化情報を出力する。予測方式制御部805は、予測方式PMがインター予測である場合には、インター処理部820にインター予測符号化情報IPEを出力する。予測方式制御部805は、予測方式PMがイントラ予測である場合には、イントラ処理部821にイントラ予測符号化情報TPEを出力する。 The prediction method control unit 805 extracts the prediction method PM in units of macroblocks adopted by the image coding device 100 from the prediction coding information PE input from the entropy decoding unit 801. Here, the prediction method PM is inter prediction or intra prediction. The prediction method control unit 805 outputs information regarding the extracted prediction method PM to the selection unit 806. Also, the prediction scheme control unit 805 takes out the prediction coding information corresponding to the prediction scheme PM extracted from the prediction coding information PE output from the entropy decoding unit 801, and stores it in the processing unit corresponding to the extracted prediction scheme PM. Predictive coding information is output. The prediction method control unit 805 outputs the inter prediction coding information IPE to the inter processing unit 820 when the prediction method PM is inter prediction. The prediction method control unit 805 outputs the intra prediction encoding information TPE to the intra processing unit 821 when the prediction method PM is intra prediction.
 選択部806は、予測方式制御部805から入力された予測方式PMの情報に従って、後述するインター処理部820が出力するインター予測画像ブロック信号、あるいは後述するイントラ処理部821が出力するイントラ予測画像ブロック信号を選択する。予測方式PMが、インター予測の場合には、インター予測画像ブロック信号を選択する。予測方式PMが、イントラ予測の場合には、イントラ予測画像ブロック信号を選択する。選択部806は、選択した予測画像ブロック信号を、加算部804に出力する。 The selection unit 806, based on the prediction scheme PM information input from the prediction scheme control unit 805, outputs an inter prediction image block signal output by an inter processing unit 820 described later or an intra predicted image block output by an intra processing unit 821 described later. Select a signal. When the prediction method PM is inter prediction, an inter prediction image block signal is selected. When the prediction method PM is intra prediction, an intra prediction image block signal is selected. The selection unit 806 outputs the selected predicted image block signal to the addition unit 804.
 加算部804は、逆直交変換部803が出力した復号差分画像ブロック信号に、選択部806が出力した予測画像ブロック信号を加算し、復号画像ブロック信号DBを生成する。加算部804は、復号した復号画像ブロック信号DBを、インター処理部820とイントラ処理部821、及び画像出力部812に出力する。 The addition unit 804 adds the predicted image block signal output from the selection unit 806 to the decoded difference image block signal output from the inverse orthogonal transform unit 803 to generate a decoded image block signal DB. The adding unit 804 outputs the decoded decoded image block signal DB to the inter processing unit 820, the intra processing unit 821, and the image output unit 812.
 続いて、インター処理部820について説明する。インター処理部820は、デブロッキング・フィルタ部807、フレームメモリ808、及び動き補償部809より構成される。
 デブロッキング・フィルタ部807は、加算部804から入力される復号画像ブロック信号DBに対して、デブロッキング・フィルタ部111で行うFIRフィルタと同じ処理を行い、その処理結果(補正ブロック信号)をフレームメモリ808に出力する。
 フレームメモリ808は、デブロッキング・フィルタ部807から補正ブロック信号を取得し、フレーム番号を同定できる情報と共に画像の一部として補正ブロック信号を保持しておく。
Next, the inter processing unit 820 will be described. The inter processing unit 820 includes a deblocking filter unit 807, a frame memory 808, and a motion compensation unit 809.
The deblocking filter unit 807 performs the same processing as the FIR filter performed by the deblocking filter unit 111 on the decoded image block signal DB input from the addition unit 804, and the processing result (correction block signal) is framed. Output to the memory 808.
The frame memory 808 acquires the correction block signal from the deblocking filter unit 807, and holds the correction block signal as a part of the image together with information that can identify the frame number.
 動き補償部809は、予測方式制御部805よりインター予測符号化情報IPEを取得し、その中から参照画像情報と予測ベクトル情報(動きベクトル)を取り出す。動き補償部809は、取り出した参照画像情報と予測ベクトル情報に基づいて、フレームメモリ808に蓄積されている画像の中から対象の画像ブロック信号(予測画像ブロック信号)を抽出する。動き補償部809は、上記予測ベクトル(動きベクトル)が1つの場合には、該当する画像ブロックをフレームメモリ808より1つ取り出し、選択部806に出力する。上記予測ベクトル(動きベクトル)が2つの場合には、該当する2つの画像ブロックをフレームメモリ808より取り出し、それらを平均化して、選択部806に出力する。インター処理部820(動き補償部809)より選択部806に出力する、この信号は、インター予測画像ブロック信号である。 The motion compensation unit 809 acquires inter prediction coding information IPE from the prediction method control unit 805, and extracts reference image information and prediction vector information (motion vector) from the inter prediction coding information IPE. The motion compensation unit 809 extracts a target image block signal (predicted image block signal) from the images stored in the frame memory 808 based on the extracted reference image information and predicted vector information. When there is one prediction vector (motion vector), the motion compensation unit 809 extracts one corresponding image block from the frame memory 808 and outputs it to the selection unit 806. When there are two prediction vectors (motion vectors), two corresponding image blocks are taken out from the frame memory 808, averaged, and output to the selection unit 806. This signal output from the inter processing unit 820 (motion compensation unit 809) to the selection unit 806 is an inter prediction image block signal.
 続いて、イントラ処理部821について説明する。イントラ処理部821は、奥行き情報利用イントラ予測部810とデプスマップ復号部811より構成される。
 デプスマップ符号化データ入力部814は、外部(例えば、画像符号化装置100)から入力されたデプスマップ符号化データE2を処理ブロック単位に分割し、イントラ処理部821に出力する。
Next, the intra processing unit 821 will be described. The intra processing unit 821 includes a depth information use intra prediction unit 810 and a depth map decoding unit 811.
The depth map encoded data input unit 814 divides the depth map encoded data E2 input from the outside (for example, the image encoding device 100) into processing blocks, and outputs them to the intra processing unit 821.
 デプスマップ復号部811は、デプスマップ符号化データ入力部814が出力したブロック単位のデプスマップ符号化データを、デプスマップ符号化部116が行った符号化方法(例えば、可変長符号化)と逆の処理(例えば、可変長復号)であるエントロピー復号して、デプスブロック復号信号を生成する。デプスマップ復号部811は、デプスブロック復号信号を奥行き情報利用イントラ予測部810に出力する。 The depth map decoding unit 811 reverses the block unit depth map encoded data output from the depth map encoded data input unit 814 to the encoding method (for example, variable length encoding) performed by the depth map encoding unit 116. The depth block decoded signal is generated by performing entropy decoding which is the above process (for example, variable length decoding). The depth map decoding unit 811 outputs the depth block decoded signal to the depth information use intra prediction unit 810.
 図14は、奥行き情報利用イントラ予測部810の構成を示す概略ブロック図である。
奥行き情報利用イントラ予測部810は、第1予測モード実施部900-1、第2予測モード実施部900-2、第n予測モード実施部900-n、デプス利用予測モード実施部901及び予測モード選択部902を含んで構成される。
FIG. 14 is a schematic block diagram illustrating a configuration of the depth information use intra prediction unit 810.
The depth information use intra prediction unit 810 includes a first prediction mode execution unit 900-1, a second prediction mode execution unit 900-2, an nth prediction mode execution unit 900-n, a depth use prediction mode execution unit 901, and a prediction mode selection. A portion 902 is included.
 予測モード選択部902は、予測方式制御部805が出力したイントラ予測符号化情報TPEの中から、画像符号化装置100の予測モード選択部202で作成した予測モードを示すインデックス(予測モード)と、予測に必要な情報をそれぞれ取り出す。なお、ここで、予測に必要な情報を取り出すのは、インデックスが示す予測モードが、予測に必要な情報が存在する予測モード(具体的には、第1予測モード、第2予測モードのサブブロック単位で予測画像を生成するモードとデプス利用予測モード)場合である。予測モード選択部902は、予測に必要な情報を取り出した場合には、該当する予測モード実施部900-1~900-n、901にその情報を出力する。予測モード選択部902は、各予測モード実施部が生成した予測画像ブロック信号の中から、上記インデックス(予測モード)が示す予測モードの予測画像ブロック信号を選択し、イントラ予測画像ブロック信号として、選択部806へ出力する。 The prediction mode selection unit 902 includes an index (prediction mode) indicating the prediction mode created by the prediction mode selection unit 202 of the image encoding device 100 from the intra prediction encoding information TPE output by the prediction method control unit 805, Each piece of information necessary for prediction is extracted. Here, the information necessary for prediction is extracted because the prediction mode indicated by the index is a prediction mode in which information necessary for prediction exists (specifically, the sub-blocks of the first prediction mode and the second prediction mode). This is a case where a prediction image is generated in units and a depth use prediction mode). When the prediction mode selection unit 902 extracts information necessary for prediction, the prediction mode selection unit 902 outputs the information to the corresponding prediction mode execution units 900-1 to 900-n and 901. The prediction mode selection unit 902 selects a prediction image block signal of the prediction mode indicated by the index (prediction mode) from the prediction image block signals generated by each prediction mode execution unit, and selects the prediction image block signal as an intra prediction image block signal. Output to the unit 806.
 第1予測モード実施部900-1、第2予測モード実施部900-2、第n予測モード実施部900-nは、画像符号化装置100の奥行き情報利用イントラ予測部115に備わる第1予測モード実施部200-1、第2予測モード実施部200-2、第n予測モード実施部200-nと同じ処理が実施される。但し、16×16画素をさらに分割したサブブロック単位で予測を行う第1予測モード実施部200-1及び、第2予測モード実施部200-2については、各サブブロックにおける予測モード(予測に必要な情報)を予測モード選択部902より入力し、対応する予測モードを各サブブロック単位で実施する。予測モードについては、図4に示した内容である。 The first prediction mode execution unit 900-1, the second prediction mode execution unit 900-2, and the nth prediction mode execution unit 900-n are provided in the depth information use intra prediction unit 115 of the image encoding device 100. The same processing as that of the execution unit 200-1, the second prediction mode execution unit 200-2, and the nth prediction mode execution unit 200-n is performed. However, for the first prediction mode execution unit 200-1 and the second prediction mode execution unit 200-2 that perform prediction in sub-block units obtained by further dividing 16 × 16 pixels, the prediction mode (necessary for prediction) in each sub-block is used. Information) is input from the prediction mode selection unit 902, and the corresponding prediction mode is executed in units of sub-blocks. About the prediction mode, it is the content shown in FIG.
 デプス利用予測モード実施部901は、予測モード選択部902より予測に必要な情報(具体的には予測の方向を示す情報)を取得し、デプスマップ復号部811よりデプスブロック復号信号を取得する。デプス利用予測モード実施部901は、これら取得した情報と信号を用いて、画像符号化装置100のデプス利用予測モード実施部201が実施したように予測画像ブロック信号を生成する。予測に必要な情報とは、前記デプス利用予測モード実施部201が選択した予測の方向に関する情報である。デプス利用予測モード実施部901の構成は、基本的にデプス利用予測モード実施部201の構成と同じである。ただし、画像符号化装置100の境界制御予測画像生成部300が水平方向の予測ブロックと垂直方向の予測ブロックを入力画像との相関性を基に選択を行う最後の処理を、デプス利用予測モード実施部901の境界制御予測画像生成部300では、予測に必要な情報を用いて選択するところが異なる。以上の処理によって、デプス利用予測モード実施部901は、符号化時のデプス利用予測モード実施部201と同じ予測画像ブロック信号を生成する。 The depth use prediction mode execution unit 901 acquires information necessary for prediction (specifically, information indicating the direction of prediction) from the prediction mode selection unit 902, and acquires a depth block decoded signal from the depth map decoding unit 811. The depth use prediction mode execution unit 901 uses the acquired information and signal to generate a predicted image block signal as performed by the depth use prediction mode execution unit 201 of the image encoding device 100. The information necessary for prediction is information regarding the direction of prediction selected by the depth use prediction mode execution unit 201. The configuration of the depth usage prediction mode execution unit 901 is basically the same as the configuration of the depth usage prediction mode execution unit 201. However, the depth control prediction mode is implemented as the final process in which the boundary control prediction image generation unit 300 of the image encoding device 100 selects the prediction block in the horizontal direction and the prediction block in the vertical direction based on the correlation between the input image and the input image. The boundary control predicted image generation unit 300 of the unit 901 is different in that selection is performed using information necessary for prediction. Through the above processing, the depth use prediction mode execution unit 901 generates the same predicted image block signal as the depth use prediction mode execution unit 201 at the time of encoding.
 次に、本実施形態に係る画像復号装置800が行う画像復号処理について説明する。図15は、本実施形態に係る画像復号装置800が行う画像復号処理を示すフローチャートである。
(ステップS601)画像復号装置800は、通信網500を介して画像符号化装置100から画像の符号化データE1とデプスマップの符号化データE2とからなる符号化データEを取得する。その後、ステップS602に進む。
Next, an image decoding process performed by the image decoding apparatus 800 according to the present embodiment will be described. FIG. 15 is a flowchart showing an image decoding process performed by the image decoding apparatus 800 according to this embodiment.
(Step S601) The image decoding apparatus 800 acquires the encoded data E including the encoded data E1 of the image and the encoded data E2 of the depth map from the image encoding apparatus 100 via the communication network 500. Thereafter, the process proceeds to step S602.
(ステップS602)符号化データ入力部813は、取得した画像の符号化データE1を予め定めた大きさ(例えば、垂直方向16画素×水平方向16画素)に対応する処理ブロックに分割して、エントロピー復号部801に出力する。また、デプスマップ符号化データ入力部814は、符号化データ入力部813に入力された符号化データと同期したデプスマップ符号化データを画像復号装置800の外部から入力し、符号化データ入力部813で実施された分割と同様の処理単位に分割して、イントラ処理部821に出力する。 (Step S602) The encoded data input unit 813 divides the acquired encoded data E1 of the image into processing blocks corresponding to a predetermined size (for example, 16 pixels in the vertical direction × 16 pixels in the horizontal direction) to generate entropy. The data is output to the decoding unit 801. Further, the depth map encoded data input unit 814 inputs depth map encoded data synchronized with the encoded data input to the encoded data input unit 813 from the outside of the image decoding apparatus 800, and the encoded data input unit 813. The data is divided into the same processing units as the division performed in the above and output to the intra processing unit 821.
 画像復号装置800は、ステップS603-ステップS608の処理をフレーム内の画像ブロック毎に繰り返す。
(ステップS603)エントロピー復号部801は、符号化データ入力部813が出力した符号化データをエントロピー復号し、差分画像ブロック符号と予測符号化情報を生成する。エントロピー復号部801は、差分画像ブロック符号を、逆量子化部802に出力し、予測符号化情報を予測方式制御部805に出力する。予測方式制御部805は、エントロピー復号部801から予測符号化情報を取得し、予測方式PMに関する情報とその予測方式PMに対応する予測符号化情報を取り出す。予測方式PMが、インター予測の場合には、予測符号化情報をインター予測符号化情報IPEとして、インター処理部820に出力する。予測方式PMが、イントラ予測の場合には、予測符号化情報をイントラ予測符号化情報TPEとして、イントラ処理部821に出力する。その後、ステップS604とステップS605に進む。なお、ステップS604とステップS605とは、各ブロックについて、並行して処理を実施するようにしてもよいし、予測方式PMに応じて、いずれか片方の処理のみを実施するようにしてもよい。
The image decoding apparatus 800 repeats the processing in steps S603 to S608 for each image block in the frame.
(Step S603) The entropy decoding unit 801 performs entropy decoding on the encoded data output from the encoded data input unit 813, and generates a difference image block code and predictive encoding information. The entropy decoding unit 801 outputs the difference image block code to the inverse quantization unit 802, and outputs the prediction coding information to the prediction scheme control unit 805. The prediction scheme control unit 805 acquires the prediction coding information from the entropy decoding unit 801, and extracts information regarding the prediction scheme PM and the prediction coding information corresponding to the prediction scheme PM. When the prediction method PM is inter prediction, the prediction coding information is output to the inter processing unit 820 as inter prediction coding information IPE. When the prediction method PM is intra prediction, the prediction coding information is output to the intra processing unit 821 as the intra prediction coding information TPE. Then, it progresses to step S604 and step S605. In step S604 and step S605, each block may be processed in parallel, or only one of the processes may be performed in accordance with the prediction method PM.
(ステップS604)インター処理部820は、予測方式制御部805が出力するインター予測符号化情報IPEと、加算部804が出力する復号画像ブロック信号DBとを取得して、インター処理を実施する。インター処理部820は、生成されたインター予測画像ブロック信号を選択部806に出力する。インター処理の内容については後述する。なお、最初の処理において、加算部804の処理が完了していない場合には、リセットされた画像ブロック信号(全ての画素値が0の画像ブロック信号)を入力するものとする。インター処理部の処理が完了すると、ステップS606に進む。 (Step S604) The inter processing unit 820 acquires the inter prediction coding information IPE output from the prediction scheme control unit 805 and the decoded image block signal DB output from the adding unit 804, and performs inter processing. The inter processing unit 820 outputs the generated inter predicted image block signal to the selection unit 806. The contents of the inter processing will be described later. In the first process, when the process of the adding unit 804 is not completed, a reset image block signal (an image block signal in which all pixel values are 0) is input. When the processing of the inter processing unit is completed, the process proceeds to step S606.
(ステップS605)イントラ処理部821は、予測方式制御部805が出力するイントラ予測符号化情報TPEと、加算部804が出力する復号画像ブロック信号DBとを取得して、イントラ予測を実施する。イントラ処理部821は、生成されたイントラ予測画像ブロック信号を選択部806に出力する。イントラ予測の処理については後述する。なお、最初の処理において、加算部804の処理が完了していない場合には、リセットされた画像ブロック信号(全ての画素値が0の画像ブロック信号)を入力するものとする。イントラ予測部821の処理が完了すると、ステップS606に進む。 (Step S605) The intra processing unit 821 acquires the intra prediction encoded information TPE output from the prediction scheme control unit 805 and the decoded image block signal DB output from the adding unit 804, and performs intra prediction. The intra processing unit 821 outputs the generated intra predicted image block signal to the selection unit 806. The intra prediction process will be described later. In the first process, when the process of the adding unit 804 is not completed, a reset image block signal (an image block signal in which all pixel values are 0) is input. If the process of the intra estimation part 821 is completed, it will progress to step S606.
(ステップS606)選択部806は、予測方式制御部805から出力された予測方式PMに関する情報を取得し、インター処理部820が出力したインター予測画像ブロック信号もしくは、イントラ処理部821が出力したイントラ予測画像信号を選択して、加算部804に出力する。その後、ステップS607に進む。 (Step S606) The selection unit 806 acquires information on the prediction method PM output from the prediction method control unit 805, and outputs the inter prediction image block signal output from the inter processing unit 820 or the intra prediction output from the intra processing unit 821. An image signal is selected and output to the adder 804. Thereafter, the process proceeds to step S607.
(ステップS607)逆量子化部802は、エントロピー復号部801から入力した差分画像ブロック符号に対して、画像符号化装置100の量子化部104で実施した量子化の逆の処理を行う。逆量子化部802は、生成された復号周波数領域信号を逆直交変換部803に出力する。逆直交変換部803は、逆量子化部802から逆量子化された復号周波数領域信号を取得し、画像符号化装置100の直交変換部103で実施した直交変換処理の逆直交変換処理を実施し、差分画像(復号差分画像ブロック信号)を復号する。逆直交変換部803は、復号された復号差分画像ブロック信号を加算部804に出力する。加算部804は、逆直交変換部803が出力した復号差分画像ブロック信号に、選択部806が出力した予測画像ブロック信号を加算して、復号画像ブロック信号DBを生成する。加算部804は、復号した復号画像ブロック信号DBを画像出力部812とインター処理部820及びイントラ処理部821に出力する。その後、ステップS608に進む。 (Step S607) The inverse quantization unit 802 performs the inverse process of the quantization performed by the quantization unit 104 of the image coding device 100 on the difference image block code input from the entropy decoding unit 801. The inverse quantization unit 802 outputs the generated decoded frequency domain signal to the inverse orthogonal transform unit 803. The inverse orthogonal transform unit 803 obtains the inversely quantized decoded frequency domain signal from the inverse quantization unit 802, and performs the inverse orthogonal transform process of the orthogonal transform process performed by the orthogonal transform unit 103 of the image encoding device 100. Then, the difference image (decoded difference image block signal) is decoded. The inverse orthogonal transform unit 803 outputs the decoded decoded difference image block signal to the adding unit 804. The adding unit 804 adds the predicted image block signal output from the selection unit 806 to the decoded difference image block signal output from the inverse orthogonal transform unit 803 to generate a decoded image block signal DB. The adding unit 804 outputs the decoded decoded image block signal DB to the image output unit 812, the inter processing unit 820, and the intra processing unit 821. Thereafter, the process proceeds to step S608.
(ステップS608)画像出力部812は、加算部804が出力した復号画像ブロック信号DBを、画像内の対応する位置に配置させ出力画像信号R’を生成する。フレーム内の全ブロックについてステップS603-S607の処理が完了していない場合、処理対象となるブロックを変更してステップS602に戻る。
 画像出力部812は、生成した出力画像信号R’を画像復号装置800の外部(表示装置600)に出力する際には、例えば上述の入力順序で5フレーム(Iピクチャ(I0)、Bピクチャ(B3)、Bピクチャ(B2)、Bピクチャ(B4)及びPピクチャ(P1))毎に出力する。
(Step S608) The image output unit 812 generates the output image signal R ′ by arranging the decoded image block signal DB output by the adding unit 804 at a corresponding position in the image. If the processes in steps S603 to S607 have not been completed for all the blocks in the frame, the block to be processed is changed and the process returns to step S602.
When outputting the generated output image signal R ′ to the outside of the image decoding device 800 (display device 600), the image output unit 812, for example, has 5 frames (I picture (I0), B picture ( B3), B picture (B2), B picture (B4) and P picture (P1)).
 図16は、ステップS604のインター処理を説明するフローチャートである。
(ステップS701)デブロッキング・フィルタ部807は、インター処理部820の外部である加算部804から復号画像ブロック信号DBを取得し、上記符号化時に行ったFIRフィルタ処理を実施する。デブロッキング・フィルタ部807は、フィルタ処理後の補正ブロック信号をフレームメモリ808に出力する。その後、ステップS702に進む。
FIG. 16 is a flowchart for explaining the inter processing in step S604.
(Step S701) The deblocking / filtering unit 807 acquires the decoded image block signal DB from the adding unit 804 that is external to the inter processing unit 820, and performs the FIR filter processing performed at the time of encoding. The deblocking filter unit 807 outputs the corrected corrected block signal to the frame memory 808. Thereafter, the process proceeds to step S702.
(ステップS702)フレームメモリ808は、デブロッキング・フィルタ部807が出力した補正ブロック信号を、フレーム番号を同定できる情報と共に画像の一部として保持しておく。その後、ステップS703に進む。
(ステップS703)動き補償部809は、予測方式制御部805からインター予測符号化情報IPEを取得し、該当する予測ブロック信号をフレームメモリより抽出する。動き補償部809は、フレームメモリより抽出した予測画像ブロック信号をインター予測画像ブロック信号として、選択部806に出力する。その後、インター処理を終了する。
(Step S <b> 702) The frame memory 808 holds the correction block signal output from the deblocking filter unit 807 as part of the image together with information that can identify the frame number. Thereafter, the process proceeds to step S703.
(Step S703) The motion compensation unit 809 acquires the inter prediction coding information IPE from the prediction scheme control unit 805, and extracts a corresponding prediction block signal from the frame memory. The motion compensation unit 809 outputs the prediction image block signal extracted from the frame memory to the selection unit 806 as an inter prediction image block signal. Thereafter, the inter processing is terminated.
 図17は、ステップS605のイントラ処理を説明するフローチャートである。
(ステップS801)デプスマップ復号部811は、デプスマップ符号化データ入力部814から処理単位に分割したデプスマップ符号化データを取得し、例えば可変長復号を用いて情報量がより多いデプスマップを復号する。デプスマップ復号部811は、復号したデプスマップ(デプスブロック復号信号)を奥行き情報利用イントラ予測部810に出力する。その後、ステップS802に進む。
FIG. 17 is a flowchart illustrating the intra processing in step S605.
(Step S801) The depth map decoding unit 811 acquires depth map encoded data divided into processing units from the depth map encoded data input unit 814, and decodes a depth map having a larger amount of information using variable length decoding, for example. To do. The depth map decoding unit 811 outputs the decoded depth map (depth block decoded signal) to the depth information use intra prediction unit 810. Thereafter, the process proceeds to step S802.
(ステップS802)第1予測モード実施部900-1~第n予測モード実施部900-nは、加算部804が出力した復号画像ブロック信号DBを用いて、予測画像ブロック信号を生成する。サブブロック単位で処理が実施される予測モード実施部、具体的には、第1予測モード実施部900-1と第2予測モード実施部900-2については、画像符号化装置100で採用された各サブブロックの予想モードを示す情報を予測モード選択部902から取得し、予測画像ブロック信号を生成する。第1予測モード実施部900-1~第n予測モード実施部900-nは、生成した第1~第n予測画像ブロック信号を予測モード選択部902に出力する。 (Step S802) The first prediction mode execution unit 900-1 to the n-th prediction mode execution unit 900-n generate a prediction image block signal using the decoded image block signal DB output from the addition unit 804. The prediction mode execution unit that performs processing in units of sub-blocks, specifically, the first prediction mode execution unit 900-1 and the second prediction mode execution unit 900-2 are employed in the image coding apparatus 100. Information indicating the prediction mode of each sub-block is acquired from the prediction mode selection unit 902, and a prediction image block signal is generated. First prediction mode execution unit 900-1 to n-th prediction mode execution unit 900-n output the generated first to n-th prediction image block signals to prediction mode selection unit 902.
 デプス利用予測モード実施部901は、加算部108が出力した復号画像ブロック信号DBと、デプスデプスマップ復号部811が出力したデプスブロック復号信号と、予測モード選択部902が出力した予測に必要な情報(具体的には、予測の方向を示す情報)とを用いて、図3のデプス利用予測モード実施部201で行った処理と同様の処理を実施し、デプス利用予測画像を生成する。デプス利用予測モード実施部901は、生成した予測画像信号を予測モード選択部902に出力する。その後、ステップS803に進む。 The depth use prediction mode execution unit 901 includes a decoded image block signal DB output from the addition unit 108, a depth block decoded signal output from the depth depth map decoding unit 811, and information necessary for prediction output from the prediction mode selection unit 902. (Specifically, information indicating the direction of prediction) is used to perform the same processing as the processing performed by the depth use prediction mode execution unit 201 in FIG. 3 to generate a depth use prediction image. The depth use prediction mode execution unit 901 outputs the generated prediction image signal to the prediction mode selection unit 902. Thereafter, the process proceeds to step S803.
(ステップS803)予測モード選択部902は、予測方式制御部805から入力したイントラ予測符号化情報TPEから、画像符号化装置100で採用された予測モードを示す情報を取り出し、該当する予測モードの予測画像ブロック信号を、イントラ予測画像ブロック信号として選択部806に出力する。予測モード選択部902は、取り出した予測モードがサブブロック単位で実施する予測モードである場合には、さらに各サブブロックの予測モードを取り出し、該当する予測モード実施部にその情報を出力する。その後、イントラ予測を終了する。予測モード選択部902は、取り出した予測モードがデプス利用予測モードである場合には、予測の方向に関する情報を取り出し、デプス利用予測モード実施部901に出力する。 (Step S803) The prediction mode selection unit 902 extracts information indicating the prediction mode employed by the image encoding device 100 from the intra prediction encoding information TPE input from the prediction method control unit 805, and predicts the corresponding prediction mode. The image block signal is output to the selection unit 806 as an intra prediction image block signal. When the extracted prediction mode is a prediction mode to be executed in units of subblocks, the prediction mode selection unit 902 further extracts the prediction mode of each subblock and outputs the information to the corresponding prediction mode execution unit. Then, intra prediction is complete | finished. When the extracted prediction mode is the depth use prediction mode, the prediction mode selection unit 902 extracts information regarding the prediction direction and outputs the information to the depth use prediction mode execution unit 901.
 なお、上述の画像符号化装置100はデプス入力部118とデプスマップ符号化部116を備え、画像復号装置800は、デプスマップ符号化データ入力部814とデプスマップ復号部811を備えていたが、これに限られない。例えば、入力画像に対応するデプスマップに関する情報を、別途の手段で画像復号装置800において利用できるようにしてもよい。例えば、オフラインでの受け渡し、あるいは外部に設置され映像情報と対応づけてデプスマップを記憶したサーバ装置から通信回線を介して、画像符号化装置100及び画像復号装置800がデプスマップを受信するようにしてもよい。そのために、映像情報を示す映像のタイトルが、通信回線を通じて検索可能であって、その映像情報が選択されたときに、対応するデプスマップが受信できるようにする。 The image encoding device 100 described above includes the depth input unit 118 and the depth map encoding unit 116, and the image decoding device 800 includes the depth map encoded data input unit 814 and the depth map decoding unit 811. It is not limited to this. For example, information regarding the depth map corresponding to the input image may be made available in the image decoding apparatus 800 by a separate means. For example, the image encoding device 100 and the image decoding device 800 are configured to receive the depth map via a communication line from a server device that stores the depth map in correspondence with video information installed outside or offline. May be. Therefore, a video title indicating video information can be searched through a communication line, and when the video information is selected, a corresponding depth map can be received.
 また、本実施形態に係る画像符号化装置100は、入力画像とは異なる視点の画像も取得して、入力画像に含まれる画素と、入力画像とは異なる視点の画像に含まれる画素との間の視差を表す値を画素値とするデプスマップを生成するデプスマップ生成部を備えてもよい。その場合、デプスマップ生成部は、生成したデプスマップをデプス入力部118に出力する。 In addition, the image encoding device 100 according to the present embodiment also acquires an image of a viewpoint different from the input image, and between the pixels included in the input image and the pixels included in the image of the viewpoint different from the input image. A depth map generation unit that generates a depth map having a pixel value as a value representing the parallax of the image may be provided. In that case, the depth map generation unit outputs the generated depth map to the depth input unit 118.
 また、本実施形態に係る画像復号装置800は、出力画像と、その出力画像と同じフレームのデプスマップに基づいて出力画像とは異なる視点の第2の出力画像を生成し、外部に出力してもよい。
 また、上述の例では、画像符号化装置100は、入力画像信号を5フレーム毎に入力していたが、本実施形態では、これに限らず、任意の数のフレーム数毎に入力してもよい。
 また、上述の例では、画像復号装置800は、出力画像信号を5フレーム毎に出力していたが、本実施形態では、これに限らず、任意の数のフレーム数毎に出力してもよい。
Further, the image decoding apparatus 800 according to the present embodiment generates a second output image having a viewpoint different from the output image based on the output image and the depth map of the same frame as the output image, and outputs the second output image to the outside. Also good.
In the above-described example, the image encoding apparatus 100 inputs the input image signal every 5 frames. However, in the present embodiment, the image encoding apparatus 100 is not limited to this and may input every arbitrary number of frames. Good.
In the above-described example, the image decoding apparatus 800 outputs the output image signal every 5 frames. However, in the present embodiment, the image decoding apparatus 800 is not limited to this and may output every arbitrary number of frames. .
 また、本実施形態では、符号化する画像を動画像としたが、静止画像であってもよい。
また、符号化する画像を、多視点の画像とし、対応するデプスマップのある視点の画像においてのみ、デプス利用の予測モードを用い、対応するデプスマップのない視点の画像については、従来の予測モードを用いるようにしてもよい。
In this embodiment, the image to be encoded is a moving image, but it may be a still image.
Further, the image to be encoded is a multi-viewpoint image, and the depth-use prediction mode is used only in the viewpoint image with the corresponding depth map. For the viewpoint image without the corresponding depth map, the conventional prediction mode is used. May be used.
 このように、本実施形態では、画面内予測を行う際に、被写体までの距離を示すデプスマップの境界において、連続するような画素予測を抑制する制御を行う予測モードを2つ持つ。従来に比べて、2つの予測モードの追加のみとなっているので、予測モード数の増加による符号量の増加を抑えつつ、予測画像の精度を向上させることができる。そして、予測モード数の増加による符号量の増加を抑えつつ、予測画像の精度が向上しているので、予測画像と入力画像との残差を最小化し、高効率の画像符号化及び復号を実現することができる。
 なお、従来の予測モードに変えて、デプス利用の予測モードを用いるようにすると、予測モード数の増加もないので、さらに符号量の増加を抑えることができる。
As described above, the present embodiment has two prediction modes for performing control to suppress continuous pixel prediction at the boundary of the depth map indicating the distance to the subject when performing intra prediction. Since only two prediction modes are added as compared with the prior art, it is possible to improve the accuracy of the predicted image while suppressing an increase in code amount due to an increase in the number of prediction modes. And the accuracy of the prediction image is improved while suppressing the increase in the amount of code due to the increase in the number of prediction modes, so the residual between the prediction image and the input image is minimized, realizing highly efficient image encoding and decoding can do.
If the depth-based prediction mode is used instead of the conventional prediction mode, the number of prediction modes does not increase, and therefore the increase in code amount can be further suppressed.
 なお、上述した実施形態における画像符号化装置100、画像復号装置800の一部、例えば、減算部102、直交変換部103、量子化部104、エントロピー符号化部105、逆量子化部106、逆直交変換部107、加算部108、予測方式制御部109、選択部110、デブロッキング・フィルタ部111、動き補償部113、動きベクトル検出部114、奥行き情報利用イントラ予測部115、デプスマップ符号化部116並びにデプスマップ復号部117、及びエントロピー復号部801、逆量子化部802、逆直交変換部803、加算部804、予測方式制御部805、選択部806、デブロッキング・フィルタ部807、動き補償部809、奥行き情報利用イントラ予測部810、並びにデプスマップ復号部811をコンピュータで実現するようにしても良い。 In addition, a part of the image coding apparatus 100 and the image decoding apparatus 800 in the above-described embodiment, for example, the subtraction unit 102, the orthogonal transformation unit 103, the quantization unit 104, the entropy coding unit 105, the inverse quantization unit 106, and the inverse Orthogonal transformation unit 107, addition unit 108, prediction scheme control unit 109, selection unit 110, deblocking filter unit 111, motion compensation unit 113, motion vector detection unit 114, depth information use intra prediction unit 115, depth map encoding unit 116, a depth map decoding unit 117, an entropy decoding unit 801, an inverse quantization unit 802, an inverse orthogonal transform unit 803, an addition unit 804, a prediction scheme control unit 805, a selection unit 806, a deblocking filter unit 807, a motion compensation unit 809, a depth information use intra prediction unit 810 and a depth map decoding unit 811 are connected. It may be realized by Yuta.
 その場合、この制御機能を実現するためのプログラムをコンピュータ読み取り可能な記録媒体に記録して、この記録媒体に記録されたプログラムをコンピュータシステムに読み込ませ、実行することによって実現しても良い。なお、ここでいう「コンピュータシステム」とは、画像符号化装置100又は画像復号装置800に内蔵されたコンピュータシステムであって、OSや周辺機器等のハードウェアを含むものとする。また、「コンピュータ読み取り可能な記録媒体」とは、フレキシブルディスク、光磁気ディスク、ROM、CD-ROM等の可搬媒体、コンピュータシステムに内蔵されるハードディスク等の記憶装置のことをいう。さらに「コンピュータ読み取り可能な記録媒体」とは、インターネット等のネットワークや電話回線等の通信回線を介してプログラムを送信する場合の通信線のように、短時間、動的にプログラムを保持するもの、その場合のサーバやクライアントとなるコンピュータシステム内部の揮発性メモリのように、一定時間プログラムを保持しているものも含んでも良い。また上記プログラムは、前述した機能の一部を実現するためのものであっても良く、さらに前述した機能をコンピュータシステムにすでに記録されているプログラムとの組み合わせで実現できるものであっても良い。 In that case, the program for realizing the control function may be recorded on a computer-readable recording medium, and the program recorded on the recording medium may be read into a computer system and executed. The “computer system” here is a computer system built in the image encoding device 100 or the image decoding device 800, and includes an OS and hardware such as peripheral devices. The “computer-readable recording medium” refers to a storage device such as a flexible medium, a magneto-optical disk, a portable medium such as a ROM or a CD-ROM, and a hard disk incorporated in a computer system. Further, the “computer-readable recording medium” is a medium that dynamically holds a program for a short time, such as a communication line when transmitting a program via a network such as the Internet or a communication line such as a telephone line, In such a case, a volatile memory inside a computer system serving as a server or a client may be included and a program that holds a program for a certain period of time. The program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system.
 また、上述した実施形態における画像符号化装置100及び画像復号装置800の一部、または全部を、LSI(Large Scale Integration)等の集積回路として実現しても良い。画像符号化装置100及び画像復号装置800の各機能ブロックは個別にプロセッサ化してもよいし、一部、または全部を集積してプロセッサ化しても良い。また、集積回路化の手法はLSIに限らず専用回路、または汎用プロセッサで実現しても良い。また、半導体技術の進歩によりLSIに代替する集積回路化の技術が出現した場合、当該技術による集積回路を用いても良い。 Further, part or all of the image encoding device 100 and the image decoding device 800 in the above-described embodiment may be realized as an integrated circuit such as an LSI (Large Scale Integration). Each functional block of the image encoding device 100 and the image decoding device 800 may be individually made into a processor, or a part or all of them may be integrated into a processor. Further, the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor. Further, in the case where an integrated circuit technology that replaces LSI appears due to progress in semiconductor technology, an integrated circuit based on the technology may be used.
 以上、本発明の一態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置であって、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記入力画像が表す被写体の境界がある前記周辺画素を用いることを抑制する画面内予測部を具備する。 As described above, according to one aspect of the present invention, when an input image is encoded, an image for performing intra prediction that predicts a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel. The encoding device, wherein when performing the intra prediction, a screen that suppresses using the peripheral pixels having a subject boundary represented by the input image between the processing target pixels and the peripheral pixels. An inner prediction unit is provided.
 また、本発明の他の態様は、上述の画像符号化装置であって、前記画面内予測部は、前記被写体の境界を、前記入力画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を具備する。 Another aspect of the present invention is the image encoding device described above, wherein the intra-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the input image. A boundary detection unit is provided.
 また、本発明の他の態様は、上述の画像符号化装置であって、前記画面内予測部は、前記周辺画素のうち、前記処理対象画素の所定の方向に隣接する画素と、前記処理対象画素との間に前記被写体の境界がないときは、前記所定の方向に隣接する画素を用いて、前記処理対象画素の画素値を予測し、前記所定の方向に隣接する画素と、前記処理対象画素との間に前記被写体の境界があるときは、前記所定の方向に隣接する画素を用いて前記処理対象画素の画素値を予測することを抑制する予測画像生成部を具備する。 Another aspect of the present invention is the image encoding device described above, wherein the intra prediction unit includes a pixel adjacent to the processing target pixel in a predetermined direction among the peripheral pixels, and the processing target. When there is no boundary between the subject and the pixel, the pixel value of the processing target pixel is predicted using the pixel adjacent in the predetermined direction, and the pixel adjacent in the predetermined direction and the processing target When there is a boundary between the subject and a pixel, a predicted image generation unit is provided that suppresses predicting the pixel value of the processing target pixel using a pixel adjacent in the predetermined direction.
 また、本発明の他の態様は、上述の画像符号化装置であって、前記画面内予測部は、前記処理対象画素の画素値を予測する際に用いる周辺画素を、前記周辺画素が表す被写体までの距離を示す情報と、前記処理対象画素が表す被写体までの距離を示す情報との差に少なくとも基づき決定する予測画像生成部を具備する。 Another aspect of the present invention is the above-described image encoding device, in which the in-screen prediction unit represents a peripheral pixel used when predicting a pixel value of the processing target pixel. A prediction image generation unit that determines at least based on the difference between the information indicating the distance to the subject and the information indicating the distance to the subject represented by the processing target pixel.
 また、本実施形態では、上述の画像符号化装置であって、前記画面内予測部は、前記処理対象画素の画素値を予測する際に用いる周辺画素を、前記周辺画素と前記処理対象画素との距離に少なくとも基づき決定する予測画像生成部を具備する。 Further, in the present embodiment, in the above-described image encoding device, the intra-screen prediction unit uses the peripheral pixels and the processing target pixels as peripheral pixels used when predicting the pixel value of the processing target pixel. A predicted image generation unit that determines at least based on the distance.
 また、本発明の他の態様は、符号化画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置であって、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記符号化画像が表す被写体の境界がある前記周辺画素を用いることを抑制する画面内予測部を具備する。 Further, according to another aspect of the present invention, when decoding an encoded image, intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel. In the image decoding device, when performing the intra prediction, it is possible to suppress use of the peripheral pixels having a subject boundary represented by the encoded image between the processing target pixels and the peripheral pixels. An in-screen prediction unit is provided.
 また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記被写体の境界を、前記符号化画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を具備する。 Another aspect of the present invention is the above-described image decoding device, wherein the in-screen prediction unit detects a boundary of the subject using information indicating a distance to the subject of the encoded image. A boundary detection unit is provided.
 また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記周辺画素のうち、前記処理対象画素の所定の方向に隣接する画素と、前記処理対象画素との間に前記被写体の境界がないときは、前記所定の方向に隣接する画素を用いて、前記処理対象画素の画素値を予測し、前記所定の方向に隣接する画素と、前記処理対象画素との間に前記被写体の境界があるときは、前記所定の方向に隣接する画素を用いて前記処理対象画素の画素値を予測することを抑制する予測画像生成部を具備する。 Another aspect of the present invention is the above-described image decoding device, wherein the intra prediction unit includes a pixel adjacent to the processing target pixel in a predetermined direction among the peripheral pixels, and the processing target pixel. If there is no boundary between the subject and the pixel, the pixel value of the processing target pixel is predicted using pixels adjacent in the predetermined direction, and the pixel adjacent to the predetermined direction and the processing target pixel A prediction image generation unit that suppresses prediction of a pixel value of the processing target pixel using pixels adjacent in the predetermined direction when the boundary of the subject is between the two.
 また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記処理対象画素の画素値を予測する際に用いる周辺画素を、前記周辺画素が表す被写体までの距離を示す情報と、前記処理対象画素が表す被写体までの距離を示す情報との差に少なくとも基づき決定する予測画像生成部を具備することを特徴とする。 Another aspect of the present invention is the above-described image decoding device, wherein the intra-screen prediction unit determines a peripheral pixel used when predicting a pixel value of the processing target pixel to a subject represented by the peripheral pixel. A predicted image generation unit that determines at least based on the difference between the information indicating the distance to the subject and the information indicating the distance to the subject represented by the processing target pixel.
 また、本発明の他の態様は、上述の画像復号装置であって、前記画面内予測部は、前記処理対象画素の画素値を予測する際に用いる周辺画素を、前記周辺画素と前記処理対象画素との距離に少なくとも基づき決定する予測画像生成部を具備する。 Another aspect of the present invention is the above-described image decoding device, wherein the intra-screen prediction unit uses the peripheral pixels and the processing target as peripheral pixels used when predicting a pixel value of the processing target pixel. A predicted image generation unit that determines at least based on the distance to the pixel;
 また、本発明の他の態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化方法であって、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記入力画像が表す被写体の境界がある前記周辺画素を用いることを抑制する過程を有する。 According to another aspect of the present invention, when an input image is encoded, intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel. In the image encoding method, when the intra prediction is performed, use of the peripheral pixels having a boundary of a subject represented by the input image between the processing target pixels and the peripheral pixels is suppressed. Have a process.
 また、本発明の他の態様は、符号化画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号方法であって、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記符号化画像が表す被写体の境界がある前記周辺画素を用いることを抑制する過程を有する。 Further, according to another aspect of the present invention, when decoding an encoded image, intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel. In the image decoding method, when the intra prediction is performed, use of the peripheral pixels having a boundary of a subject represented by the encoded image between the processing target pixels and the peripheral pixels is suppressed. Have a process.
 また、本発明の他の態様は、入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置のコンピュータを、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記入力画像が表す被写体の境界がある前記周辺画素を用いることを抑制する画面内予測部として機能させるためのプログラムである。 According to another aspect of the present invention, when an input image is encoded, intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel. When the computer of the image encoding device performs the intra prediction, the use of the peripheral pixels having the boundary of the subject represented by the input image between the processing target pixels and the peripheral pixels is suppressed. It is a program for functioning as an in-screen prediction unit.
 また、本発明の他の態様は、符号化画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置のコンピュータを、前記画面内予測を行なう際に、前記周辺画素のうち、前記処理対象画素との間に前記符号化画像が表す被写体の境界がある前記周辺画素を用いることを抑制する画面内予測部として機能させるためのプログラムである。 Further, according to another aspect of the present invention, when decoding an encoded image, intra-screen prediction is performed to predict the pixel value of the processing target pixel using the pixel values of peripheral pixels around the processing target pixel. When the computer of the image decoding apparatus performs the intra prediction, the use of the peripheral pixels having a boundary of the subject represented by the encoded image between the peripheral pixels and the processing target pixel is suppressed. It is a program for functioning as an in-screen prediction unit.
 以上、この発明の実施形態を、図面を参照して詳述してきたが、具体的な構成はこの実施形態に限られるものではなく、この発明の要旨を逸脱しない範囲の設計変更等も含まれる。 The embodiment of the present invention has been described in detail with reference to the drawings. However, the specific configuration is not limited to this embodiment, and design changes and the like within a scope not departing from the gist of the present invention are included. .
 10…動画像伝送システム
 100…画像符号化装置
 101…画像入力部
 102…減算部
 103…直交変換部
 104…量子化部
 105…エントロピー符号化部
 106…逆量子化部
 107…逆直交変換部
 108…加算部
 109…予測方式制御部
 110…選択部
 111…デブロッキング・フィルタ部
 112…フレームメモリ部
 113…動き補償部
 114…動きベクトル検出部
 115…奥行き情報利用イントラ予測部
 116…デプスマップ符号化部
 117…デプスマップ復号部
 118…デプス入力部
 120…インター予測部
 121…イントラ予測部
 200-1…第1予測モード実施部
 200-2…第2予測モード実施部
 200-n…第n予測モード実施部
 201…デプス利用予測モード実施部
 202…予測モード選択部
 300…境界制御予測画像生成部
 301…境界予測制御部
 302…被写体境界検出部
 500…通信網
 600…表示装置
 800…画像復号装置
 801…エントロピー復号部
 802…逆量子化部
 803…逆直交変換部
 804…加算部
 805…予測方式制御部
 806…選択部
 807…デブロッキング・フィルタ部
 808…フレームメモリ
 809…動き補償部
 810…奥行き情報利用イントラ予測部
 811…デプスマップ復号部
 812…画像出力部
 813…符号化データ入力部
 814…デプスマップ符号化データ入力部
 820…インター処理部
 821…イントラ処理部
 900-1…第1予測モード実施部
 900-2…第2予測モード実施部
 900-n…第n予測モード実施部
 901…デプス利用予測モード実施部
 902…予測モード選択部
DESCRIPTION OF SYMBOLS 10 ... Moving image transmission system 100 ... Image coding apparatus 101 ... Image input part 102 ... Subtraction part 103 ... Orthogonal transformation part 104 ... Quantization part 105 ... Entropy coding part 106 ... Inverse quantization part 107 ... Inverse orthogonal transformation part 108 DESCRIPTION OF SYMBOLS ... Addition unit 109 ... Prediction method control unit 110 ... Selection unit 111 ... Deblocking filter unit 112 ... Frame memory unit 113 ... Motion compensation unit 114 ... Motion vector detection unit 115 ... Depth information utilization intra prediction unit 116 ... Depth map encoding 117: Depth map decoding unit 118 ... Depth input unit 120 ... Inter prediction unit 121 ... Intra prediction unit 200-1 ... First prediction mode execution unit 200-2 ... Second prediction mode execution unit 200-n ... nth prediction mode Execution unit 201 ... depth use prediction mode execution unit 202 ... prediction mode selection unit 300 ... boundary Control predicted image generation unit 301 ... Boundary prediction control unit 302 ... Subject boundary detection unit 500 ... Communication network 600 ... Display device 800 ... Image decoding device 801 ... Entropy decoding unit 802 ... Inverse quantization unit 803 ... Inverse orthogonal transformation unit 804 ... Addition 805 ... Prediction method control unit 806 ... Selection unit 807 ... Deblocking filter unit 808 ... Frame memory 809 ... Motion compensation unit 810 ... Depth information use intra prediction unit 811 ... Depth map decoding unit 812 ... Image output unit 813 ... Encoding Data input unit 814 ... Depth map encoded data input unit 820 ... Inter processing unit 821 ... Intra processing unit 900-1 ... First prediction mode execution unit 900-2 ... Second prediction mode execution unit 900-n ... nth prediction mode Implementation unit 901 ... Depth use prediction mode implementation unit 902 ... Prediction mode selection unit

Claims (10)

  1.  入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置であって、
     前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部
     を具備することを特徴とする画像符号化装置。
    When encoding an input image, an image encoding device that performs in-screen prediction for predicting a pixel value of a processing target pixel using pixel values of peripheral pixels around the processing target pixel,
    An image code comprising: an in-screen prediction unit that determines a predicted value for each pixel based on information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels. Device.
  2.  前記画面内予測部は、
     前記処理対象画素と前記周辺画素との画素間の距離に基づいて画素ごとの予測値を決定すること
     を特徴とする請求項1に記載の画像符号化装置。
    The intra prediction unit
    The image coding apparatus according to claim 1, wherein a prediction value for each pixel is determined based on a distance between pixels of the processing target pixel and the surrounding pixels.
  3.  前記画面内予測部は、
     前記被写体の境界を、前記入力画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を備え、
     前記処理対象画素と所定の方向に隣接する画素との間に前記境界が検出されなかった場合には、前記所定の方向に隣接する画素を用いて前記処理対象画素の予測値を予測すること
     を特徴とする請求項1に記載の画像符号化装置。
    The intra prediction unit
    A subject boundary detection unit that detects a boundary of the subject using information indicating a distance to the subject of the input image;
    When the boundary is not detected between the processing target pixel and a pixel adjacent in a predetermined direction, the prediction value of the processing target pixel is predicted using the pixel adjacent in the predetermined direction. The image encoding apparatus according to claim 1, wherein the apparatus is an image encoding apparatus.
  4.  符号化画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置であって、
     前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部
    を具備することを特徴とする画像復号装置。
    When decoding an encoded image, an image decoding device that performs in-screen prediction that predicts a pixel value of a processing target pixel using pixel values of surrounding pixels around the processing target pixel,
    An intra-screen prediction unit that determines a predicted value for each pixel based on information indicating a distance to the subject in the processing target pixel and information indicating a distance to the subject in the peripheral pixels when performing the intra-screen prediction. An image decoding apparatus comprising:
  5.  前記画面内予測部は、
     前記処理対象画素と前記周辺画素との画素間の距離に基づいて画素ごとの予測値を決定すること
     を特徴とする請求項4に記載の画像復号装置。
    The intra prediction unit
    The image decoding apparatus according to claim 4, wherein a prediction value for each pixel is determined based on a distance between the processing target pixel and the peripheral pixel.
  6.  前記画面内予測部は、
     前記被写体の境界を、前記入力画像の被写体までの距離を示す情報を用いて検出する被写体境界検出部を備え、
     前記処理対象画素と所定の方向に隣接する画素との間に前記境界が検出されなかった場合には、前記所定の方向に隣接する画素を用いて前記処理対象画素の予測値を予測すること
     を特徴とする請求項4に記載の画像復号装置。
    The intra prediction unit
    A subject boundary detection unit that detects a boundary of the subject using information indicating a distance to the subject of the input image;
    When the boundary is not detected between the processing target pixel and a pixel adjacent in a predetermined direction, the prediction value of the processing target pixel is predicted using the pixel adjacent in the predetermined direction. The image decoding device according to claim 4, characterized in that:
  7.  入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化方法であって、
     前記画面内予測を行う際に、画面内予測部が前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する過程
     を有することを特徴とする画像符号化方法。
    When encoding an input image, an image encoding method for performing intra prediction to predict a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel,
    When performing the intra prediction, the intra prediction unit determines a prediction value for each pixel based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. An image encoding method comprising the steps of:
  8.  入力画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号方法であって、
     前記画面内予測を行う際に、画面内予測部が前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する過程
     を有することを特徴とする画像復号方法。
    When decoding an input image, an image decoding method for performing intra prediction to predict a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel,
    When performing the intra prediction, the intra prediction unit determines a prediction value for each pixel based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. An image decoding method comprising the steps of:
  9.  入力画像を符号化する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像符号化装置を、
     前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部
     として機能させるためのプログラム。
    When encoding an input image, an image encoding device that performs in-screen prediction for predicting a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel,
    When performing the intra prediction, an intra prediction unit that determines a prediction value for each pixel based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. A program to make it work.
  10.  入力画像を復号する際に、処理対象画素の周辺にある周辺画素の画素値を用いて、前記処理対象画素の画素値を予測する画面内予測を行う画像復号装置を、
     前記画面内予測を行う際に、前記処理対象画素における被写体までの距離を示す情報と前記周辺画素における被写体までの距離を示す情報とに基づいて画素ごとの予測値を決定する画面内予測部
     として機能させるためのプログラム。
    When decoding an input image, an image decoding device that performs in-screen prediction for predicting a pixel value of the processing target pixel using pixel values of peripheral pixels around the processing target pixel,
    When performing the intra prediction, an intra prediction unit that determines a prediction value for each pixel based on information indicating the distance to the subject in the processing target pixel and information indicating the distance to the subject in the peripheral pixels. A program to make it work.
PCT/JP2012/063503 2011-05-25 2012-05-25 Image encoding device, image decoding device, image encoding method, image decoding method and program WO2012161318A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011-117425 2011-05-25
JP2011117425A JP2014150297A (en) 2011-05-25 2011-05-25 Image encoding device, image decoding device, image encoding method, image decoding method, and program

Publications (1)

Publication Number Publication Date
WO2012161318A1 true WO2012161318A1 (en) 2012-11-29

Family

ID=47217383

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/063503 WO2012161318A1 (en) 2011-05-25 2012-05-25 Image encoding device, image decoding device, image encoding method, image decoding method and program

Country Status (2)

Country Link
JP (1) JP2014150297A (en)
WO (1) WO2012161318A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116320395A (en) * 2022-12-27 2023-06-23 维沃移动通信有限公司 Image processing method, device, electronic equipment and readable storage medium

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5711636B2 (en) * 2011-09-26 2015-05-07 日本電信電話株式会社 Image encoding method, image decoding method, image encoding device, image decoding device, image encoding program, image decoding program
JP5729825B2 (en) * 2011-09-26 2015-06-03 日本電信電話株式会社 Image encoding method, image decoding method, image encoding device, image decoding device, image encoding program, and image decoding program
JP5759357B2 (en) * 2011-12-13 2015-08-05 日本電信電話株式会社 Video encoding method, video decoding method, video encoding device, video decoding device, video encoding program, and video decoding program

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009090884A1 (en) * 2008-01-18 2009-07-23 Panasonic Corporation Image encoding method and image decoding method
JP2010056701A (en) * 2008-08-27 2010-03-11 Nippon Telegr & Teleph Corp <Ntt> Intra-image predictive encoding method, intra-image predictive decoding method, devices thereof, programs therefor, and recording medium recording the programs

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009090884A1 (en) * 2008-01-18 2009-07-23 Panasonic Corporation Image encoding method and image decoding method
JP2010056701A (en) * 2008-08-27 2010-03-11 Nippon Telegr & Teleph Corp <Ntt> Intra-image predictive encoding method, intra-image predictive decoding method, devices thereof, programs therefor, and recording medium recording the programs

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
FENG ZOU ET AL.: "EDGE-BASED ADAPTIVE DIRECTIONAL INTRA PREDICTION", 28TH PICTURE CODING SYMPOSIUM, IEEE, 8 December 2010 (2010-12-08), pages 366 - 369 *
MIN-KOO KANG ET AL.: "Geometry-based Block Partitioning for Efficient Intra Prediction in Depth Video Coding", VISUAL INFORMATION PROCESSING AND COMMUNICATION, PROC. OF THE SPIE-IS&T ELECTRIC IMAGING, SPIE, vol. 7543, 17 January 2010 (2010-01-17), pages 75430A-1 - 75430A-11 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116320395A (en) * 2022-12-27 2023-06-23 维沃移动通信有限公司 Image processing method, device, electronic equipment and readable storage medium

Also Published As

Publication number Publication date
JP2014150297A (en) 2014-08-21

Similar Documents

Publication Publication Date Title
KR101773693B1 (en) Disparity vector derivation in 3d video coding for skip and direct modes
US9020030B2 (en) Smoothing overlapped regions resulting from geometric motion partitioning
EP3672249A1 (en) Inter frame prediction method and device and codec for video images
US8204127B2 (en) Method and apparatus for encoding and decoding image by using multiple reference-based motion prediction
WO2018026887A1 (en) Geometry transformation-based adaptive loop filtering
EP3453173A1 (en) Control-point based intra direction representation for intra coding
CN111971960B (en) Method for processing image based on inter prediction mode and apparatus therefor
JP6039178B2 (en) Image encoding apparatus, image decoding apparatus, method and program thereof
US8204120B2 (en) Method for intra prediction coding of image data
WO2011130186A2 (en) Fixed point implementation for geometric motion partitioning
KR20170093833A (en) Coding of intra modes
KR20120118463A (en) Image processing device, method, and program
WO2012161318A1 (en) Image encoding device, image decoding device, image encoding method, image decoding method and program
CN113940078A (en) Simplification of bidirectional optical flow computation in video coding and decoding
AU2020347025B2 (en) Image encoding/decoding method and device for performing bdof, and method for transmitting bitstream
JP6232117B2 (en) Image encoding method, image decoding method, and recording medium
WO2013077304A1 (en) Image coding device, image decoding device, and methods and programs thereof
KR20230149297A (en) Intra prediction method and device based on intra prediction mode derivation
CN116647683A (en) Quantization processing method and device
JP2013179554A (en) Image encoding device, image decoding device, image encoding method, image decoding method, and program
WO2013035452A1 (en) Image encoding method, image decoding method, and apparatuses and programs thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12788713

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12788713

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP