WO2013073422A1 - Dispositif de codage vidéo - Google Patents

Dispositif de codage vidéo Download PDF

Info

Publication number
WO2013073422A1
WO2013073422A1 PCT/JP2012/078784 JP2012078784W WO2013073422A1 WO 2013073422 A1 WO2013073422 A1 WO 2013073422A1 JP 2012078784 W JP2012078784 W JP 2012078784W WO 2013073422 A1 WO2013073422 A1 WO 2013073422A1
Authority
WO
WIPO (PCT)
Prior art keywords
inter
image
prediction
encoding
screen prediction
Prior art date
Application number
PCT/JP2012/078784
Other languages
English (en)
Japanese (ja)
Inventor
信博 知原
昌史 高橋
山口 宗明
雅俊 近藤
Original Assignee
株式会社日立国際電気
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立国際電気 filed Critical 株式会社日立国際電気
Publication of WO2013073422A1 publication Critical patent/WO2013073422A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/57Motion estimation characterised by a search window with variable size or shape
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/43Hardware specially adapted for motion estimation or compensation
    • H04N19/433Hardware specially adapted for motion estimation or compensation characterised by techniques for memory access

Definitions

  • the present invention relates to a technique for encoding a moving image.
  • video coding technology has become an indispensable technology due to the increase in video distribution content due to the development of broadband networks and the use of large-capacity storage media such as DVDs and large-screen video display devices.
  • a technology for encoding at a high resolution in the moving image encoding technology is indispensable.
  • the moving image encoding process is a process of converting an original image (input image) input to the moving image encoding apparatus into a stream having a smaller data amount.
  • the international standard H.264 is available as one of moving image encoding technologies capable of encoding with high resolution and high image quality.
  • the H.264 / AVC standard (Non-Patent Document 1) exists.
  • an input image in a wide transmission band for example, about 1.5 Gbps in HD-SDI (High Definition Serial Digital Interface)
  • a low transmission band for example, about 15 Mbps in terrestrial digital broadcasting.
  • intra prediction a plurality of prediction methods are prepared according to combinations of block sizes and prediction directions serving as prediction units.
  • inter-screen prediction a plurality of prediction methods are prepared according to the size of a block which is a prediction unit.
  • H. in H.264 / AVC an encoding method with high image quality and high compression is realized by dynamically selecting these prediction methods according to the target image quality and code amount.
  • FIG. 1 is a diagram illustrating a configuration of an image encoding device that performs H.264 / AVC encoding processing.
  • the mode selection unit 930 selects the intra prediction unit 910. Then, the stream 91 is obtained from the original image 90 through the intra prediction unit 910, the orthogonal transform unit 940, the quantization unit 950, and the variable length coding unit 980. In the encoding process using inter-screen prediction, the mode selection unit 930 selects the inter-screen prediction unit 920. Then, the stream 91 is obtained from the original image 90 through the inter-screen prediction unit 920, the orthogonal transform unit 940, the quantization unit 950, and the variable length encoding unit 980.
  • the quantization granularity is adjusted using the quantization coefficient D84 determined by the encoding control unit 990. If the quantization granularity is small, the code amount is large, but the image is likely to be closer to the original image, and if the quantization granularity is large, the code amount is small but the image is likely to be separated from the original image.
  • H. In H.264 / AVC encoding encoding with high image quality and high compression efficiency is realized by dynamically selecting the quantization coefficient D84 according to the target image quality and code amount.
  • the in-screen prediction unit 910 receives an original image 90 and a reconstructed image 92 that is a peripheral image.
  • the reconstructed image 92 is an image configured by adding the restored difference image 97 output from the inverse orthogonal transform unit 970 and the predicted image 95 output by the mode selection unit 930, and is the same as the original image (that is, It is an image group of a plurality of past frames including a frame (currently being encoded).
  • the reconstructed image 92 referred to in intra-screen prediction is the same frame as the original image 90.
  • the intra-screen prediction unit 910 selects an appropriate intra-screen prediction mode (intra-screen prediction method) from the original image 90 and the reconstructed image 92 by the intra-screen prediction process, and represents the mode information of the intra-screen prediction mode.
  • An intra-screen prediction information D81, an intra-screen prediction image 93 which is a prediction result, and an intra-screen prediction error representing a difference between the original image 90 and the intra-screen prediction image 93 are generated.
  • the inter-screen prediction unit 920 receives the input of the original image 90 and the reconstructed image 92 generated from the original image before and after (past or future), and receives the inter-screen prediction information D82, the inter-screen prediction image 94, and the original image An inter-screen prediction error representing a difference between 90 and the inter-screen prediction image 94 is generated.
  • the mode selection unit 930 uses the intra-screen prediction error input from the intra-screen prediction unit 910 and the inter-screen prediction error input from the inter-screen prediction unit 920 according to the encoding mode selection algorithm, and performs intra-screen prediction and inter-screen prediction. Determine any coding mode of prediction.
  • the intra prediction image 93 is output as the prediction image 95
  • the inter prediction image 94 is output as the prediction image 95.
  • the encoding mode selection algorithm greatly affects the code amount and image quality of the stream 91, and therefore there are various methods depending on the content of the original image 90 to be encoded and the purpose of video encoding.
  • the orthogonal transform unit 940 generates a frequency component D83 from the difference image 96 that is a difference between the original image 90 and the predicted image 95 by orthogonal transform processing.
  • the unit of prediction (block) and the unit of orthogonal transformation are not necessarily the same, and the difference image 96 may be further divided and orthogonally transformed.
  • the quantization unit 950 performs a quantization process from the quantization coefficient D84 input from the encoding control unit 990 and the frequency component D83 input from the orthogonal transform unit 940, and generates a quantized value D85 with a reduced amount of information. Output.
  • the inverse quantization unit 960 performs an inverse quantization process on the quantized value D85 to generate a restored frequency component D86.
  • the inverse orthogonal transform unit 970 performs an inverse orthogonal transform process on the reconstructed frequency component D86 to generate a reconstructed difference image 97. Then, the generated restored differential image 97 and the predicted image 95 output by the mode selection unit 930 are added together in the same block, and stored as a reconstructed image 92 in a storage device such as a memory.
  • variable length encoding unit 980 encodes the quantized value D85 and the intra-screen prediction information D81 or the inter-screen prediction information D82 into a data string having a smaller data amount, and outputs it as a stream 91.
  • the stream buffer 1000 once buffers the stream 91, and then outputs it to the transmission path or a subsequent decoder.
  • the code amount of the buffered stream 91 is output to the encoding control unit 990 as code amount information D87.
  • the encoding control unit 990 determines the quantization coefficient D84 according to the rate control algorithm using the code amount information D87 input from the stream buffer 1000, and outputs this to the quantization unit 950.
  • Patent Document 1 discloses a method for improving objective image quality by referring to a predicted image closer to an original image, in which the quantization coefficient is reduced in inter-screen prediction by managing the reference relationship in inter-screen prediction in units of pictures. Is disclosed. Further, in Patent Document 1, when a sharp image and a dull image are repeatedly displayed every other frame, a visual characteristic called “repetitive illusion of a dull image” that looks like a sharp image as a whole by human vision is utilized. A method for improving subjective image quality is also disclosed.
  • FIG. 22 is a diagram for explaining a method for improving the encoding efficiency using the “repetitive illusion of a sharp image” in the image encoding technique described in Patent Document 1.
  • FIG. 22 when alternately displaying a high-quality image with a reduced quantization coefficient and a low-quality image with a reduced quantization amount by increasing the quantization coefficient every other frame, “sharpness” is displayed. Due to the “repetitive illusion of images”, the entire moving image appears to be equivalent to a high-quality moving image with a small quantization coefficient in human vision.
  • the code output from the moving image encoding device is input to the moving image decoding device, and a decoding process is executed to display the same image.
  • the time difference from display to display is the delay time, and in a general moving picture encoding apparatus and moving picture decoding apparatus, the delay is about several hundred milliseconds to several seconds.
  • image data is transmitted at a fixed bit rate determined from the data amount of the entire image data, and the transmitted image data is stored and processed in a buffer.
  • code data having a bit rate or higher
  • data in the buffer is insufficient for processing in this area, and waiting for transmission of code data to the buffer occurs. If all processing is not completed for this area, this area cannot be displayed, resulting in an increase in delay time in displaying the entire image.
  • the delay time is determined by the time for processing the assumed maximum code generation amount, and this delay time is greatly influenced by the difference between the bit rate and the maximum code generation amount, that is, the variation amount of the generated code amount.
  • FIG. 23 is a diagram for explaining a code amount variation by quantization control in units of pictures in the image coding technique described in Patent Document 1.
  • the present invention provides a moving image coding capable of performing high-quality and low-delay coding processing by suppressing fluctuations in the code amount while improving the image quality in the moving image coding processing.
  • An object is to provide an apparatus, a moving image encoding method, and a moving image encoding program.
  • a moving image encoding apparatus that encodes a moving image including a plurality of images arranged in time series by intra prediction or inter prediction.
  • a plurality of images are divided into a reference region that is a region that refers to a pixel value in the intra prediction or the inter-screen prediction and a non-reference region that is a region that does not refer to a pixel value, and
  • a reference control unit that determines the reference region and the non-reference region in the plurality of images so that the reference region and the non-reference region in the plurality of images are switched at a predetermined interval in FIG.
  • a quantization unit that performs quantization with a width (gradation) coarser than the reference region for the non-reference region when the image is encoded.
  • the reference control unit may be configured such that a ratio of the reference area to the non-reference area in a certain unit area in any of the certain unit areas included in the image.
  • the reference area and the non-reference area may be determined so as to be the same.
  • the video encoding device further includes an intra-screen prediction unit that performs the encoding by the intra-screen prediction on the plurality of images, and the intra-screen prediction unit includes: In the image, the encoding is performed with reference to only the pixels in the reference area according to the positional relationship between the current macroblock that is the macroblock to be encoded and the reference area or the non-reference area. It may be.
  • the video encoding device further includes an inter-screen prediction unit that performs the encoding by the inter-screen prediction on the plurality of images, and the inter-screen prediction unit includes:
  • the motion vector search range may refer to at least one of the plurality of images serving as the reference region.
  • FIG. 1 shows an example of a structure of the moving image encoder which concerns on one Embodiment of this invention.
  • H The figure which shows the kind of prediction direction in the prediction in a screen with respect to 4x4 block in a H.264 / AVC specification.
  • limiting information The schematic diagram which shows a response
  • limiting information The schematic diagram which shows a response
  • Conventional H.264 The block diagram of the image coding apparatus which performs a H.264 / AVC encoding process. The figure explaining the improvement method of the encoding efficiency using the conventional "repetitive illusion of a sharp image”. The figure explaining the code amount fluctuation
  • FIG. 1 is a diagram illustrating an example of a configuration of a video encoding apparatus according to the present embodiment.
  • the video encoding apparatus 1 includes an intra-screen prediction unit 110, an inter-screen prediction unit 120, a mode selection unit 130, an orthogonal transform unit 140, a quantization unit 150, an inverse quantization unit 160, and an inverse orthogonal transform.
  • LSI Large Scale Integration
  • FPGA Field-Programmable Gate Array
  • various image processing for performing moving image encoding processing is possible. It can be applied to equipment. Further, for example, it can be realized by a configuration similar to that of a general computer such as a memory (not shown) such as a CPU (Central Processing Unit) and a RAM (Random Access Memory), a storage device such as a hard disk, and a network interface. That is, the function of each component of the moving image encoding device 1 can be realized by, for example, the CPU reading and executing a program stored in a hard disk or the like.
  • the intra-screen prediction unit 110 acquires the original image 10, the reconstructed image 12 that is an image located around the original image 10, and intra-screen prediction restriction information D11 (described later) output from the reference control unit 210. Then, the intra-screen prediction information D1, the intra-screen prediction image 13 that is the prediction result of the intra-screen prediction, and the intra-screen prediction error representing the difference between the original image 10 and the intra-screen prediction image 13 are generated.
  • the intra-screen prediction information D1 includes an intra-screen prediction block type that is information indicating the block size for which intra-screen prediction has been performed, and intra-screen prediction mode information indicating the direction of intra-screen prediction.
  • the inter-screen prediction unit 120 includes the original image 10 and a past original image (on the time axis, an original image located in front of the original image 10. The same applies hereinafter) or a future original image (on the time axis,
  • the reconstructed image 12 generated from the original image located after the original image 10 (the same applies hereinafter) and the inter-screen prediction restriction information D12 (described later) output from the reference control unit 210 are acquired, and the inter-screen is acquired.
  • the prediction information D2, the inter-screen prediction image 14 which is the prediction result of the inter-screen prediction, and the inter-screen prediction error representing the difference between the original image 10 and the inter-screen prediction image 14 are generated.
  • the inter-screen prediction information D2 includes an inter-screen prediction block type indicating a block size when inter-screen prediction is performed, and motion vector information as a result of motion compensation.
  • the mode selection unit 130 selects a prediction mode (intra-screen prediction) according to the mode selection algorithm from the intra-screen prediction error output from the intra-screen prediction unit 110 and the inter-screen prediction error output from the inter-screen prediction unit 120. Or, prediction between screens) is determined.
  • a prediction mode intra-screen prediction
  • the intra-screen prediction image 13 is output as the predicted image 15.
  • inter-screen prediction image 14 is output as the predicted image 15.
  • the orthogonal transform unit 140 generates a frequency component D3 from the difference image 16 that is a difference between the original image 10 and the predicted image 15 by orthogonal transform processing.
  • the quantization unit 150 performs a quantization process from the quantization coefficient D4 output from the encoding control unit 190 and the frequency component D3 output from the orthogonal transform unit 140, and obtains a quantization value D5 with a reduced amount of information. Output.
  • the inverse quantization unit 160 performs an inverse quantization process on the quantization value D5 output from the quantization unit 150 to generate a restored frequency component D6.
  • the inverse orthogonal transform unit 170 performs an inverse orthogonal transform process on the restoration frequency component D6 output from the inverse quantization unit 160 to generate the restoration difference image 17.
  • the restored difference image 17 and the predicted image 15 selected by the mode selection unit 130 are added together and stored as a reconstructed image 12 in a storage device such as a memory.
  • a storage device such as a memory.
  • the inverse quantization unit 160 and the inverse orthogonal transform unit 170 may stop processing.
  • variable length encoding unit 180 encodes the quantized value D5 and the intra-screen prediction information D1 or the inter-screen prediction information D2 into a data string having a smaller data amount, and outputs the data stream 11 as a stream 11.
  • the stream buffer 200 acquires the stream 11 output from the variable-length encoding unit 180, buffers the stream 11, and then puts the stream 11 in a first-in first-out manner in a transmission path connected to the moving image encoding apparatus 1, or a subsequent decoder. Output in the method.
  • the code amount (accumulated amount) of the buffered stream 11 is output to the encoding control unit 190 as code amount information D7.
  • the encoding control unit 190 acquires the code amount information D7 from the stream buffer 200, and calculates the quantization coefficient D4 according to the rate control algorithm. Further, if the region determination result D13 (described later) of the current MB (macroblock, the same applies hereinafter) output from the reference control unit 210 indicates “reference region”, the region is left with the quantization coefficient unchanged. If the determination result D13 indicates “non-reference region”, the offset value is added to the quantization coefficient (or quantization parameter), and the result is output to the quantization unit 150.
  • the “reference area” is an area that is permitted to be referred to by others, that is, when performing a macroblock prediction process in inter-screen prediction or intra-screen prediction, inter-screen prediction is searched for inter-screen reference.
  • An area that is a range means an area that is the target of pixel reference within the screen in intra prediction, and ⁇ non-reference area '' is an area that is prohibited from being referred to by others.
  • An area that is not used as a search range refers to an area that is not subject to pixel reference within the screen in the intra prediction.
  • the reference control unit 210 determines the positions of the reference area and the non-reference area for each picture, holds the position management information, and includes the current MB, which is the macro block targeted for prediction processing, in either the non-reference area or the reference area Judge whether or not Then, according to the determination result, an area determination result D13 indicating whether the current MB is included in the reference area or the non-reference area is output to the encoding control unit 190. Further, based on the managed reference area information, the intra-screen prediction restriction information D11 is provided to the intra-screen prediction unit 110, and the inter-screen prediction restriction information D12 is provided to the inter-screen prediction unit 120. This will be described in detail later.
  • the reference control unit 210 which is a feature of the video encoding apparatus according to the present embodiment, will be described in detail below with reference to the drawings.
  • FIG. 2 is a diagram for explaining an example of a reference area and non-reference area management method for each picture executed by the reference control unit 210 of the video encoding device 1 according to the present embodiment.
  • the reference control unit 210 divides each picture into left and right, determines one as a reference area and the other as a non-reference area, and performs control for switching each picture.
  • the reference control unit 210 sets the left half of the encoding target picture 403 as a reference area and the right half as a non-reference area in the encoding target picture 403 that is the first picture in time series (display order; the same applies hereinafter). In the subsequent pictures (pictures 402 and 401), the area that was the non-reference area in the previous picture in time series is set as the reference area, and the area that was the reference area is set as the non-reference area. Also, the reference control unit 210 holds position management information indicating the position (left half or right half) of the reference area or non-reference area for each picture for the current and past several pictures. Then, based on this position management information, it is determined whether the current MB being processed is included in the reference region or the non-reference region, and the determination result is output to the encoding control unit 190 as a region determination result D13.
  • this specific example is characterized in that the left and right reference areas and the non-reference areas are switched for each picture, and the reference area and the non-reference area in the first picture 403 may be reversed (that is, The right half of the picture 403 may be a reference area and the left half may be a non-reference area).
  • FIG. 3 is a diagram for explaining the code amount per macro block line (hereinafter referred to as “MBL”) in the encoding target picture.
  • FIG. 4 is a diagram showing a change in code amount for each MBL.
  • the moving image encoding process is performed on the encoding target picture 401 in MB units according to the raster scan order indicated by the arrows.
  • the left half of the encoding target picture 401 is a reference area and the right half is a non-reference area, a non-reference area with a small amount of code and a reference area with a large amount of code within 1 MBL
  • the area included in is always constant in any MBL (see FIG. 3). Therefore, in the case of this specific example, as shown in FIG. 4, the code amount is almost the same in 1 MBL units, and the fluctuation between pictures is small, so that the code amount fluctuation can be suppressed to within 1 MBL unit.
  • the encoding target picture is divided into left and right, but the present invention is not limited to this. If the ratio of the reference area and the non-reference area included in every fixed number of MBs is always the same, and the non-reference area and the reference area are switched at regular intervals, the “repetitive illusion of sharp images” can be used. It is possible to realize a delay. Therefore, for example, a method of dividing a picture diagonally, a method of dividing a picture into three or more regions, and a method of dividing by a stripe pattern may be used.
  • the delay of the code amount is suppressed and the delay is reduced while maintaining the image quality. Is possible.
  • the picture area is divided into a high-quality reference area and a low-quality non-reference area on the left and right sides of the picture, coding efficiency is obtained when pixel values in the non-reference area are referred to in intra prediction and inter prediction. Gets worse.
  • the reference control unit 210 provides intra-screen prediction restriction information D11 to the intra-screen prediction unit 110 for intra-screen prediction of the current MB and inter-screen prediction for inter-screen prediction so that image quality degradation does not propagate from the non-reference region.
  • the inter-screen prediction restriction information D12 is output to the unit 120.
  • inter-screen prediction restriction information D12 that the reference control unit 210 outputs to the inter-screen prediction unit 120 will be described.
  • the inter picture prediction restriction information D12 includes a reference picture number.
  • the “reference picture number” is a number indicating a reference image on which a motion vector search is performed in the inter-screen prediction process.
  • the inter-screen prediction process is a process of searching for a pixel block similar to the current MB within a search range centered on the position of the current MB or a position predicted therefrom.
  • 5 to 7 are diagrams conceptually illustrating an example of the inter-screen prediction restriction information D12 that the reference control unit 210 outputs to the inter-screen prediction unit 120 in this specific example.
  • the reference control unit 210 included the search range 41 of the encoding target picture 401 in the past in the coding order (that is, the left half region was the reference region). )
  • the picture number of the picture 403 which is the image temporally closest to the encoding target picture 401 is included and designated in the inter-screen prediction restriction information D12.
  • the coding is performed on the image in the coding order in the right half area in the past.
  • the picture number of the picture 402 that is the image temporally closest to the target picture 401 is specified by being included in the inter-screen prediction restriction information D12.
  • the left half area is referred to in the past.
  • the reference area of the picture 403 that is the most recent image that was the area (more specifically, the same area as the search area 41 in the reference area; the same applies to FIG. 15) and the nearest area in which the right half was the reference area in the past.
  • the picture numbers of both images are designated so as to refer to both the reference area of the picture 402 that is the image of the image (more specifically, the same area portion as the search area 41 in the reference area, which is the same in FIG. 15). .
  • the inter-screen prediction unit 120 that has received such inter-screen prediction restriction information D12 basically performs inter-screen prediction with reference to only the reference area of the designated picture number. Therefore, according to this specific example, the inter-picture prediction always uses the reference picture encoded with a small quantization coefficient. As a result, the inter-picture prediction accuracy is improved and the moving picture coding efficiency is also improved. To do.
  • the in-screen prediction restriction information D11 includes mode restriction information.
  • the reference control unit 210 sets the mode restriction information so that the intra-screen prediction unit 110 refers to the pixel value only from the reference region without referring to the pixel value from the non-reference region.
  • the “mode restriction information” will be described below.
  • FIG. 9 is a diagram for explaining an example of the restriction of the intra prediction mode in each positional relationship between the reference area and the current MB.
  • the reference control unit 210 has a non-reference area on the left side of the current picture 401a, a current MB in the reference area, and is adjacent to the right side of the non-reference area. Designates mode restriction information indicating that the use of the intra prediction mode referring to the pixel value on the left side is prohibited as the intra prediction restriction information D11 (FIG. 9C, “position A” column).
  • the reference control unit 210 does not specify mode restriction information.
  • the intra-screen prediction restriction information D11 may include information indicating that the prediction mode is not restricted.
  • FIG. 10 is a diagram for explaining another example of a method for managing a reference area and a non-reference area for each picture performed by the reference control unit 210.
  • the reference control unit 210 divides each picture into upper and lower parts, and performs control for switching each picture as a reference area and the other as a non-reference area.
  • the reference control unit 210 sets the upper half of the encoding target picture 503, which is the first picture in time series, as a reference area, the lower half as a non-reference area, and the subsequent pictures ( In the pictures 502 and 501), on the time axis, an area that was a non-reference area in the previous picture is set as a reference area, and an area that was a reference area is set as a non-reference area. Further, the reference control unit 210 holds position management information indicating the position of this reference area or non-reference area for the current and past several pictures. Then, based on this position management information, it is determined whether the current MB being processed is included in the reference region or the non-reference region, and the determination result is output to the encoding control unit 190 as a region determination result D13.
  • FIG. 11 is a diagram for explaining the code amount per picture in the picture to be encoded. Further, FIG. 12 is a diagram illustrating a change in code amount for each picture.
  • the moving image encoding process is performed on the encoding target picture 501 in MB units according to the raster scan order indicated by the arrows.
  • the code amount becomes almost the same code amount in one picture unit, and the code amount fluctuation can be suppressed to within one picture unit.
  • the non-reference area and the reference area are alternately displayed on the upper and lower sides of the encoding target picture. Can be used to maintain subjective image quality. Also, the amount of codes can be reduced.
  • This specific example is characterized in that the upper and lower reference areas and the non-reference areas are switched for each picture, and the reference area and the non-reference area in the first picture 503 may be reversed (that is, The lower half of the picture 503 may be a reference area, and the upper half may be a non-reference area).
  • the inter-screen prediction restriction information D12 output from the reference control unit 210 to the inter-screen prediction unit 120 includes a reference picture number, and the “reference picture number” It is a number indicating a reference image for performing motion vector search.
  • 13 to 15 are diagrams conceptually illustrating an example of the inter-screen prediction restriction information D12 that the reference control unit 210 outputs to the inter-screen prediction unit 120 in this specific example.
  • the reference control unit 210 refers to the upper half area in the past on the time axis.
  • the picture number of the picture 503 that is the image temporally closest to the encoding target picture 501 among the images that were the area is included and designated in the inter-screen prediction restriction information D12.
  • the encoding target of the images whose lower half area was the reference area in the past on the time axis is specified by being included in the inter-screen prediction restriction information D12.
  • the upper half area is referred to in the past.
  • the picture numbers of both images are designated so as to refer to both the reference area of the picture 503 which is the most recent image that was the area and the reference area of the picture 502 which is the most recent image whose lower half was the reference area in the past. .
  • the inter-picture prediction always uses the reference picture encoded with a small quantization coefficient.
  • the inter-picture prediction accuracy is improved and the moving picture coding efficiency is also improved. To do.
  • moving images generally have a large lateral movement.
  • the search range can be ensured horizontally longer than in the case of the specific example 1, and thus there is an advantage that the inter-screen prediction accuracy in the horizontal direction can be improved.
  • FIG. 16 is a diagram for explaining an example of the restriction of the intra prediction mode in each positional relationship between the reference area and the current MB.
  • the intra-screen prediction restriction information D11 output from the reference control unit 210 to the intra-screen prediction unit 110 includes mode restriction information.
  • the reference control unit 210 sets the upper pixel when the upper side of the encoding target picture 501a is a non-reference area and the current MB is adjacent to the lower side of the non-reference area.
  • the mode restriction information indicating that the in-screen prediction mode that refers to the value is prohibited is designated as the in-screen prediction restriction information D11 (FIG. 16B, “position C” column).
  • the reference control unit 210 does not specify mode restriction information.
  • the intra-screen prediction restriction information D11 may include information indicating that the prediction mode is not restricted.
  • the prediction mode in the horizontal direction of the picture is not limited in the intra prediction mode, there is an advantage that the prediction accuracy is improved compared to the above specific example 1 for an image having a horizontal movement. .
  • FIG. 17 is a diagram for explaining another example of a method for managing a reference area and a non-reference area for each picture performed by the reference control unit 210.
  • the reference control unit 210 divides each picture into three in the horizontal direction, and performs control for switching each of the pictures as a reference area and the remaining one as a non-reference area for each picture. Do.
  • the reference control unit 210 sets the right area of the encoding target picture 604 that is the first picture in time series as a non-reference area, the center and the left as reference areas, and the next picture In the following (pictures 603, 602, 601), if the area that was the non-reference area in the previous picture is right, the area is the center, left if it is the center, and if it is left, the right area is the non-reference area.
  • the reference control unit 210 refers to the current MB being processed based on the position management information indicating the position (left, center, right) of this reference area or non-reference area, as in the first specific example and the second specific example.
  • the region determination result D13 indicating whether the region is included in the region or the non-reference region is output to the encoding control unit 190.
  • the picture is divided into three parts, one of which is a non-reference area and two of which are reference areas, and the non-reference areas are moved in order of right, center, and left in time series.
  • the present invention is not limited to this.
  • the number of divisions and the order of movement of the non-reference areas may be other methods as long as the setting pattern of the non-reference areas and the reference areas is repeated in a predetermined cycle in time series.
  • each picture may be divided into more regions. This period does not need to be linked with the GOP period (IDR frame insertion interval).
  • FIG. 18 is a diagram illustrating the code amount per 1 MBL in the encoding target picture.
  • the moving image encoding process is performed on the encoding target picture 601 in MB units according to the raster scan order indicated by the arrows.
  • the proportion of non-reference areas with low image quality in a picture is narrower than in the first specific example, and when played back as a moving image, the interval at which non-reference areas appear at the same location in each picture is It becomes longer than the specific example 1. Therefore, even when the frame rate is as low as 30 fps, it is possible to receive the effect of improving the subjective image quality of the “repetitive illusion of sharp images”.
  • FIGS. 19 and 20 are diagrams conceptually illustrating an example of the inter-screen prediction restriction information D12 that the reference control unit 210 outputs to the inter-screen prediction unit 120 in this specific example.
  • the reference control unit 210 encodes the encoding target picture among the images whose left area was the reference area in the past on the time axis.
  • the picture number of the image temporally closest to 501 is specified by being included in the inter-screen prediction restriction information D12.
  • the search range 61 exists in the center area of the encoding target picture 601, the time for the encoding target picture 601 in the image whose center area has been the reference area in the past on the time axis is temporal.
  • the picture number of the closest image is included and specified in the inter-screen prediction restriction information D12.
  • the search range 61 exists in the right region of the encoding target picture 601. That is, the picture number of the image temporally closest to the encoding target picture 601 among the images whose right region has been the reference region in the past on the time axis is specified by being included in the inter-screen prediction restriction information D12.
  • the search range 61 extends over a plurality of areas such as the left and center of the encoding target picture 601 or the center and right
  • the picture number of the image is included in the inter-screen prediction restriction information D12 so as to refer to the reference area of the image in which the area at the same position as the position where the search range extends is the reference area.
  • the search range 61 extends over the left and center areas of the encoding target picture 601, but in this case, the picture that is the latest image in which the left area was the reference image in the past. Both the reference image 603 and the picture 602 that is the most recent image whose central area was the reference area in the past may be used.
  • the reference control unit 210 specifies the picture numbers of the pictures 602 and 603 by including them in the inter-screen prediction restriction information D12.
  • inter-picture prediction always uses a reference image encoded with a small quantization coefficient.
  • inter-picture prediction accuracy is improved and moving picture coding is performed. Efficiency is also improved.
  • the restriction on intra prediction is the same as in the first specific example. That is, when the current MB is adjacent to the non-reference area (for example, as shown in FIG. 20), the reference control unit 210 sets the mode restriction information so as not to refer to the pixel value from the non-reference area, and the intra-screen prediction restriction. Designated as information D11.
  • a sharp image and a sharp image are mixed in a moving image that is encoded and decoded according to the specific example described above.
  • the “repetitive illusion of a sharp image” is effective for human vision, but can be a disturbance to image processing such as image recognition.
  • Periodic sharp changes can be detected on the decoding side by observing the code amount or SAD value per macroblock, and it can be estimated whether it is a reference area or a non-reference area, but position management on the encoding side Information cannot be completely restored, and an extra processing load is generated.
  • transcoding which is once decoded and re-encoded with another method when it is desired to change the encoding method, is widely performed, but once a sharply repeated moving image material maintains its sharp pattern. If transcoding is not performed, the image quality is significantly degraded.
  • variable length encoding unit 180 encodes information (position management information) for distinguishing between a reference area and a non-reference area, and user information (for example, H.264 / AVC encoding). SEI (supplemental enhancement information) message) in the stream and output from the stream buffer. Then, the decoding device that receives the stream decodes the information and provides it to downstream image processing.
  • FIG. 24 is a configuration diagram of the transcoder according to the present embodiment.
  • the moving picture decoding apparatus 2 is the same as the H.264 standard. This is a general decoder that decodes an H.264 / AVC encoded stream and sequentially outputs image data. The motion information (motion vector) used for motion compensation and the position management information decoded from the stream are output together.
  • the moving picture decoding apparatus 1 ′ is almost the same as the moving picture decoding apparatus 1 described in the first to third specific examples.
  • the reference control unit 210 of the video decoding device 1 arbitrarily determines the positions of the reference region and the non-reference region, whereas the reference control unit 210 ′ of the video encoding device 1 ′ of the present example.
  • the position management information received from the video decoding device 2 is followed to determine the position.
  • the motion information from the video decoding device 2 is provided to the inter-screen prediction unit 120 ′, which is useful for reducing the amount of motion search processing.
  • the moving picture decoding apparatus 1 ′ and the moving picture decoding apparatus 2 have been described as being integrated so that data can be transferred through a memory or the like.
  • the position management information can be included in the ancillary data (auxiliary data).
  • a moving image encoding program for causing a computer to execute a moving image encoding method for encoding a moving image including a plurality of images arranged in time series by intra prediction or inter prediction.
  • the plurality of images are divided into a reference region that is a region that refers to a pixel value in the intra prediction or inter-screen prediction, and a non-reference region that is a region that does not refer to a pixel value, and the time series
  • a first processing step for determining the reference region and the non-reference region in the plurality of images so that the reference region and the non-reference region in the plurality of images are switched at a predetermined interval set forth above;
  • a moving image encoding program for executing a process including: a second processing step of performing quantization on the non-reference area with a coarser width than the reference area when encoding the plurality of images .
  • the present invention relates to MPEG (trademark), VC-1 (SMPTE 421M), H.264, etc.
  • the present invention can be applied to a moving picture coding system based on predictive coding such as H.265 / HEVC, and includes a video recording apparatus including a digital video camera using such a moving picture coding system, a video transmission system, a television studio device, a video phone / conference It can be used in a device or software for a personal computer that realizes these functions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un dispositif de codage vidéo pour réaliser un traitement de codage de haute qualité et de latence faible par la suppression des variations dans la quantité de code tout en améliorant également la qualité d'image subjective. Un dispositif de codage vidéo pour coder une vidéo comprend une pluralité d'images alignées côte à côte dans une série temporelle par une prévision intra-trame ou une prévision inter-trame, la pluralité d'images étant segmentée en une région de référence, qui est une région pour référencer des valeurs de pixels dans la prévision intra-trame ou dans la prévision inter-trame, et une région de non référence, qui est une région où les valeurs de pixels ne sont pas référencées, et la région de référence et la région de non référence dans la pluralité d'images étant déterminées de sorte que la région de référence et la région de non référence dans la pluralité des images alternent avec un intervalle prédéterminé qui est établi à l'avance sur la série temporelle. Pendant le codage de la pluralité d'images, la région de non référence est quantifiée par des étapes plus grossières que la région de référence.
PCT/JP2012/078784 2011-11-17 2012-11-07 Dispositif de codage vidéo WO2013073422A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011-251879 2011-11-17
JP2011251879A JP2013110466A (ja) 2011-11-17 2011-11-17 動画像符号化装置

Publications (1)

Publication Number Publication Date
WO2013073422A1 true WO2013073422A1 (fr) 2013-05-23

Family

ID=48429485

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/078784 WO2013073422A1 (fr) 2011-11-17 2012-11-07 Dispositif de codage vidéo

Country Status (2)

Country Link
JP (1) JP2013110466A (fr)
WO (1) WO2013073422A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111009208B (zh) * 2019-12-27 2021-07-06 Tcl华星光电技术有限公司 面板显示的优化方法、显示面板及存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0723392A (ja) * 1993-06-16 1995-01-24 Mitsubishi Electric Corp 映像信号符号化復号化方式
JPH0870458A (ja) * 1994-08-29 1996-03-12 Sanyo Electric Co Ltd 符号量制御方法
JP2002064829A (ja) * 2000-08-23 2002-02-28 Kddi Research & Development Laboratories Inc 動画像の量子化制御装置
JP2005236459A (ja) * 2004-02-18 2005-09-02 Nippon Hoso Kyokai <Nhk> 動画像符号化装置、その方法及びそのプログラム
JP2011035444A (ja) * 2009-07-29 2011-02-17 Sony Corp 画像処理装置及び画像処理方法

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3934568B2 (ja) * 2003-03-04 2007-06-20 松下電器産業株式会社 動画符号化方法および装置
JP5238523B2 (ja) * 2009-01-13 2013-07-17 株式会社日立国際電気 動画像符号化装置、動画像復号化装置、および、動画像復号化方法
JP5067471B2 (ja) * 2010-11-29 2012-11-07 株式会社日立製作所 復号化方法、復号化装置、及びプログラム

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0723392A (ja) * 1993-06-16 1995-01-24 Mitsubishi Electric Corp 映像信号符号化復号化方式
JPH0870458A (ja) * 1994-08-29 1996-03-12 Sanyo Electric Co Ltd 符号量制御方法
JP2002064829A (ja) * 2000-08-23 2002-02-28 Kddi Research & Development Laboratories Inc 動画像の量子化制御装置
JP2005236459A (ja) * 2004-02-18 2005-09-02 Nippon Hoso Kyokai <Nhk> 動画像符号化装置、その方法及びそのプログラム
JP2011035444A (ja) * 2009-07-29 2011-02-17 Sony Corp 画像処理装置及び画像処理方法

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MASATOSHI TAKADA ET AL.: "Development on ultra-low delay video codec for broadcasting system", HITACHI KOKUSAI ELECTRIC TECHNICAL JOURNAL, 19 March 2010 (2010-03-19), pages 29 - 35 *
TAKAHIKO FUKINUKI: "Repetition of Sharp/Blurred TV Pictures and Its Application to Frame Interpolation (TFI) : Extension of Signal Processing of Visual Perception", THE JOURNAL OF THE INSTITUTE OF IMAGE INFORMATION AND TELEVISION ENGINEERS, vol. 63, no. 4, 1 April 2009 (2009-04-01), pages 549 - 552 *

Also Published As

Publication number Publication date
JP2013110466A (ja) 2013-06-06

Similar Documents

Publication Publication Date Title
KR100667806B1 (ko) 영상 부호화 및 복호화 방법 및 장치
JP4908180B2 (ja) 動画像符号化装置
US8630347B2 (en) Video decoding apparatus and video decoding method
JP4755093B2 (ja) 画像符号化方法および画像符号化装置
US20070199011A1 (en) System and method for high quality AVC encoding
US20070025444A1 (en) Coding Method
US9307241B2 (en) Video encoding method and a video encoding apparatus using the same
JP4875007B2 (ja) 動画像符号化装置、動画像符号化方法、及び、動画像復号化装置
US20070098078A1 (en) Method and apparatus for video encoding/decoding
JP5137687B2 (ja) 復号装置及び復号方法、プログラム
JP4774315B2 (ja) 画像復号化装置及び画像復号化方法
US20150016518A1 (en) Video Decoding with Reduced Complexity Deblocking
EP3057319A1 (fr) Décodeur vidéo à résolution adaptative
JP7343817B2 (ja) 符号化装置、符号化方法、及び符号化プログラム
JP5529161B2 (ja) ビデオストリームをブラウジングする方法及び装置
US20070014364A1 (en) Video coding method for performing rate control through frame dropping and frame composition, video encoder and transcoder using the same
JP2011250400A (ja) 動画像符号化装置及び動画像符号化方法
US9930352B2 (en) Reducing noise in an intraframe appearance cycle
WO2011108146A1 (fr) Dispositif de codage d&#39;image, système de codage/décodage d&#39;image, procédé de codage d&#39;image, et procédé d&#39;affichage d&#39;image
WO2013073422A1 (fr) Dispositif de codage vidéo
JP2009218965A (ja) 画像処理装置、それを搭載した撮像装置、および画像再生装置
WO2012165428A1 (fr) Dispositif de codage vidéo
JP5656575B2 (ja) 画像符号化装置
KR102674037B1 (ko) 화상 부호화 장치, 화상 부호화 방법, 화상 복호 장치, 화상 복호 방법, 및 컴퓨터 프로그램
KR20170095047A (ko) 동적 프레임 삭제 장치 및 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12848816

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12848816

Country of ref document: EP

Kind code of ref document: A1