CN106331710B - Image processing apparatus and method - Google Patents

Image processing apparatus and method Download PDF

Info

Publication number
CN106331710B
CN106331710B CN201610811282.2A CN201610811282A CN106331710B CN 106331710 B CN106331710 B CN 106331710B CN 201610811282 A CN201610811282 A CN 201610811282A CN 106331710 B CN106331710 B CN 106331710B
Authority
CN
China
Prior art keywords
unit
image
prediction
intra prediction
neighborhood pixels
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610811282.2A
Other languages
Chinese (zh)
Other versions
CN106331710A (en
Inventor
佐藤数史
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Verona Media International Ltd
Original Assignee
Verona Media International Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2009244753A external-priority patent/JP5597968B2/en
Application filed by Verona Media International Ltd filed Critical Verona Media International Ltd
Publication of CN106331710A publication Critical patent/CN106331710A/en
Application granted granted Critical
Publication of CN106331710B publication Critical patent/CN106331710B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The present invention relates to image processing apparatus and method that noise remove can be executed according to image and bit rate.Low-pass filtering setting unit 93 corresponds to the filter factor of intraprediction mode information and quantization parameter from the filter factor setting being stored in built-in filter factor memory 94.Adjacent image setting unit 81 is filtered using the filter factor being arranged by low-pass filtering setting unit 93 come the neighborhood pixels value to the current block from frame memory 72.Forecast image generates unit 82 and executes intra prediction using the neighborhood pixels value by filtering processing from adjacent image setting unit 81, and generates forecast image.Present invention may apply to use such as H.264/AVC picture coding device of said shank.

Description

Image processing apparatus and method
The application be same applicant the applying date be on June 23rd, 2010, application No. is 201080028028.6 (PCT/JP2010/060605), the division Shen of the Chinese invention patent application of entitled " image processing apparatus and method " Please.
Technical field
The present invention relates to a kind of image processing apparatus and methods, more particularly to a kind of image that can be improved forecasting efficiency Processing unit and method.
Background technique
In recent years, the device for carrying out compressed encoding to image by following manner is widely used: utilizing for by image Coded format of the information as Digital Signal Processing, also, the distinctive redundancy of image information (redundancy) is utilized, In, at this time to send and store efficient information as target, mended by the orthogonal transformation and movement of discrete cosine transform etc. It repays to compress image.The example of the coding method includes MPEG (motion characteristics planning) etc..
Particularly, MPEG2 (ISO/IEC 13818-2) is defined as general image coded format, also, being includes interlacing The standard of both scan image and sequential scan image and standard-resolution image and HD image.For example, MPEG2 is existing It is utilized with being widely used in the wide scope for having been used for professional purpose and consumer use.For example, by being pressed using MPEG2 Contracting format, in the case where the horizontally interlaced image of the standard resolution with 720 × 480 pixels, distribution 4 is to 8Mbps's Size of code (bit rate).In addition, for example, by utilizing MPEG2 compressed format, in the high-resolution with 1920 × 1088 pixels In the case where the horizontally interlaced image of rate, the size of code (bit rate) of distribution 18 to 22Mbps.It is thereby achieved that high compression rate With excellent picture quality.
For MPEG2, mainly to be suitble to the high image quality of broadcast purposes as target, still, do not handle than MPEG1's The low size of code of size of code (bit rate), that is, the coded format of compression ratio with higher.It is general with personal digital assistant And it is expected that the demand to this coded format will be increased from now on, also, in response to this, have been carried out The standardization of MPEG4 coded format.About coding format, specification is confirmed to be international standard ISO/ in December, 1998 IEC 14496-2。
In addition, in recent years, having passed through the originally estimated image coding for video conference purposes and be used as H.26L The standardization of the standard of (ITU-T Q6/16 VCEG).For H.26L, it is known that, the tradition with such as MPEG2 or MPEG4 Coded format is compared, although requiring bigger calculation amount for its coding and decoding, realizes higher code efficiency.This Outside, currently, as movable a part of MPEG4, in order to realize higher code efficiency, have been carried out also using not by H.26L support but by this H.26L based on function standardization, as enhancing compression Video coding conjunctive model.Make For standardized timetable (schedule), H.264 with MPEG-4 Part10 (advanced video coding, hereinafter referred to as H.264/AVC) become international standard in March, 2003.
In addition, having been completed that the business including such as RGB, 4:2:2 or 4:4:4 made at 2 months 2005 as its extension With required encoding tool, 8x8DCT and quantization matrix as defined in MPEG-2 FRExt (fidelity range extension, Fidelity Range Extension) standardization.Therefore, H.264/AVC have become or even can suitably express electricity The coded format for the film noise for including in shadow, also, have been used for such as Blu-ray disc (Blu-Ray Disc, registered trademark) Deng be widely applied.
But now, the demand for further high compression coding has been increased, for example, anticipated compression has about The image of 4000 × 2000 pixels, that is, four times of high visual pattern.Alternatively, having increased for further high compression The demand of coding, such as, it is contemplated that high visual pattern is issued in the environment with limited transmission capacity as internet. Therefore, it for the above-mentioned VCEG (=visual coding expert group) under the control of ITU-T, has been consecutively carried out and has improved volume The related research of code efficiency.
Now, why H.264/AVC the format that can be provided realizes high volume compared with traditional MPEG2 format etc. A kind of factor of code efficiency is that intra-frame prediction method (intra prediction method) is utilized.
For intra-frame prediction method, about luminance signal, it is determined that four kind of 16 × 16 pixel macroblock unit and nine kind 4 × The intra prediction mode of 4 pixels and 8 × 8 pixel block units.About color difference signal, it is determined that four kind of 8 × 8 pixel block unit Intra prediction mode.Intra prediction mode for color difference signal can be independent with the intra prediction mode for luminance signal Ground is set.
On how to show the residual error (residual) after this intra prediction, for each intra prediction mode There are specific pattern (pattern).
As the method eliminated this redundancy and further increase code efficiency, non-patent literature 1 proposes following Method.
That is, in advance using the training signal of processed offline by the way that normally H.264/AVC format executes in frame Image coded treatment (intra image encoding processing), about each piece, for each intra prediction mould Formula executes the orthogonal transformation of Karhunen-Lo é ve transformation etc., also, calculates optimal transformation coefficient.
Then, in actual coded treatment, following processing are used: being converted using by Karhunen-Lo é ve above-mentioned For the orthogonal transform coefficient of each model-based optimization, rather than the orthogonal transformation as defined in H.264/AVC format.
In addition, non-patent literature 2 is proposed intra prediction above-mentioned and inter-prediction (inter prediction) group The method of conjunction.
That is, for non-patent literature 2, about the motion vector information obtained in inter-prediction, not only for Current block, and for the neighborhood pixels value around current block, produce difference information.It is being generated in this manner about working as Preceding piece of difference information and about intra prediction is executed between the difference information of neighborhood pixels, generates second order difference information.Then, Orthogonal transformation and quantization are carried out to the second order difference information of generation, and it is exported together with compression image in downstream.
Therefore, code efficiency is further improved.
In addition, as described above, macroblock size is 16 × 16 pixels with H.264/AVC format.But for that will lead to Cross the big picture frame of such as UHD (ultrahigh resolution: 4000 × 2000 pixels) of next-generation coding method processing, 16 × 16 The macroblock size of pixel is not optimal.
Thus, for example, non-patent literature 3 etc. proposes the size that macroblock size is enlarged into 32 × 32 pixels.
Quotation list
Non-patent literature
Non-patent literature 1: " Improved Intra Coding ", VCEG-AF15, ITU-Telecommunications Standardization Sector STUDY GROUP Question 6 Video coding Experts Group (VCEG),20-21 April 2007
Non-patent literature 2: " Second Order Prediction (SOP) in P Slice ", Sijia Chen, Jinpeng Wang,Shangwen Li and,Lu Yu,VCEG-AD09,ITU-Telecommunications Standardization Sector STUDY GROUP Question 6Video coding Experts Group (VCEG),16-18July 2008
Non-patent literature 3: " Video Coding Using Extended Block Sizes ", VCEG-AD09, ITU- Telecommunications Standardization Sector STUDY GROUP Question 16- Contribution 123,Jan 2009
Summary of the invention
Technical problem
Now, for H.264/AVC format, using the block of 8 × 8 above-mentioned pixels as increment execute intra prediction it Before, low-pass filtering treatment is executed to the pixel value of neighborhood pixels.Therefore, the noise for including in neighborhood pixels is eliminated, also, is mentioned High correlation, so as to realize higher code efficiency.
But the degree of the noise regardless of being included is with input picture, quantization parameter value, intra prediction mode etc. Difference all secures the low-pass filter for removing noise with H.264/AVC format.That is, being schemed according to input Picture, quantization parameter value, intra prediction mode etc., the low-pass filter are not still optimal.
In addition, being only increment intra prediction mode to above-mentioned using the block of 8 × 8 pixels for H.264/AVC format The pixel values of neighborhood pixels perform noise remove, also, for other modes, this is still not applicable.
The present invention is made in view of the situation, also, the present invention realizes noise corresponding with image and bit rate and goes It removes, to improve forecasting efficiency.
Solution to problem
Image processing apparatus according to the first aspect of the invention includes: decoding device, is configured as to being used as decoding The image of the current block of the object of processing is decoded;Filter set parts, be configured as according to coding parameter be arranged with it is described Corresponding, for the current block the intra prediction of current block and for pass through the neighborhood pixels of the current block The coefficient of filtering processing;And intra prediction component, it is configured as: using the system by the filtering set parts setting Number makes the neighborhood pixels by the filtering processing, and executes intra prediction to the current block.
Coding parameter may include the mode of the intra prediction of current block or the quantization parameter of current block;Decoding device can It is decoded with the quantization parameter of the mode of the intra prediction to the current block or the current block;And the filtering is set Setting component can be according to the mode by the decoded intra prediction of the decoding device or by the decoded quantization of the decoding device Coefficient described in parameter setting.
Image processing apparatus can also include: filter factor storage unit, be configured as storing the coefficient;Wherein, institute Coefficient is stated to be calculated to that study image is used to obtain as the Minimum Residual between the study block and forecast image of the object of coded treatment Difference, also, the coefficient is stored in the filter factor storage unit, and the forecast image is by the study block It executes intra prediction corresponding with the mode of the intra prediction of the study block or the study quantization parameter of block and obtains ?;And the filtering set parts will be stored in the coefficient in the filter factor storage unit and institute The corresponding coefficient of quantization parameter for stating the mode or the current block of the intra prediction of current block is provided as the coefficient.
(wherein, the coefficient can be remained n according to the register capacity of processor by the filter factor storage unit N is integer) bit value.
The decoding device can be decoded the coefficient, and use is by the filtering processing at coding side The preceding neighborhood pixels calculate the coefficient, so that obtaining about by the intra prediction executed for the current block The residual error of forecast image is the smallest, also, with the mode of the current block and the intra prediction of the current block or described The quantization parameter of current block is correspondingly provided with the coefficient;And the filtering set parts can will be by the decoding device It is in the decoded coefficient, opposite with the quantization parameter of the mode of the intra prediction of the current block or the current block The coefficient answered is provided as the coefficient.
The coefficient can be made of filter factor and deviant.
The filter factor can be made of three taps.
The filter factor can have the symmetry centered on the coefficient for corresponding to 0 phase.
The decoding device can be decoded the mode of the intra prediction of the current block;And
It is described filtering set parts can by the mode of intra prediction vertical mode or horizontal pattern be taken as the first kind Not, and other modes can be taken as second category, belongs to the first kind in the mode of the intra prediction of the current block In the case where not, the coefficient corresponding with the first category is can be set in the filtering set parts, is solved In the case where the mode of the intra prediction of the current block of code belongs to the second category, the filtering set parts can be set Set the coefficient corresponding with the second category.
Image processing apparatus can also include: receiving part, be configured as receive instruction for the neighborhood pixels whether Execute the flag information of the filtering processing;Wherein, based on the flag information received by the receiving part, the filtering is set Set whether component setting executes the filtering processing for the neighborhood pixels.
The receiving part can be using macro block as flag information described in Incremental receipt;And it is based on being connect by the receiving part The flag information received, the filtering set parts can be set for whether the neighborhood pixels execute the filtering processing.
The receiving part can be using block as flag information described in Incremental receipt;And it is based on being received by the receiving part The flag information arrived, whether it is increment to the neighborhood pixels execution filtering using block that the filtering set parts can be set Processing.
Image processing method according to the first aspect of the invention includes the following steps: that image processing apparatus is executed to making Image for the current block of the object of decoding process is decoded;It is used for according to coding parameter setting is corresponding with the current block The coefficient for the filtering processing to carry out to the neighborhood pixels of the current block, the neighborhood pixels are in the frame of the current block Prediction;And using the coefficient being set, the filtering processing is carried out to the neighborhood pixels, and to described current Block executes intra prediction.
Image processing apparatus according to the second aspect of the invention includes: filtering set parts, is configured as according to coding The coefficient of filtering processing of the parameter setting for be carried out to the neighborhood pixels of the current block of the object of decoding process to be used as, institute State intra prediction of the neighborhood pixels for the current block;Intra prediction component, is configured as: use is by the filtering setting unit The coefficient of part setting, makes the neighborhood pixels by the filtering processing, and executes intra prediction to the current block; And addressable part, it is configured as encoding the image of the current block.
Coding parameter may include the mode of the intra prediction of current block or the quantization parameter of current block;And the filter Described in wave set parts can be arranged according to the quantization parameter of the mode of the intra prediction of the current block or the current block Coefficient;And addressable part can corresponding modes or the current block to the intra prediction of the current block quantization parameter It is encoded.
Image processing apparatus can also include: filter factor storage unit, be configured as storing the coefficient;Wherein, institute Coefficient is stated to be calculated to that study image is used to obtain as the Minimum Residual between the study block and forecast image of the object of coded treatment Difference, also, the coefficient is stored in the filter factor storage unit, and the forecast image is by the study block It executes intra prediction corresponding with the mode of the intra prediction of the study block or the study quantization parameter of block and obtains ?;And the filtering set parts will be stored in the coefficient in the filter factor storage unit and institute The corresponding coefficient of quantization parameter for stating the mode or the current block of the intra prediction of current block is provided as the coefficient.
Image processing apparatus can also include: filter factor calculating unit, be configured as calculating the coefficient, so that described Residual error between current block and forecast image is minimum, and the forecast image is by the mould according to the intra prediction of the current block The quantization parameter of formula or the current block, using the neighborhood pixels before carrying out the filtering processing to the study Block executes intra prediction and obtains;The filtering set parts will be by the calculated system of the filter factor calculating unit Coefficient setting in number, corresponding with the quantization parameter of the mode of the intra prediction of the current block or the current block As the coefficient.
The coefficient can be made of filter factor and deviant.
The filtering set parts can be set whether execute the filtering processing to neighborhood pixels;And the coding unit Part can be encoded to indicating whether to execute by the flag information of the filtering processing of the filtering set parts setting.
Image processing method according to the second aspect of the invention includes the following steps: that image processing apparatus is executed according to volume The coefficient of filtering processing of the code parameter setting for be carried out to the neighborhood pixels of the current block of the object of decoding process to be used as, The neighborhood pixels are used for the intra prediction of the current block;Using the coefficient being set, to the neighborhood pixels into The row filtering processing, and intra prediction is executed to the current block;And the image of the current block is encoded.
According to the first aspect of the invention, the image of the current block of the object of decoding process to be used as is decoded, root According to coding parameter, coefficient, filtering processing of the coefficient for carry out to the neighborhood pixels of current block, wherein described neighbouring are set Pixel is used for the intra prediction of current block.Then, neighborhood pixels are filtered using the coefficient being set, also, Intra prediction is executed to current block.
According to the second aspect of the invention, being used for according to coding parameter setting will working as to the object of decoding process to be used as The coefficient for the filtering processing that preceding piece of neighborhood pixels carry out, the neighborhood pixels are used for the intra prediction of current block, using The coefficient of setting is filtered neighborhood pixels, executes intra prediction to current block, also, carry out to the image of current block Coding.
It note that above-mentioned image processing apparatus can be independent device, alternatively, can be one image coding of composition The internal block of device or picture decoding apparatus.
Beneficial effects of the present invention
According to first invention, image can be decoded.In addition, according to the second invention, it can be according to image and bit Rate executes noise remove.
According to the second invention, image can be encoded.In addition, according to first invention, it can be according to image and bit Rate executes noise remove.
Detailed description of the invention
[Fig. 1] Fig. 1 is the block diagram of the configuration for the embodiment that diagram applies picture coding device of the invention.
[Fig. 2] Fig. 2 is the diagram for describing processing sequence in the case where 16 × 16 pixel intra prediction mode.
[Fig. 3] Fig. 3 is the diagram for illustrating the type of 4 × 4 pixel intra prediction modes of luminance signal.
[Fig. 4] Fig. 4 is the diagram for illustrating the type of 4 × 4 pixel intra prediction modes of luminance signal.
[Fig. 5] Fig. 5 is the diagram for describing the direction of 4 × 4 pixel intra predictions.
[Fig. 6] Fig. 6 is the diagram for describing 4 × 4 pixel intra predictions.
[Fig. 7] Fig. 7 is the diagram for the coding for describing 4 × 4 pixel intra prediction modes of luminance signal.
[Fig. 8] Fig. 8 is the diagram for illustrating the type of 8 × 8 pixel intra prediction modes of luminance signal.
[Fig. 9] Fig. 9 is the diagram for illustrating the type of 8 × 8 pixel intra prediction modes of luminance signal.
[Figure 10] Figure 10 is the diagram for illustrating the type of 16 × 16 pixel intra prediction modes of luminance signal.
[Figure 11] Figure 11 is the diagram for illustrating the type of 16 × 16 pixel intra prediction modes of luminance signal.
[Figure 12] Figure 12 is the diagram for describing 16 × 16 pixel intra predictions.
[Figure 13] Figure 13 is the diagram of the type of the intra prediction mode of graphical color difference signal.
[Figure 14] Figure 14 is the configuration example of the neighborhood pixels interpolation filtering switch unit in pictorial image 1 and intraprediction unit The block diagram of son.
[Figure 15] Figure 15 is the diagram for describing the calculating of filter factor.
[Figure 16] Figure 16 is the flow chart for describing the coded treatment of the picture coding device in Fig. 1.
[Figure 17] Figure 17 is the flow chart for describing the prediction processing in the step S21 in Figure 16.
[Figure 18] Figure 18 is the flow chart for describing the intra-prediction process in the step S31 in Figure 17.
[Figure 19] Figure 19 is the flow chart for describing the prediction processing of the intraframe motion in the step S32 in Figure 17.
[Figure 20] Figure 20 is that the neighborhood pixels interpolation filtering switch unit and the another of intraprediction unit in pictorial image 1 are matched Set the block diagram of example.
[Figure 21] Figure 21 is the process for describing another example of the intra-prediction process in the step S31 in Figure 17 Figure.
[Figure 22] Figure 22 is the block diagram of the configuration for the embodiment that diagram applies picture decoding apparatus of the invention.
[Figure 23] Figure 23 is the configuration of the neighborhood pixels interpolation filtering switch unit in pictorial image 22 and intraprediction unit The block diagram of example.
[Figure 24] Figure 24 is the flow chart for describing the decoding process of the picture decoding apparatus in Figure 22.
[Figure 25] Figure 25 is the flow chart for describing the prediction processing in the step S138 in Figure 24.
[Figure 26] Figure 26 is the another of neighborhood pixels interpolation filtering switch unit in pictorial image 22 and intraprediction unit The block diagram of configuration example.
[Figure 27] Figure 27 is the flow chart for describing another example of the prediction processing in the step S138 in Figure 24.
[Figure 28] Figure 28 is the block diagram of the configuration for the embodiment that diagram applies learning device of the invention.
[Figure 29] Figure 29 is the configuration of the neighborhood pixels interpolation filtering computing unit in pictorial image 28 and intraprediction unit The block diagram of example.
[Figure 30] Figure 30 is the flow chart for describing the intra-prediction process using learning device in Figure 28.
[Figure 31] Figure 31 is the block diagram of the configuration for another embodiment that diagram applies picture coding device of the invention.
[Figure 32] Figure 32 is the diagram for describing re prediction processing.
[Figure 33] Figure 33 is the block diagram of the configuration for another embodiment that diagram applies picture decoding apparatus of the invention.
[Figure 34] Figure 34 is the block diagram of the configuration for the another embodiment that diagram applies picture coding device of the invention.
[Figure 35] Figure 35 is the block diagram of the configuration example of the intraprediction unit in pictorial image 34.
[Figure 36] Figure 36 is the process for describing another example of the intra-prediction process in the step S31 in Figure 17 Figure.
[Figure 37] Figure 37 is the process for describing the another example of the intra-prediction process in the step S31 in Figure 17 Figure.
[Figure 38] Figure 38 is the process for describing another example of the intra-prediction process in the step S31 in Figure 17 Figure.
[Figure 39] Figure 39 is the block diagram of the configuration for the another embodiment that diagram applies picture decoding apparatus of the invention.
[Figure 40] Figure 40 is the configuration of the neighborhood pixels interpolation filtering control unit and intraprediction unit in pictorial image 39 The block diagram of example.
[Figure 41] Figure 41 is the flow chart for describing the another example of the prediction processing in the step S138 in Figure 24.
[Figure 42] Figure 42 is the block diagram of the configuration for another embodiment that diagram applies picture coding device of the invention.
[Figure 43] Figure 43 is the block diagram of the configuration for another embodiment that diagram applies picture decoding apparatus of the invention.
[Figure 44] Figure 44 is the diagram for illustrating the example of extension block size.
[Figure 45] Figure 45 is the block diagram for illustrating the configuration example of hardware of computer.
[Figure 46] Figure 46 is the block diagram for the main configuration example that diagram applies television receiver of the invention.
[Figure 47] Figure 47 is the block diagram for the main configuration example that diagram applies cellular phone of the invention.
[Figure 48] Figure 48 is the block diagram for the main configuration example that diagram applies hdd recorder of the invention.
[Figure 49] Figure 49 is the block diagram for the main configuration example that diagram applies camera of the invention.
Specific embodiment
The embodiment of the present invention is described hereinafter with reference to attached drawing.Note that be described in the following order.
1. first embodiment (neighborhood pixels interpolation filtering switching: the example of intra prediction)
2. second embodiment (neighborhood pixels interpolation filtering switching: the example of re prediction)
3. 3rd embodiment (neighborhood pixels interpolation filtering on/off (ON/OFF) control: the example of intra prediction)
4. fourth embodiment (neighborhood pixels interpolation filtering opens/closes control: the example of re prediction)
<1. first embodiment>
[configuration example of picture coding device]
Fig. 1 shows be used as the configuration for applying the embodiment of picture coding device of image processing apparatus of the invention.
The picture coding device 51 is used and for example H.264 (is hereinafter retouched with MPEG-4Part10 (advanced video coding) State as H.264/AVC) format come to image carry out compressed encoding.
For the example in Fig. 1, picture coding device 51 rearranges buffer 62, meter by A/D converting unit 61, picture Calculate unit 63, orthogonal transform unit 64, quantifying unit 65, lossless coding unit 66, storage buffer 67, inverse quantization unit 68, inverse orthogonal transformation unit 69, computing unit 70, de-blocking filter 71, frame memory 72, switch 73, intraprediction unit 74, Neighborhood pixels interpolation filtering switch unit 75, motion prediction/compensating unit 76, forecast image selecting unit 77 and rate control list Member 78 is constituted.
A/D converting unit 61 carries out Analog-digital Converter to input picture, and the input picture is output to picture weight It is new to arrange buffer 62 to store.Picture rearranges buffer 62 and will be used to show according to depositing according to GOP (picture group) The image of the frame of the sequence of storage is rearranged for the sequence of the frame for coding.
Computing unit 63 from rearranged by picture subtracted in the image that buffer 62 is read selected by forecast image it is single The forecast image from intraprediction unit 74 or the forecast image from motion prediction/compensating unit 76 of 77 selection of member, And its difference information is output to orthogonal transform unit 64.Orthogonal transform unit 64 is to the difference information from computing unit 63 The orthogonal transformation of such as discrete cosine transform, Karhunen-Lo é ve transformation or the like is carried out, and exports its transformation coefficient.Amount Change the transformation coefficient that unit 65 exports orthogonal transform unit 64 to quantify.
The quantization transform coefficient of output as quantifying unit 65 is input into lossless coding unit 66, in lossless volume The lossless coding and compression of such as variable length code, arithmetic coding or the like are carried out in code unit 66 to it.
Lossless coding unit 66 obtains the information of instruction intra prediction etc. from intraprediction unit 74, and pre- from movement Survey/compensating unit 76 obtains the information of instruction inter-frame forecast mode etc..It note that the information for hereinafter indicating intra prediction By referred to as intraprediction mode information.Similarly, hereinafter, indicate that the information of inter-prediction will be referred to as inter-prediction mould Formula information.
The transformation coefficient of 66 pairs of lossless coding unit quantizations encodes, also to the information of instruction intra prediction, instruction Information, quantization parameter of inter-frame forecast mode etc. are encoded, and take these information as the head information in compression image A part.Coded data is supplied to storage buffer 67 to store by lossless coding unit 66.
For example, executing the lossless of such as variable length code, arithmetic coding or the like for lossless coding unit 66 Coded treatment.The example of variable length code includes CAVLC (the context-adaptive variable length determined by H.264/AVC format Degree coding).The example of arithmetic coding includes CABAC (context adaptive binary arithmetic coding).
Storage buffer 67 is using the data supplied from lossless coding unit 66 as the pressure by H.264/AVC said shank Contract drawing picture is output to storage device or the transmitting path in downstream not shown in this Figure etc..
In addition, being also inputted to inverse quantization unit 68 from the quantization transform coefficient that quantifying unit 65 exports, inverse amount has been carried out Change, then, the further progress inverse orthogonal transformation at the inverse orthogonal transformation unit 69.It will have been carried out by computing unit 70 inverse orthogonal The output of transformation is added with the forecast image supplied from forecast image selecting unit 77, and is changed into the figure of local decoder Picture.Then de-blocking filter 71 is supplied to frame and deposits from removal block distortion (block distortion) in decoding image Reservoir 72, to store.Image before de-blocking filter 71 executes deblocking filtering processing is also supplied to frame memory 72, To store.
The reference picture being stored in frame memory 72 is output in motion prediction/compensating unit 76 or frame in advance by switch 73 Survey unit 74.
For example, rearranging I picture, B picture and the P figure of buffer 62 from picture for the picture coding device 51 Piece is supplied to intraprediction unit 74 as the image of intra prediction to be carried out (also referred to as processing in frame).In addition, from picture Face rearranges the figure of the B picture and P picture of the reading of buffer 62 as inter-prediction to be carried out (also referred to as interframe processing) As being supplied to motion prediction/compensating unit 76.
Intraprediction unit 74 based on from picture rearrange buffer 62 read the image that carry out intra prediction and The reference picture supplied from frame memory 72 executes the intra-prediction process of all candidate intra prediction modes, to generate prediction Image.
Before intra-prediction process, intraprediction unit 74 executes filtering processing, the neighborhood pixels to neighborhood pixels It is the intra prediction for each current block and with scheduled positional relationship and the neighbouring pixel of current block.According to pre- out of frame Survey the intra prediction mode etc. that unit 74 is supplied, which uses is arranged by neighborhood pixels interpolation filtering switch unit 75 Filter factor.That is, intraprediction unit 74 uses for the intra-prediction process of all candidate intra prediction modes The neighborhood pixels of the filtering processing using the filter factor being arranged by neighborhood pixels interpolation filtering switch unit 75 are carried out.
Intraprediction unit 74 calculates the cost function value (cost about the intra prediction mode for producing forecast image Function value), also, select wherein calculated cost function value give the intra prediction mode of minimum value as Best intra prediction mode.Intraprediction unit 74 is by the forecast image generated in best intra prediction mode and for corresponding The calculated cost function value of best intra prediction mode be supplied to forecast image selecting unit 77.
In the case where forecast image selecting unit 77 has selected the forecast image generated in best intra prediction mode, The information for indicating best intra prediction mode is supplied to lossless coding unit 66 by intraprediction unit 74.From intra prediction In the case where unit 74 transfers information, lossless coding unit 66 encodes the information, and takes the information as pressure A part of head information in contract drawing picture.
Neighborhood pixels interpolation filtering switch unit 75 stores by using training image in the Figure 28 being described later on It practises and executes study at device 251 and obtain, corresponding with quantization parameter and intra prediction mode filter factor.
The quantization parameter from Rate control unit 78 is supplied to neighborhood pixels interpolation filtering switch unit 75 and comes from frame The intraprediction mode information of interior prediction unit 74.Neighborhood pixels interpolation filtering switch unit 75 is arranged and comes from rate control list The quantization parameter filter factor corresponding with the intra prediction mode from intraprediction unit 74 of member 78.Neighborhood pixels interpolation It filters switch unit 75 and the filter factor of setting is supplied to intraprediction unit 74.
Note that neighborhood pixels interpolation filtering switch unit 75 can execute in quantization parameter and intra prediction mode The study and storage of only one rather than two corresponding filter factors.
, can in addition, although neighborhood pixels interpolation filtering switch unit 75 stores the filter factor of preparatory off-line learning Instead in line computation filter factor.In this case, the filtering being arranged by neighborhood pixels interpolation filtering switch unit 75 Coefficient is output to lossless coding unit 66 to be sent to decoding side, as shown in dotted arrows.
Motion prediction/compensating unit 76 executes motion prediction and compensation deals for all candidate inter-frame forecast modes. Specifically, it is supplied to motion prediction/compensating unit 76 from picture via switch 73 and rearranges carrying out for the reading of buffer 62 The image that interframe is handled and the reference picture from frame memory 72.Motion prediction/compensating unit 76 is based on carrying out at interframe The image and reference picture of reason detect the motion vector of all candidate inter-frame forecast modes, based on motion vector to reference picture Processing is compensated, and generates forecast image.
In addition, motion prediction/compensating unit 76 calculates cost function value for all candidate inter-frame forecast modes.Movement Prediction/compensating unit 76 determines that the prediction mode of the offer minimum value in calculated cost function value is pre- as best interframe Survey mode.
The forecast image and its cost function value that motion prediction/compensating unit 76 will generate in best inter-frame forecast mode It is supplied to forecast image selecting unit 77.It has selected to generate in best inter-frame forecast mode in forecast image selecting unit 77 In the case where forecast image, motion prediction/compensating unit 76 will indicate the information (inter-frame forecast mode of best inter-frame forecast mode Information) it exports to lossless coding unit 66.
It is single to note that motion vector information, flag information, reference frame information etc. are output to lossless coding as needed Member 66.Lossless coding unit 66 also carries out such as variable length code to the information from motion prediction/compensating unit 76, calculates The lossless coded treatment of art coding etc., and be inserted into the head of compression image.
Forecast image selecting unit 77 is based on the cost exported from intraprediction unit 74 or motion prediction/compensating unit 76 Functional value determines optimum prediction mode from best intra prediction mode and best inter-frame forecast mode.Forecast image selecting unit Then 77 select the forecast image in determining optimum prediction mode, and be supplied into computing unit 63 and 70.At this point, pre- The selection information of forecast image is supplied to intraprediction unit 74 or motion prediction/compensating unit by altimetric image selecting unit 77 76。
Rate control unit 78 is single based on the compression image quantization state modulator quantization being stored in storage buffer 67 The rate of the quantization operation of member 65, not will lead to overflow or underflow.
The quantization parameter for being used for rate control at quantifying unit 65 is supplied to lossless coding unit 66, to the amount Change parameter and carry out lossless coded treatment, and inserts it into the head of compression image.The quantization parameter is supplied to neighbouring Pixel interpolating filters switch unit 75, and is used to set up the filtering system of the filtering processing for apply to neighborhood pixels Number.
[according to the description of the intra-prediction process of H.264/AVC format]
Firstly, the intra prediction mode that description is determined by H.264/AVC format.
Firstly, the intra prediction mode that luminance signal will be described.For the intra prediction mode of luminance signal, determine in frame In 4 × 4 prediction modes, frame in 8 × 8 prediction modes and frame 16 × 16 prediction modes three systems.In the presence of for determining block list The mode of position, for each macro block Setting pattern.In addition, being directed to each macro block, face can be independently given with luminance signal Intra prediction mode is arranged in colour difference signal.
In addition, for each 4 × 4 pixel current block, can be set nine kinds in the case where intra-frame 4 * 4 forecasting model Prediction mode one of works as prediction mode.In addition, in frame in the case where 8 × 8 prediction mode, for each 8 × 8 pixel Current block can be set nine kinds of prediction modes and one of work as prediction mode.In addition, in frame the case where 16 × 16 prediction mode In, four kinds of prediction modes can be set to 16 × 16 pixel current macros and one of work as prediction mode.
It note that hereinafter, 16 × 16 prediction modes in 8 × 8 prediction modes and frame in intra-frame 4 * 4 forecasting model, frame It will also be known respectively as 4 × 4 pixel intra prediction modes, 8 × 8 pixel intra prediction modes and 16 × 16 pictures in due course Plain intra prediction mode.
For the example in Fig. 2, investing each piece of number 1 to 25 indicates that the bit stream sequence of its block (decodes the processing of side Sequence).It note that about luminance signal, macro block is divided into 4 × 4 pixels, also, executes the DCT of 4 × 4 pixels.Only exist In frame in the case where 16 × 16 prediction modes, as shown in 1 block, the DC ingredient of aggregation block generates 4 × 4 matrixes, also, Orthogonal transformation is also carried out to the matrix.
On the other hand, about color difference signal, macro block be divided into 4 × 4 pixels and execute 4 × 4 pixels DCT it Afterwards, as shown in block 16 and 17, the DC ingredient of aggregation block generates 2 × 2 matrixes, and also carries out orthogonal transformation to the matrix.
Note that, about 8 × 8 prediction mode in frame, this can be suitable only for following situations: where with high image quality or Higher image quality (high profile or a profile beyond this) carries out 8 × 8 orthogonal transformations to current macro.
Fig. 3 and Fig. 4 is nine kind of 4 × 4 pixel intra prediction mode (Intra_4 × 4_pred_ for showing luminance signal Mode diagram).Eight kinds of modes other than showing the mode 2 of average value (DC) prediction correspond respectively in Fig. 5 use number The direction that word 0,1,3 to 8 indicates.
Nine kinds of intra_4 × 4_pred_mode will be described referring to Fig. 6.For the example in Fig. 6, pixel a to p expression is wanted The pixel of the current block handled in frame is carried out, pixel value A to M indicates the pixel value for belonging to the pixel of contiguous block.Specifically, picture Plain a to p is the image to be processed that the reading of buffer 62 is rearranged from picture, also, pixel value A to M is to store from frame The pixel value for decoding image that device 72 reads and is referenced.
In the case where the intra prediction mode shown in figs. 3 and 4, the pixel value A for the pixel for belonging to contiguous block is used To M, the following predicted pixel values for generating pixel a to p.Here, pixel value be " available " indicate the pixel value be available and Not such reason: the pixel is located in the edge of picture frame or is not yet encoded.On the other hand, pixel value be " can not " indicate cause the pixel value to be disabled for this reason: the pixel be located in the edge of picture frame or Not yet it is encoded.
Mode 0 is vertical prediction mode, and is suitable only for the case where pixel value A to D is " available ".In this feelings Under condition, the predicted pixel values of pixel a to p are generated as following expressions (1):
Predicted pixel values=A of pixel a, e, i and m
Predicted pixel values=B of pixel b, f, j and n
Predicted pixel values=C of pixel c, g, k and o
Predicted pixel values=D of pixel d, h, l and p...(1)
Mode 1 is horizontal prediction mode, and is suitable only for the case where pixel value I to L is " available ".In this feelings Under condition, the predicted pixel values of pixel a to p are generated as following expressions (2):
Predicted pixel values=I of pixel a, b, c and d
Predicted pixel values=J of pixel e, f, g and h
Predicted pixel values=K of pixel i, j, k and l
Predicted pixel values=L of pixel m, n, o and p...(2)
Mode 2 is DC prediction mode, also, when pixel value A, B, C, D, I, J, K and L are entirely " available ", such as table Predicted pixel values are equally generated up to formula (3).
(A+B+C+D+I+J+K+L+4)>>3...(3)
In addition, generating prediction pixel as expression formula (4) when pixel value A, B, C and D are entirely " not available " Value.
(I+J+K+L+2)>>2...(4)
In addition, generating prediction pixel as expression formula (5) when pixel value I, J, K and L are entirely " not available " Value.
(A+B+C+D+2)>>2...(5)
It note that, when pixel value A, B, C, D, I, J, K and L are entirely " not available ", 128 are used as predicted pixel values.
Mode 3 is Diagonal_Down_Left (diagonal down-left) prediction mode, and be suitable only for pixel value A, B, C, the case where D, I, J, K, L and M are " available ".In this case, generating pixel a to p as following expressions (6) Predicted pixel values:
The predicted pixel values of pixel a=(A+2B+C+2) > > 2
The predicted pixel values of pixel b and e=(B+2C+D+2) > > 2
The predicted pixel values of pixel c, f and i=(C+2D+E+2) > > 2
The predicted pixel values of pixel d, g, j and m=(D+2E+F+2) > > 2
The predicted pixel values of pixel h, k and n=(E+2F+G+2) > > 2
The predicted pixel values of pixel l and o=(F+2G+H+2) > > 2
The predicted pixel values of pixel p=(G+3H+2) > > 2...(6)
Mode 4 is Diagonal_Down_Right (lower-right diagonal position) prediction mode, and be suitable only for pixel value A, B, C, the case where D, I, J, K, L and M are " available ".In this case, generating pixel a to p as following expressions (7) Predicted pixel values:
The predicted pixel values of pixel m=(J+2K+L+2) > > 2
The predicted pixel values of pixel i and n=(I+2J+K+2) > > 2
The predicted pixel values of pixel e, j and o=(M+2I+J+2) > > 2
The predicted pixel values of pixel a, f, k and p=(A+2M+I+2) > > 2
The predicted pixel values of pixel b, g and l=(M+2A+B+2) > > 2
The predicted pixel values a=(A+2B+C+2) of pixel c and h > > 2
The predicted pixel values of pixel d=(B+2C+D+2) > > 2...(7)
Mode 5 is Diagonal_Vertical_Right (right Vertical Diagonal) prediction mode, and is suitable only for pixel The case where value A, B, C, D, I, J, K, L and M are " available ".In this case, pixel is generated as following expressions (8) The predicted pixel values of a to p:
The predicted pixel values of pixel a and j=(M+A+1) > > 1
The predicted pixel values of pixel b and k=(A+B+1) > > 1
The predicted pixel values of pixel c and l=(B+C+1) > > 1
The predicted pixel values of pixel d=(C+D+1) > > 1
The predicted pixel values of pixel e and n=(I+2M+A+2) > > 2
The predicted pixel values of pixel f and o=(M+2A+B+2) > > 2
The predicted pixel values of pixel g and p=(A+2B+C+2) > > 2
The predicted pixel values of pixel h=(B+2C+D+2) > > 2
The predicted pixel values of pixel i=(M+2I+J+2) > > 2
The predicted pixel values of pixel m=(I+2J+K+2) > > 2.. (8)
Mode 6 is Horizontal_Down (lower level) prediction mode, and be suitable only for pixel value A, B, C, D, I, J, the case where K, L and M are " available ".In this case, the prediction picture of pixel a to p is generated as following expressions (9) Element value:
The predicted pixel values of pixel a and g=(M+I+1) > > 1
The predicted pixel values of pixel b and h=(I+2M+A+2) > > 2
The predicted pixel values of pixel c=(M+2A+B+2) > > 2
The predicted pixel values of pixel d=(A+2B+C+2) > > 2
The predicted pixel values of pixel e and k=(I+J+1) > > 1
The predicted pixel values of pixel f and l=(M+2I+J+2) > > 2
The predicted pixel values of pixel i and o=(J+K+1) > > 1
The predicted pixel values of pixel j and p=(I+2J+K+2) > > 2
The predicted pixel values of pixel m=(K+L+1) > > 1
The predicted pixel values of pixel n=(J+2K+L+2) > > 2...(9)
Mode 7 is Vertical_Left (lower vertical) prediction mode, and be suitable only for pixel value A, B, C, D, I, J, K, the case where L and M is " available ".In this case, the prediction picture of pixel a to p is generated as following expressions (10) Element value:
The predicted pixel values of pixel a=(A+B+1) > > 1
The predicted pixel values of pixel b and i=(B+C+1) > > 1
The predicted pixel values of pixel c and j=(C+D+1) > > 1
The predicted pixel values of pixel d and k=(D+E+1) > > 1
The predicted pixel values of pixel l=(E+F+1) > > 1
The predicted pixel values of pixel e=(A+2B+C+2) > > 2
The predicted pixel values of pixel f and m=(B+2C+D+2) > > 2
The predicted pixel values of pixel g and n=(C+2D+E+2) > > 2
The predicted pixel values of pixel h and o=(D+2E+F+2) > > 2
The predicted pixel values of pixel p=(E+2F+G+2) > > 2...(10)
Mode 8 is Horizontal_Up (upper level) prediction mode, and be suitable only for pixel value A, B, C, D, I, J, K, the case where L and M is " available ".In this case, the prediction picture of pixel a to p is generated as following expressions (11) Element value:
The predicted pixel values of pixel a=(I+J+1) > > 1
The predicted pixel values of pixel b=(I+2J+K+2) > > 2
The predicted pixel values of pixel c and e=(J+K+1) > > 1
The predicted pixel values of pixel d and f=(J+2K+L+2) > > 2
The predicted pixel values of pixel g and i=(K+L+1) > > 1
The predicted pixel values of pixel h and j=(K+3L+2) > > 2
Predicted pixel values=L of pixel k, l, m, n, o and p...(11)
Next, 4 × 4 pixel intra prediction mode (Intra_4 × 4_pred_ that luminance signal will be described referring to Fig. 7 Mode coded format).For the example of Fig. 7, the current block C as encoding target of 4 × 4 pixel of composition is shown, also, is shown The block A and block B neighbouring with current block of 4 × 4 pixels is constituted out.
In this case, it is conceivable that in Intra_4 × 4_pred_mode in current block C and block A and block B Intra_4 × 4_pred_mode has high correlation.Coded treatment is executed according to following using the correlation, so as to reality Existing higher code efficiency.
Specifically, Intra_4 × 4_pred_mode in block A and block B is taken as respectively for the example in Fig. 7 Intra_4 × 4_pred_modeA and Intra_4 × 4_pred_modeB, also, MostProbableMode be defined as with Lower expression formula (12):
MostProbableMode=Min (Intra_4 × 4_pred_modeA, Intra_4 × 4_pred_modeB) ... (12)。
That is, the block for being assigned lesser mode_number in block A and block B is taken as MostProbableMode。
Referred to as prev_intra4 × 4_pred_mode_flag [luma4 × 4Blkldx] and rem_intra4x4_ Two values of pred_mode [luma4x4Blkldx] are defined as the parameter about current block C in bit stream, also, pass through Decoding process is executed based on the processing of pseudo-code shown in following expressions (13), it is hereby achieved that about block C's The value of Intra_4 × 4_pred_mode and Intra4 × 4PredMode [luma4 × 4Blkldx].
If(prev_intra4×4_pred_mode_flag[luma4×4Blkldx])
Intra4 × 4PredMode [luma4 × 4Blkldx]=MostProbableMode
else
if(rem_intra4×4_pred_mode[luma4×4Blkldx]<MostProbableMode)
Intra4 × 4PredMode [luma4 × 4Blkldx]=rem_intra4 × 4_pred_mode [luma4 × 4Blkldx]
else
Intra4 × 4PredMode [luma4 × 4Blkldx]=rem_intra4 × 4_pred_mode [luma4 × 4Blkldx]+1...(13)
Next, 8 × 8 pixel intra prediction modes will be described.Fig. 8 and Fig. 9 is nine kind of 8 × 8 picture for showing luminance signal The diagram of plain intra prediction mode (intra_8 × 8_pred_mode).
Hypothesis, the pixel value in current 8 × 8 pieces are taken as (0≤x≤7 p [x, y];0≤y≤7), also, such as p [- 1, -1] ..., p [- 1,15], p [- 1,0] ..., [p-1,7] equally indicate contiguous block pixel value.
About 8 × 8 pixel intra prediction modes, low-pass filtering is carried out to neighborhood pixels before generating predicted value.Now, Hypothesis, pixel value p [- 1, -1] before low-pass filtering treatment ..., p [- 1,15], p [- 1,0] ..., p [- 1, 7] indicate, also, pixel value p'[-1 after this process, -1] ..., p'[-1,15], p'[-1,0] ..., p'[- 1,7] it indicates.
Firstly, in the case where p [- 1, -1] is " available ", the calculating p'[0 as following expressions (14), -1], In the case where " not available ", the calculating p'[0 as following expressions (15), -1].
P'[0, -1]=(p [- 1, -1]+2*p [0, -1]+p [1, -1]+2) > > 2... (14)
P'[0, -1]=(3*p [0, -1]+p [1, -1]+2) > > 2... (15)
Calculate p'[x as following expressions (16), -1] (x=0 ..., 7).
P'[x, -1]=(p [x-1, -1]+2*p [x, -1]+p [x+1, -1]+2) > > 2... (16)
In the case where p [x, -1] (x=8 ..., 15) is " available ", p' is calculated as following expressions (17) [x, -1] (x=8 ..., 15).
P'[x, -1]=(p [x-1, -1]+2*p [x, -1]+p [x+1, -1]+2) > > 2
P'[15, -1]=(p [14, -1]+3*p [15, -1]+2) > > 2... (17)
In the case where p [- 1, -1] is " available ", it is calculated as follows p'[-1, -1].Specifically, in p [0, -1] In the case where being all " available " with both p [- 1,0], calculate p'[-1 as expression formula (18), -1], also, p [- 1, 0] in the case where being " not available ", the calculating p'[-1 as expression formula (19), -1].In addition, being " unavailable in p [0, -1] " in the case where, the calculating p'[-1 as expression formula (20), -1].
P'[-1, -1]=(p [0, -1]+2*p [- 1, -1]+p [- 1,0]+2) > > 2... (18)
P'[-1, -1]=(3*p [- 1, -1]+p [0, -1]+2) > > 2... (19)
P'[-1, -1]=(3*p [- 1, -1]+p [- 1,0]+2) > > 2... (20)
When p [- 1, y] (y=0 ..., 7) is " available ", p'[-1, y is calculated as follows] (y=0 ..., 7). Specifically, firstly, in the case where p [- 1, -1] is " available ", p'[-1 is calculated as expression formula (21), 0], and In the case where p [- 1, -1] is " not available ", the calculating p'[-1 as expression formula (22), 0]
P'[-1,0]=(p [- 1, -1]+2*p [- 1,0]+p [- 1,1]+2) > > 2... (21)
P'[-1,0]=(3*p [- 1,0]+p [- 1,1]+2) > > 2... (22)
In addition, calculating p'[-1, y as following expressions (23)] (y=1 ..., 6), also, such as expression formula (24) Equally calculate p'[-1,7].
P [- 1, y]=(p [- 1, y-1]+2*p [- 1, y]+p [- 1, y+1]+2) > > 2... (23)
P'[-1,7]=(p [- 1,6]+3*p [- 1,7]+2) > > 2... (24)
Using p' calculated in this way, according to the prediction generated in the intra prediction mode shown in figs. 8 and 9 as follows Value.
Mode 0 is vertical prediction mode, also, is only " available " Shi Shiyong in p [x, -1] (x=0 ..., 7). Predicted value pred8 × 8 are generated as following expressions (25)L[x,y]。
pred8×8L[x, y]=p'[x, -1] x, y=0 ..., 7... (25)
Mode 1 is horizontal prediction mode, also, is only " available " Shi Shiyong in p [- 1, y] (y=0 ..., 7). Predicted value pred8 × 8 are generated as following expressions (26)L[x,y]。
pred8×8L[x, y]=p'[-1, y] x, y=0 ..., 7... (26)
Mode 2 is DC prediction mode, also, generates predicted value pred8 × 8 according to followingL[x,y].Specifically, in p In the case where both [x, -1] (x=0 ..., 7) and p [- 1, y] (y=0 ..., 7) are " available ", such as expression formula (27) predicted value pred8 × 8 are equally generatedL[x,y]。
[mathematical expression 1]
It is " available " in p [x, -1] (x=0 ..., 7) and p [- 1, y] (y=0 ..., 7) is " not available " In situation, predicted value pred8 × 8 are generated as expression formula (28)L[x,y]。
[mathematical expression 2]
It is " not available " in p [x, -1] (x=0 ..., 7) and p [- 1, y] (y=0 ..., 7) is " available " In situation, predicted value pred8 × 8 are generated as expression formula (29)L[x,y]。
[mathematical expression 3]
The case where both p [x, -1] (x=0 ..., 7) and p [- 1, y] (y=0 ..., 7) are " not available " In, predicted value pred8 × 8 are generated as expression formula (30)L[x,y]。
pred8×8L[x, y]=128... (30)
Here, expression formula (30) indicates the case where 8 bits input.
Mode 3 is Diagonal_Down_Left_prediction mode (diagonal down-left prediction mode), also, according to It is following to generate predicted value pred8 × 8L[x,y].Specifically, Diagonal_Down_Left_prediction mode only exists P [x, -1], x=0 ..., 15 be " available " Shi Shiyong, also, generate as following expressions (31) in x=7 and y= Predicted pixel values in the case where 7, also, other predicted pixel values are generated as following expressions (32).
pred8×8L[x, y]=(p'[14, -1]+3*p [15, -1]+2) > > 2... (31)
pred8×8L[x, y]=(p'[x+y, -1]+2*p'[x+y+1, -1]+p'[x+y+2, -1]+2) > > 2... (32)
Mode 4 is Diagnonal_Down_Right_prediction mode (down-right prediction mode), also, is pressed Predicted value pred8 × 8 are generated according to followingL[x,y].Specifically, Diagonal_Down_Right_prediction mode is only Only at p [x, -1], x=0 ..., 7 and p [- 1, y], y=0 ..., 7 be " available " Shi Shiyong, also, such as following expression Formula (33) equally generates predicted pixel values in the case where x>y, also, generates in x<y as following expressions (34) Predicted pixel values in situation.In addition, generating predicted pixel values in the case where x=y as following expressions (35).
pred8×8L[x, y]=(p'[x-y-2, -1]+2*p'[x-y-1, -1]+p'[x-y, -1]+2) > > 2... (33)
pred8×8L[x, y]=(p'[-1, y-x-2]+2*p'[-1, y-x-1]+p'[-1, y-x]+2) > > 2... (34)
pred8×8L[x, y]=(p'[0, -1]+2*p'[-1, -1]+p'[-1,0]+2) > > 2... (35)
Mode 5 is Vertical_Right_prediction mode (right vertical prediction mode), also, is produced according to following Raw predicted value pred8 × 8L[x,y].Specifically, Vertical_Right_prediction mode is only in p [x, -1], x =0 ..., 7 and p [- 1, y], y=-1 ..., 7 be " available " Shi Shiyong.Now, fixed as following expressions (36) Adopted zVR.
ZVR=2*x-y... (36)
At this point, generating prediction as following expressions (37) in the case where zVR is 0,2,4,6,8,10,12 or 14 Pixel value, also, in the case where zVR is 1,3,5,7,9,11 or 13, prediction pixel is generated as following expressions (38) Value.
pred8×8L[x, y]=(p'[x- (y > > 1) -1, -1]+p'[x- (y > > 1), -1]+1) > > 1... (37)
pred8×8L[x, y]=(p'[x- (y > > 1) -2, -1]+2*p'[x- (y > > 1) -1, -1]+p'[x- (y > > 1), - 1]+2)>>2...(38)
In addition, generating predicted pixel values as following expressions (39), also, removing this in the case where zVR is -1 Except in the case where, specifically, in the case where zVR is -2, -3, -4, -5, -6 or -7, as following expressions (40) Generate predicted pixel values.
pred8×8L[x, y]=(p'[-1,0]+2*p'[-1, -1]+p'[0, -1]+2) > > 2... (39)
pred8×8L[x, y]=(p'[-1, y-2*x-1]+2*p'[-1, y-2*x-2]+p'[-1, y-2*x-3]+2) > > 2...(40)
Mode 6 is Horizontal_Down_prediction mode (lower horizontal prediction mode), also, is produced according to following Raw predicted value pred8 × 8L[x,y].Specifically, Horizontal_Down_prediction mode is only in p [x, -1], x =0 ..., 7 and p [- 1, y], y=-1 ..., 7 be " available " Shi Shiyong.Now, fixed as following expressions (41) Adopted zHD.
ZHD=2*y-x... (41)
At this point, generating prediction as following expressions (42) in the case where zHD is 0,2,4,6,8,10,12 or 14 Pixel value, also, in the case where zHD is 1,3,5,7,9,11 or 13, prediction pixel is generated as following expressions (43) Value.
pred8×8L[x, y]=(p'[-1, y- (x > > 1) -1]+p'[-1, y- (x > > 1)+1] > > 1... (42)
pred8×8L[x, y]=(p'[-1, y- (x > > 1) -2]+2*p'[-1, y- (x > > 1) -1]+p'[-1, y- (x > > 1)]+2)>>2...(43)
In addition, generating predicted pixel values as following expressions (44), also, removing this in the case where zHD is -1 Except in the case where, specifically, in the case where zHD is -2, -3, -4, -5, -6 or -7, as following expressions (45) Generate predicted pixel values.
pred8×8L[x, y]=(p'[-1,0]+2*p'[-1, -1]+p'[0, -1]+2) > > 2... (44)
pred8×8L[x, y]=(p'[x-2*Y-1, -1]+2*p'[x-2*y-2, -1]+p'[x-2*y-3, -1]+2) > > 2...(45)
Mode 7 is Vertical_Left_prediction mode (left vertical prediction mode), also, is generated according to following Predicted value pred8 × 8L[x,y].Specifically, Vertical_Left_prediction mode is only in p [x, -1], x= 0 ..., 15 be " available " Shi Shiyong, and in the case where y=0,2,4 or 6, prediction is generated as following expressions (46) Pixel value, also, in the case where in addition to this, that is, in the case where y=1,3,5 or 7, as following expressions (47) Generate predicted pixel values.
pred8×8L[x, y]=(p'[x+ (y > > 1), -1]+p'[x+ (y > > 1)+1, -1]+1) > > 1... (46)
pred8×8L[x, y]=(p'[x+ (y > > 1), -1]+2*p'[x+ (y > > 1)+1, -1]+p'[x+ (y > > 1)+2, - 1]+2)>>2...(47)
Mode 8 is Horizontal_Up_prediction mode (upper horizontal prediction mode), also, is generated according to following Predicted value pred8 × 8L[x,y].Specifically, Horizontal_Up_prediction mode is only in p [- 1, y], y= 0 ..., 7 be " available " Shi Shiyong.Now, zHU is defined as following expressions (48).
ZHU=x+2*y... (48)
At this point, generating prediction as following expressions (49) in the case where the value of zHU is 0,2,4,6,8,10 or 12 Pixel value, also, in the case where the value of zHU is 1,3,5,7,9 or 11, prediction pixel is generated as following expressions (50) Value.
pred8×8L[x, y]=(p'[-1, y+ (x > > 1)]+p'[-1, y+ (x > > 1)+1]+1) > > 1... (49)
pred8×8L[x, y]=(p'[-1, y+ (x > > 1)] ... (50)
In addition, in the case where the value of zHU is 13, the generation predicted pixel values as following expressions (51), also, In addition to this in the case where, that is, in the case where the value of zHU is greater than 13, prediction pixel is generated as following expressions (52) Value.
pred8×8L[x, y]=(p'[-1,6]+3*p'[-1,7]+2) > > 2... (51)
pred8×8L[x, y]=p'[-1,7] ... (52)
Next, 16 × 16 pixel intra prediction modes will be described.Figure 10 and Figure 11 is four kind 16 for showing luminance signal The diagram of × 16 pixel intra prediction modes (Intra_16 × 16_pred_mode).
2 four kinds of intra prediction modes will be described referring to Fig.1.For the example in Figure 12, processing in frame to be carried out is shown Current macro A, also, P (x, y);X, y=-1,0 ..., 15 indicate pixel value with current macro A neighbouring pixel.
Mode 0 is vertical prediction mode, and only at P (x, -1);X, y=-1,0 ..., 15 be " available " Shi Shi With.In this case, the predicted pixel values Pred of each pixel of current macro A is generated as following expressions (53) (x,y)。
Pred (x, y)=P (x, -1);X, y=0 ..., 15... (53)
Mode 1 is horizontal prediction mode, and only at P (- 1, y);X, y=-1,0 ..., 15 be " available " Shi Shi With.In this case, the predicted pixel values Pred of each pixel of current macro A is generated as following expressions (54) (x,y)。
Pred (x, y)=P (- 1, y);X, y=0 ..., 15... (54)
Mode 2 is DC prediction mode, also, in P (x, -1) and P (- 1, y);X, y=-1,0 ..., 15 be entirely " can " in the case where, the predicted pixel values Pred of each pixel of current macro A is generated as following expressions (55) (x,y)。
[mathematical expression 4]
Wherein x, y=0 ..., 15 ... (55)
In addition, at P (x, -1);X, y=-1,0 ..., 15 be " not available " in the case where, such as following expressions (56) Equally generate the predicted pixel values Pred (x, y) of each pixel of current macro A.
[mathematical expression 5]
Wherein x, y=0 ..., 15 ... (56)
In addition, at P (- 1, y);X, y=-1,0 ..., 15 be " not available " in the case where, such as following expressions (57) Equally generate the predicted pixel values Pred (x, y) of each pixel of current macro A.
[mathematical expression 6]
Wherein x, y=0 ..., 15 ... (57)
In P (x, -1) and P (- 1, y);X, y=-1,0 ..., 15 be entirely " not available " in the case where, 128 are used as Predicted pixel values.
Mode 3 is plane prediction mode, and only in P (x, -1) and P (- 1, y);X, y=-1,0 ..., 15 all It is " available " Shi Shiyong.In this case, each pixel that current macro A is generated as following expressions (58) Predicted pixel values Pred (x, y).
[mathematical expression 7]
Pred (x, y)=Clip1 ((a+b (x-7)+c (y-7)+16) > > 5)
A=16 (P (- 1,15)+P (15, -1))
B=(5H+32) > > 6
C=(5V+32) > > 6
Firstly, the intra prediction mode by description about color difference signal.Figure 13 is four kinds of frames for showing color difference signal The diagram of inner estimation mode (Intra_chroma_pred_mode).It can independently be set with the intra prediction mode of luminance signal Set the intra prediction mode of color difference signal.About color difference signal intra prediction mode and above-mentioned luminance signal 16 × 16 pixel intra prediction modes are consistent.
However, 16 × 16 pixel intra prediction modes of luminance signal take 16 × 16 block of pixels as target, it is still, another Aspect, the intra prediction mode about color difference signal take 8 × 8 block of pixels as target.In addition, such as above-mentioned Figure 10 and Figure 13 Shown, MODE NUMBER between the two does not correspond to.
It is now assumed that our 16 × 16 pixel intra prediction modes in accordance with the luminance signal described above in reference to Figure 12 In current block A pixel value and neighborhood pixels value definition.For example, it is assumed that saying, with the current macro A handled in frame to be carried out The pixel value of (in the case where color difference signal, 8 × 8 pixels) neighbouring pixel is taken as P (x, y);X, y=-1,0 ..., 7。
Mode 0 is DC prediction mode, also, in P (x, -1) and P (- 1, y);X, y=-1,0 ..., 7 be entirely " available " in the case where, as following expressions (59) generate current macro A each pixel predicted pixel values Pred (x, y)。
[mathematical expression 8]
Wherein x, y=0 ..., 7 ... (59)
In addition, at P (- 1, y);X, y=-1,0 ..., 7 be " not available " in the case where, such as following expressions (60) Equally generate the predicted pixel values Pred (x, y) of each pixel of current macro A.
[mathematical expression 9]
Wherein x, y=0 ..., 7 ... (60)
In addition, at P (x, -1);X, y=-1,0 ..., 7 be " not available " in the case where, such as following expressions (61) Equally generate the predicted pixel values Pred (x, y) of each pixel of current macro A.
[mathematical expression 10]
Wherein x, y=0 ..., 7 ... (61)
Mode 1 is horizontal prediction mode, and only at P (- 1, y);X, y=-1,0 ..., 7 be " available " Shi Shi With.In this case, the predicted pixel values Pred of each pixel of current macro A is generated as following expressions (62) (x, y).
Pred (x, y)=P (- 1, y);X, y=0 ..., 7... (62)
Mode 2 is vertical prediction mode, and only at P (x, -1);X, y=-1,0 ..., 7 be " available " Shi Shi With.In this case, the predicted pixel values Pred of each pixel of current macro A is generated as following expressions (63) (x, y).
Pred (x, y)=P (x, -1);X, y=0 ..., 7... (63)
Mode 3 is plane prediction mode, and only in P (x, -1) and P (- 1, y);X, y=-1,0 ..., 7 be " can " Shi Shiyong.In this case, the prediction of each pixel of current macro A is generated as following expressions (64) Pixel value Pred (x, y).
[mathematical expression 11]
Pred (x, y)=Clip1 (a+b (x-3)+c (y-3)+16) > > 5;X, y=0 ..., 7
A=16 (P (- 1,7)+P (7, -1))
B=(17H+16) > > 5
C=(17V+16) > > 5
As described above, the intra prediction mode of luminance signal includes nine kinds of predictions of 4 × 4 pixels and 8 × 8 pixel block units Four kinds of prediction modes of mode and 16 × 16 pixel macroblock units.For each macro block unit, these block units are set Mode.The intra prediction mode of color difference signal includes four kinds of prediction modes of 8 × 8 pixel macroblock units.It can believe with brightness Number intra prediction mode be provided independently from the intra prediction mode of color difference signal.
In addition, about 4 × 4 pixel intra prediction modes (intra-frame 4 * 4 forecasting model) of luminance signal and 8 × 8 frame of pixels Inner estimation mode (8 × 8 prediction mode in frame) is arranged pre- in a kind of frame for 4 × 4 pixels and 8 × 8 pixel intensity blocks Survey mode.16 × 16 pixel intra prediction modes (16 × 16 prediction mode in frame) and color difference signal about luminance signal A kind of prediction mode is arranged for a macro block in intra prediction mode.
It note that the prediction mode of these types corresponds to the direction indicated with above-mentioned number 0,1,3 to 8 in Fig. 5. Prediction mode 2 is average value prediction.
As described previously for the intra prediction according to H.264/AVC format, only with expression formula (14) to expression formula (24) before executing intra prediction with the block incremental of 8 × 8 above-mentioned pixels, with determining filter factor to the pixel of neighborhood pixels Value executes filtering processing.On the contrary, using picture coding device 51, the intra prediction for executing all intra prediction modes it Before, filtering processing is executed to the pixel value of neighborhood pixels with according to the filter factor of the block to be predicted setting.
[configuration example of intraprediction unit and neighborhood pixels interpolation filtering switch unit]
Figure 14 is the detailed of neighborhood pixels interpolation filtering switch unit 75 and intraprediction unit 74 shown in pictorial image 1 The block diagram of configuration example.
In the case where example in Figure 14, intraprediction unit 74 is produced by adjacent image setting unit 81, forecast image Raw unit 82 and optimum prediction mode determination unit 83 are constituted.
Neighborhood pixels interpolation filtering switch unit 75 is by prediction mode buffer 91, quantization parameter buffer 92 and low pass filtered Wave setting unit 93 is constituted.It note that low-pass filtering setting unit 93 has built-in filter factor memory 94.
From frame memory 72 to adjacent image setting unit 81 for the neighborhood pixels value of the current block applied to intra prediction. Although the diagram of switch 73 is omitted in Figure 14, in fact, being set from frame memory 72 to adjacent image via switch 73 It sets unit 81 and executes supply.It note that in the case where intra prediction, without the pixel of the deblocking filtering of de-blocking filter 71 Value is used as neighborhood pixels value.
Adjacent image setting unit 81 is deposited using the filter factor being arranged by low-pass filtering setting unit 93 to from frame The neighborhood pixels value of the current block of reservoir 72 executes filtering processing, and pre- by being supplied to by the neighborhood pixels value of filtering processing Altimetric image generates unit 82.
The information that its mode is the intra prediction mode being presently processing is supplied to pre- by forecast image generation unit 82 Survey pattern buffer 91.Forecast image generates unit 82 and uses the neighbour by filtering processing from adjacent image setting unit 81 Nearly pixel value is generated pre- with being supplied to the intra prediction mode of prediction mode buffer 91 to execute intra prediction to current block Altimetric image.The forecast image of generation is supplied to optimum prediction mode determination unit 83 together with intraprediction mode information.
Carrying out in frame in advance for the reading of buffer 62 is rearranged to the supply of optimum prediction mode determination unit 83 from picture The image of survey generates the forecast image and its intraprediction mode information that unit 82 generates by forecast image.
Optimum prediction mode determination unit 83 is calculated using the information of supply about in the frame for having generated forecast image The cost function value of prediction mode, and the intra prediction mode for generating the minimum value in calculated cost function value is determined For best intra prediction mode.Optimum prediction mode determination unit 83 is by the forecast image of best intra prediction mode and accordingly Cost function value is output to forecast image selecting unit 77.
In addition, the feelings for the forecast image for having selected to generate in best intra prediction mode in forecast image selecting unit 77 In condition, the information for indicating best intra prediction mode is supplied to lossless coding unit by optimum prediction mode determination unit 83 66。
The storage of prediction mode buffer 91 generates the intraprediction mode information of unit 82 from forecast image.Quantization parameter Buffer 92 stores the quantization parameter from Rate control unit 78.
Low-pass filtering setting unit 93 reads the intraprediction mode information of current block from prediction mode buffer 91, and And quantization parameter corresponding with current block is read from quantization parameter buffer 92.Low-pass filtering setting unit 93 is out of be stored in It sets filter factor in filter factor memory 94 and filter factor corresponding with the information is set, also, by the filtering system of setting Number is supplied to adjacent image setting unit 81.
Filter factor memory 94 stores filter factor corresponding with quantization parameter and by the Figure 28 being described later on Learning device 251 at learnt using training image and the intra prediction mode that obtains.For example, for per a piece of (slice), it calculates and stores filter factor, as described below.
[description of the calculating of optimum filtering coefficient]
Next, referring to Fig.1 5 descriptions to be used for the calculating side of the optimum filtering coefficient of the filtering processing to neighborhood pixels Method.It note that in the example of Figure 15, the example that vertical prediction (vertical prediction) is executed to the current block of 4 × 4 pixels be shown, But the case where can be applied to any intra prediction mode is described below.
Intra prediction for the block incremental of 8 × 8 pixels for above expression formula (14) to expression formula (24) being used to describe, The 3- tap filtering coefficient of { 1,2,1 } is defined for the low-pass filter of neighborhood pixels, but we are by { c0,c1,c2View For the general type of 3- tap.In addition, for the present invention, also the 4th parameter c of introducing3As deviant.
Although note that the 3- tap filter is described as to be arranged for each segment in the following description, But the 3- tap filter is without being limited thereto, also, for example, can be arranged for entire sequence or for each GOP.
In example in Figure 15, akm(0≤k, m≤3) are the pixel value for the pixel for including in current block, also, bm (- 1≤m≤4) are the pixel values of the neighborhood pixels for vertical prediction.
Firstly, by neighborhood pixels value bmThe 3- tap filtering processing of execution is shown in following expressions (65) to generate B'm (0≤m≤3) out.
[mathematical expression 12]
b′m=c0*bm-l+cl*bm+c2*bm+1+c3(0≤m≤3)…(65)
That is, we will say in the case where executing filtering processing and using filter factor, also using corresponding Deviant, as shown in expression formula (65), in addition it is following without particularly referring to.In other words, filter factor and deviant are to use In the coefficient of filtering processing.In an identical manner, in the case where decoding side is sent in a manner of encoding in filter factor, I Will say, corresponding deviant is also sent in a manner of coding.
Now, if we say that the predicted pixel values when intra prediction mode is n are pij(b'm,n);0≤i, j≤3, Then about intra predicting pixel values, following expressions (66) is set up, because passing through the linear list described above in reference to Fig. 2 to 14 Prediction pixel is generated up to formula.
[mathematical expression 13]
pij(b′m, n)
=pij(c0*bm-1+c1*bm+c2*bm+1+c3, n)
=c0*pij(bm-1, n) and+c1*pij(bm, n) and+c2*pij(bm+1, n) and+c3…(66)
At this point, about with aijThe prediction square error such as following expressions of current block Ω as original image pixels value (67) shown in.
[mathematical expression 14]
Now, if we indicate the set of the intra block with intra prediction mode n coding in current clip with Φ, Prediction square error about the block for belonging to Φ and indicated with following expressions (68).
[mathematical expression 15]
In above-mentioned expression formula (68), Err (Ω ∈ Φ) is considered as c by us0、c1、c2、c3Function, that is, Err (Ω ∈ Φ;c0, c1, c2, c3), so making Err (Ω ∈ Φ;c0, c1, c2, c3) minimize c0、c1、c2、c3It will be in current clip Optimum filtering coefficient value.That is, being enough to obtain c in the case where following expressions (69) are set up0、c1、c2、c3
[mathematical expression 16]
That is, obtaining simultaneous equations shown in following expressions (70) from expression formula (69).
[mathematical expression 17]
Formula deformation is carried out to the expression formula (70) using matrix, obtains expression formula (71).
[mathematical expression 18]
The expression formula (71) is solved, optimum filtering coefficient and deviant { c can be obtained for current clip0,c1,c2, c3}。
It note that by solving the simultaneous equations in expression formula (70), obtain optimum filtering coefficient and deviant { c0,c1, c2,c3It is used as floating point values, still, for example, the picture coding device 51 in Fig. 1 and the corresponding picture decoding apparatus in Figure 22 In the case where 151, these are rounded to 8 bit coefficients.
That is, even if filter factor is floating-point, filter factor memory 94 is also according to the register of such as processor Length keeps these as n-bit (wherein, n is integer) value.
It, can also be for pre- in other frames by the way that other intra-frame prediction methods are also applied with method same as mentioned above Survey method obtains optimum filtering coefficient.In addition, by identical method, not only for intra-frame 4 * 4 forecasting model, and for 8 × 8 prediction modes in frame, in frame 16 × 16 prediction modes and color difference signal intra prediction mode, can obtain best Filter factor.
Although obtaining a filter factor for each intra prediction mode in the above description, the filtering Coefficient is without being limited thereto, and it is possible to carry out such arrangement: where for all intra prediction modes, only obtains one Filter factor.Particularly, for the intra prediction mode described above in reference to Fig. 2 to Figure 14, for it is vertical (it is vertical, Vertical) and horizontal (horizontal, horizontal) mode uses predicted pixel values as former state, still, holds for other modes Certain average treatment of row or weighted average processing are to generate prediction pixel, so its characteristic is different.Therefore, vertical water flat-die is executed The classification of the two categories of formula and other modes, also, calculates the filter factor of each classification, may be implemented code efficiency into One step improves.In addition, for example, there may be a filter factors for intra-frame 4 * 4 forecasting model, use for luminance signal A filter factor in 8 × 8 prediction modes in frame and a filter factor for 16 × 16 prediction modes in frame.Example Such as, for color difference signal, filter factor can be obtained respectively for Cb/Cr.
In addition, in the above description, for the filter factor for low-pass filtering treatment, having used { c0,c1,c2Three Tap, still, this is not limited to 3 taps, and the filter of any number of tap also can be used.That is, obtaining tap Filter factor+deviant of number.However, as number of taps increases, it is desirable that the number of the simultaneous equations of solution also increases.
Furthermore, it is possible to carry out such arrangement: where prepare and applies the filter factor different according to picture frame, such as CIF (CLV Common Intermediate Format, Common Intermediate Format)/QCIF (a quarter CIF, Quarter CIF), SD (standard definition), HD (fine definition) etc..
In addition, calculating filtering system by minimizing intra prediction (prediction square error) for the above method Number.But filter factor calculation method is without being limited thereto, also, in the case where needing to send filter factor to decoding side, It can execute comprising the optimization for sending the bit of filter factor.
In addition, for above-mentioned filter factor, it will be assumed that the symmetry of coefficient, as shown in following expressions (72).
C0=C2... (72)
That is, filter factor is calculated, to have symmetry, such as { c about the center coefficient for corresponding to 0 phase0,c1, c0Equally.Therefore, three simultaneous equations as shown in above-mentioned expression formula (70) can be reduced to 2.As a result, it is possible to reduce Calculation amount.
It is suitable for the filter factor of input picture and in an adaptive way to neighborhood pixels execution low pass filtered by being arranged Wave processing, using the above method, is able to use the forecast image for being suitable for the image, quantization parameter and prediction mode to execute volume Code, so as to improve code efficiency.
It is calculated about above-mentioned optimum filtering coefficient, it is contemplated that two methods.A kind of method is processed offline, that is, its In, before executing coded treatment, filter factor is calculated using the picture signal for training in advance, to make all figures As signal optimizes.Study later with reference to Figure 28 description as the processed offline is handled, also, is calculated by study processing Filter factor and deviant out is stored in the filter factor memory 94 in Figure 14.
Second method is online processing, that is, continuously calculates optimum filtering coefficient for each segment.In this feelings Under condition, decoding side is sent by calculated filter factor and deviant.It executes and is used as second later with reference to Figure 20 description The example of the case where processed offline of method.
[description of the coded treatment of picture coding device]
Next, referring to Fig.1 6 flow chart to be described to the coded treatment of the picture coding device 51 in Fig. 1.
In step s 11, A/D converting unit 61 executes Analog-digital Converter to input picture.In step s 12, picture Rearrange buffer 62 and store image supply from A/D converting unit 61, and execution from for show the sequence of picture to Sequence for coding rearranges.
In step s 13, computing unit 63 calculates between the image rearranged in step s 12 and forecast image Difference.In the case where executing inter-prediction, via forecast image selecting unit 77 by forecast image from motion prediction/compensating unit 76 are supplied to computing unit 63, also, in the case where executing intra prediction, via forecast image selecting unit 77 by prognostic chart As being supplied to computing unit 63 from intraprediction unit 74.
Compared with raw image data, the data volume of differential data is smaller.Therefore, in the case where no variation to original The case where beginning image is encoded is compared, can be with amount of compressed data.
In step S14, orthogonal transform unit 64 carries out orthogonal transformation to the difference information supplied from computing unit 63.Tool It says to body, executes the orthogonal transformation of discrete cosine transform, Karhunen-Lo é ve transformation etc., also, output transform coefficient. In step S15, quantifying unit 65 quantifies transformation coefficient.In the quantization, speed control, so that description be retouched later The processing in step S25 stated.
Following local decoder is carried out to the difference information quantified in this way.Specifically, in step s 16, inverse quantization list Member 68 carries out inverse quantization to the transformation coefficient quantified by quantifying unit 65 using characteristic corresponding with the characteristic of quantifying unit 65. In step S17, inverse orthogonal transformation unit 69 is using characteristic corresponding with the characteristic of orthogonal transform unit 64 come to by inverse amount The inverse-quantized transformation coefficient for changing unit 68 carries out inverse orthogonal transformation.
In step S18, forecast image and local solution that computing unit 70 will be inputted via forecast image selecting unit 77 The difference information of code is added, and generates the image (input that the image corresponds to computing unit 63) of local decoder.In step In rapid S19, de-blocking filter 71 is filtered the image exported from computing unit 70.Therefore, block distortion is eliminated.In step In S20, frame memory 72 stores filtered image.It note that the image of the filtering processing without de-blocking filter 71 Also it is supplied to frame memory 72 from computing unit 70, to store.
In the step s 21, intraprediction unit 74 and motion prediction/compensating unit 76 are performed both by image prediction processing.Tool It says to body, in the step s 21, intraprediction unit 74 executes intra-prediction process with intra prediction mode.Motion prediction/compensation Unit 76 executes motion prediction and compensation deals with inter-frame forecast mode.
The details of the prediction processing in step S21 is described later with reference to Figure 17, still, according to present treatment, is executed all Candidate modes in prediction processing, also, calculate the cost function value in all candidate modes.Then, base Best intra prediction mode is selected in calculated cost function value, also, will pass through the frame in best intra prediction mode The forecast image and its cost function value that interior prediction generates are supplied to forecast image selecting unit 77.
It note that at this point, 74 use of intraprediction unit is cut by neighborhood pixels interpolation filtering before intra-prediction process The filter factor for changing the setting of unit 75 to execute filtering processing to the neighborhood pixels of the intra prediction for current block.Then, exist Intra prediction is executed at intraprediction unit 74 using the neighborhood pixels by filtering processing, also, generates forecast image.
In step S22, forecast image selecting unit 77 is based on from intraprediction unit 74 and motion prediction/compensating unit One of best intra prediction mode and best inter-frame forecast mode are determined as optimum prediction mode by the cost function value of 76 outputs. Then forecast image selecting unit 77 selects the forecast image in determining optimum prediction mode, and it is single to be supplied into calculating Member 63 and 70.As described above, the forecast image is for the calculating in step S13 and S18.
It note that the selection information of the forecast image is supplied to intraprediction unit 74 or motion prediction/compensating unit 76.In the case where having selected the forecast image in best intra prediction mode, intraprediction unit 74 will be indicated in optimum frame The information (that is, intraprediction mode information) of prediction mode is supplied to lossless coding unit 66.
In the case where having selected the forecast image in best inter-frame forecast mode, motion prediction/compensating unit 76 will refer to Show that the information of best inter-frame forecast mode is exported to lossless coding unit 66, also, as needed, it will be with best inter-prediction The corresponding information of mode is exported to lossless coding unit 66.The example of information corresponding with best inter-frame forecast mode includes fortune Moving vector information, flag information and reference frame information.That is, when having selected and the interframe as best inter-frame forecast mode When the corresponding forecast image of prediction mode, motion prediction/compensating unit 76 by inter-frame forecast mode information, motion vector information and Reference frame information is output to lossless coding unit 66.
In step S23, lossless coding unit 66 encodes the quantization transform coefficient exported from quantifying unit 65. Specifically, carrying out the lossless coding and compression of variable length code, arithmetic coding etc. to difference image.At this point, Be input in step S22 lossless coding unit 66 best intraprediction mode information from intraprediction unit 74 or Information corresponding with the best inter-frame forecast mode from motion prediction/compensating unit 76 and come from Rate control unit 78 Quantization parameter etc. be also encoded, and be added with head information.
In step s 24, storage buffer 67 stores difference image as compression image.It is stored in storage buffer 67 Compression image read in due course, and via transmitting path be sent to decoding side.
In step s 25, Rate control unit 78 passes through quantization based on the compression image being stored in storage buffer 67 The rate of the quantization operation of state modulator quantifying unit 65, not will lead to overflow or underflow.
The quantization parameter for being used for rate control at quantifying unit 65 is supplied to lossless coding unit 66, above-mentioned Step S23 in lossless coded treatment is carried out to the quantization parameter, and insert it into the head of compression image.In addition, The quantization parameter is supplied to neighborhood pixels interpolation filtering switch unit 75, and for being arranged for execute to neighborhood pixels Filtering processing filter factor, the filtering processing is executed before intra prediction.
[description of prediction processing]
Next, the flow chart in referring to Fig.1 7 is described the processing of the prediction in the step S21 in Figure 16.
It is the figure in the block handled in frame to be carried out rearranging the image to be processed that buffer 62 is supplied from picture As in the case where, the decoding image to be referred to is read from frame memory 72, also, is supplied to intra prediction via switch 73 Unit 74.
In step S31, intraprediction unit 74 is using the image of supply with all candidate intra prediction modes to wanting The pixel of the block of processing executes intra prediction.It note that and be used as and to join without the pixel of the deblocking filtering of de-blocking filter 71 The decoded pixel examined.
The details for referring to Fig.1 8 being described the intra-prediction process in step S31 is still managed according to this, and setting is best Filter factor, also, filtering processing is executed to neighborhood pixels using the filter factor of setting.Then, using have been carried out filter The neighborhood pixels of wave processing execute intra prediction, to generate forecast image.
Above-mentioned processing is executed on all candidate intra prediction modes, for all candidate intra prediction mode meters Cost function value is calculated, also, determines best intra prediction mode based on calculated cost function value.Resulting prediction The cost function value of image and best intra prediction mode is supplied to forecast image selecting unit 77.
The feelings that the image to be processed that buffer 62 is supplied is the image of interframe processing to be carried out are being rearranged from picture In condition, the image to be referred to is read from frame memory 72, also, is supplied to motion prediction/compensating unit via switch 73 76.In step s 32, these images are based on, motion prediction/compensating unit 76 executes interframe movement prediction processing.That is, Motion prediction/compensating unit 76 from the image that frame memory 72 is supplied with reference to executing in all candidate inter-frame forecast modes Motion prediction process.
The details of interframe movement in step S32 prediction processing is described later with reference to Figure 19, also, according to present treatment, The motion prediction process in all candidate inter-frame forecast modes is executed, still, for all candidate inter-frame forecast mode meters Calculate cost function value.
In step S33, motion prediction/compensating unit 76 compares calculated about inter-prediction mould in step s 32 The cost function value of formula, and the prediction mode for providing minimum value is determined as best inter-frame forecast mode.Motion prediction/compensation Then the forecast image and its cost function value that generate in best inter-frame forecast mode are supplied to forecast image to select by unit 76 Select unit 77.
[description of intra-prediction process]
Next, the flow chart in referring to Fig.1 7 is described the intra-prediction process in the step S31 in Figure 17.It please infuse Meaning, for the example in Figure 18, the case where by luminance signal is described, as an example.
In the step S25 in above-mentioned Figure 16, Rate control unit 78 is for the quantization parameter applied to current block.In step In rapid S41, quantization parameter buffer 92 obtains the quantization parameter for current block from Rate control unit 78 and stores the quantization Parameter.
In step S42, it is pre- out of 4 × 4 pixels, 8 × 8 pixels and 16 × 16 pixels frame that forecast image generates unit 82 A kind of intra prediction mode is selected in survey mode.The intraprediction mode information of selection is stored in prediction mode buffer 91 In.
Low-pass filtering setting unit 93 reads intraprediction mode information from prediction mode buffer 91, also, from quantization Parameter buffer 92 reads quantization parameter value.In step S43, low-pass filtering setting unit 93 is then from being stored in filter factor It is arranged in the filter factor gone out for each fragment computations in memory 94 and corresponds to the intra prediction mode and quantization The filter factor of parameter.The filter factor of setting is supplied to adjacent image setting unit 81.
In step S44, adjacent image setting unit 81 is using the filter factor of setting come the neighborhood pixels to current block Value executes filtering processing, also, forecast image will be supplied to generate unit 82 by the neighborhood pixels value of filtering processing.
In step S43, forecast image generates unit 82 using the neighborhood pixels value by filtering processing in step S42 The intra prediction mode of middle selection executes intra prediction to current block, and generates forecast image.
To optimum prediction mode determination unit 83 supply from picture rearrange buffer 62 read to carry out frame The image of interior prediction generates the forecast image and its intraprediction mode information that unit 82 generates by forecast image.
In step S46, optimum prediction mode determination unit 83 is calculated using the information of supply has produced prediction The cost function value of the intra prediction mode of image.Here, based on one of high complexity mode or the technology of low-complexity mode Carry out the calculating of executory cost functional value.The JM (conjunctive model) of the H.264/AVC reference software of format is used as to determine these Mode.
Specifically, temporarily, coded treatment is gone to all candidate modes in high complexity mode, As the processing in step S45.For the cost function value that each prediction mode calculating is indicated with following expressions (73), also, Select the prediction mode for providing its minimum value as optimum prediction mode.
Cost (Mode)=D+ λ R... (73)
D indicates original image and decodes the difference (distortion) between image, and R indicates the generation of the generation comprising orthogonal transform coefficient Code amount, also, λ indicates the Lagrange multiplier that be provided as the function of quantization parameter QP.
On the other hand, in low-complexity mode, forecast image is generated, also, extremely for all candidate modes The head bit of more calculation of motion vectors information, prediction mode information, flag information etc., as the processing in step S45.For each Prediction mode calculates the cost function value indicated with following expressions (74), also, selects to provide the prediction mode of its minimum value As optimum prediction mode.
Cost (Mode)=D+QPtoQuant (QP)+Header_Bit... (74)
D indicates original image and decodes the difference (distortion) between image, and Header_Bit indicates the head about prediction mode Bit, also, QptoQuant is provided as the function of the function of quantization parameter QP.
In low-complexity mode, forecast image is generated only for all prediction modes, also, without executing coding Processing and decoding process, so as to reduce calculation amount.
In step S47, whether optimum prediction mode determination unit 83 is determined terminates for all intra prediction modes Processing.That is, in step S47, it is determined whether for all of 4 × 4 pixels, 8 × 8 pixels and 16 × 16 pixels Intra prediction mode performs the processing of step S42 to S46.
In the case where determining that being not yet directed to all intra prediction modes ends processing in step S47, processing returns to To step S42, also, repeat subsequent processing.
In the case where determining in step S47 for all intra prediction mode end for the treatment of, processing is proceeded to Step S48.In step S48, its calculated cost function value is the frame of minimum value by optimum prediction mode determination unit 83 Inner estimation mode is determined as best intra prediction mode.
The forecast image and its corresponding cost functional value of best intra prediction mode are supplied to forecast image selecting unit 77。
In the case where forecast image selecting unit 77 has selected the forecast image generated in best intra prediction mode, The information for indicating best intra prediction mode is supplied to lossless coding unit 66 by optimum prediction mode determination unit 83.Then This information is encoded at lossless coding unit 66, also, the information is added with the head information of compression image (upper The step S23 in Figure 16 stated).
It note that be stored in filter factor memory 94 and calculated filter factor handled also similarly by study It is stored in the picture decoding apparatus 151 in the Figure 22 being described later on, from the filter factor and compression figure without that will be arranged The head information of picture is added and sends.
Therefore, in the case where H.264/AVC, there are 51 quantization parameters, there are the nine of 4 × 4 pixels and 8 × 8 pixels Kind intra prediction mode, also, when considering these combinations, need 51 × 9=459 huge filter factors.It need not will close It is sent to decoding side in the information of such huge filter factor, so as to the case where not increasing the expense of coefficient information Lower realization processing.
[description of interframe movement prediction processing]
Next, the flow chart in referring to Fig.1 9 is described the processing of the interframe movement prediction in the step S32 in Figure 17.
In step S61, motion prediction/compensating unit 76 is directed to eight kinds of frames being made of 16 × 16 pixels to 4 × 4 pixels Between each in prediction mode determine motion vector and reference picture.That is, for will be with each inter-prediction mould The block of formula processing determines motion vector and reference picture.
In step S62, based on pre- for eight kinds of interframe being made of 16 × 16 pixels to 4 × 4 pixels in step S61 Each motion vector determined in survey mode, motion prediction/compensating unit 76 carry out motion prediction and benefit to reference picture Repay processing.According to the motion prediction and compensation deals, the forecast image in each inter-frame forecast mode is generated.
In step S63, motion prediction/compensating unit 76 is generated about for being made of 16 × 16 pixels to 4 × 4 pixels Eight kinds of inter-frame forecast modes in each determine motion vector motion vector information, with compression image addition.
Also using the motion vector information generated when being calculated as this functional value in following step S64, also, pre- In the case where altimetric image selecting unit 77 has finally chosen corresponding forecast image, by the motion vector information of generation and prediction mould Formula information and reference frame information are exported together to lossless coding unit 66.
In step S64, motion prediction/compensating unit 76 is directed to eight kinds of frames being made of 16 × 16 pixels to 4 × 4 pixels Between each in prediction mode calculate cost function value shown in above-mentioned expression formula (73) or expression formula (74).When upper When determining best inter-frame forecast mode in the step S34 in the Figure 17 stated, use calculated cost function value here.
Next, as the second method for calculating optimum filtering coefficient following situations will be described referring to Figure 20 Example: where execute online processing, that is, continuously calculate optimum filtering coefficient for each segment.
Now, in this case it is necessary to send the filter factor gone out in coding side for each fragment computations to Side is decoded, also, is sent and has been broken down into the filter factors of a variety of situations and leads to the deterioration of code efficiency.Therefore, for piece Section only sends a filter factor, alternatively, each prediction mode for each block size only sends a filter factor, or Person only sends a filter factor for the prediction mode type of horizontal forecast, vertical prediction etc..
In addition, in the case where above-mentioned processed offline, to use intra prediction mode and quantization parameter as based on The example for calculating the parameter of filter factor is described.On the other hand, in the case where online processing, for calculating filter factor A large amount of parameter increases treating capacity, thus, it will be described with Figure 20 about the example for only using intra prediction mode for parameter Example.Although being described omitting, of course, it is possible to quantization parameter be only used, alternatively, two kinds of parameters can be used.
[the other configurations example of intraprediction unit and neighborhood pixels interpolation filtering switch unit]
Figure 20 is the intraprediction unit 74 being shown in following situations and neighborhood pixels interpolation filtering switch unit 75 The block diagram of another configuration example: online processing is executed for each segment, continuously to calculate optimum filtering coefficient.
In the case where example in Figure 20, intraprediction unit 74 and neighborhood pixels interpolation filtering switch unit 75 it Between insertion switch 101, also, it is different from situation shown in Figure 14, by switching on and off switch 101, intraprediction unit 74 Execute intra prediction twice.That is, being executed in the state that switch 101 disconnects for intraprediction unit 74 H.264/AVC intra prediction defined in, also, calculate the filter factor for being suitable for intra prediction.In the shape that switch 101 is connected In state, frame is executed with the filter factor being arranged by neighborhood pixels interpolation filtering switch unit 75 in calculated filter factor Interior prediction.
Intraprediction unit 74 in Figure 20 generates unit 112 and best by adjacent image setting unit 111, forecast image Prediction mode determination unit 113 is constituted.
Neighborhood pixels interpolation filtering switch unit 75 is by prediction mode buffer 121, optimum filtering computing unit 122 and low Pass filter setting unit 123 is constituted.
From frame memory 72 to neighborhood pixels setting unit 111 for all current of the current clip applied to intra prediction The neighborhood pixels value of block.In the case where Figure 20, the diagram of switch 73 is also omited.It note that in the case where intra prediction, Pixel value without the deblocking filtering of de-blocking filter 71 is used as neighborhood pixels value.
In the case where switch 101 is in an off state, 111 use of neighborhood pixels setting unit is used only for H.264/ The filter factor of intra prediction mode defined in AVC is filtered come the neighborhood pixels value to the current block from frame memory 72 Wave processing, and be supplied into forecast image and generate unit 112.That is, only use expression formula (14) above is extremely In the case where 8 × 8 prediction modes of expression formula (24) description, the neighborhood pixels value by filtering processing is supplied to forecast image Generate unit 112.In all other situations, the neighborhood pixels value of the current block from frame memory 72 is supplied as former state Unit 112 is generated to forecast image.
In the case where switch 101 is in (on) state of connection, it is arranged from low-pass filtering setting unit 123 to neighborhood pixels Unit 111 supplies filter factor.Therefore, neighborhood pixels setting unit 111 uses the filter being arranged by low-pass filtering setting unit 123 Wave system number is filtered come the neighborhood pixels value to the current block from frame memory 72, and will be by filtering processing Neighborhood pixels value is supplied to forecast image to generate unit 112.
Forecast image generates unit 112 using the neighborhood pixels value from neighborhood pixels setting unit 111 come with all Intra prediction mode executes intra prediction to current block, and generates forecast image.By the forecast image and intra prediction of generation Pattern information is supplied to optimum prediction mode determination unit 113 together.
Rearranged that buffer 62 reads from picture to the supply of optimum prediction mode determination unit 113 in frame The image of prediction generates the forecast image and its intraprediction mode information that unit 112 generates by forecast image.
Optimum prediction mode determination unit 113 calculates the intra prediction for having generated forecast image using the information of supply The cost function value of mode, and the intra prediction mode of the minimum value generated in calculated cost function value is determined as most Good intra prediction mode.
In the case where switch 101 is in (off) state of disconnection, optimum prediction mode determination unit 113 will be in optimum frame The information of prediction mode is supplied to prediction mode buffer 121.Switch 101 in an ON state in the case where, optimum prediction The forecast image of best intra prediction mode and corresponding cost function value are supplied to forecast image to select by pattern determining unit 113 Select unit 77.
In addition, the feelings for the forecast image for having selected to generate in best intra prediction mode in forecast image selecting unit 77 In condition, the information for indicating best intra prediction mode is supplied to lossless coding unit by optimum prediction mode determination unit 113 66。
Prediction mode buffer 121 stores the intraprediction mode information from optimum prediction mode determination unit 113.
The figure for the intra prediction that buffer 62 is read has been rearranged from picture to the supply of optimum filtering computing unit 122 The neighborhood pixels value of picture and the current block from frame memory 72.Optimum filtering computing unit 122 is from prediction mode buffer 121 Read each piece of intra prediction mode for including in current clip.Then optimum filtering computing unit 122 uses the information to The optimum filtering coefficient for calculating the intra prediction mode of current clip, as described in above referring to Fig.1 5, also, optimum filtering meter It calculates unit 122 and calculated filter factor is supplied to low-pass filtering setting unit 123.
Low-pass filtering setting unit 123 is arranged for current block in the filter factor of calculated current clip Filter factor, connect the terminal of switch 101, and the filter factor of setting be supplied to neighborhood pixels setting unit 111.This Outside, the filter factor for being used for current clip is supplied to lossless coding unit 66 by low-pass filtering setting unit 123.
[other descriptions of intra-prediction process]
Next, the neighborhood pixels interpolation filtering switch unit 75 and the frame that will be described referring to the flow chart in Figure 21 in Figure 20 The intra-prediction process that interior prediction unit 74 executes.It note that the intra-prediction process is pre- in the frame of the step S31 in Figure 17 Survey another example of processing.
Firstly, switch 101 is in an off state.Supplying from frame memory 72 to neighborhood pixels setting unit 111 will carry out The neighborhood pixels value of all current blocks of the current clip of intra prediction.111 use of neighborhood pixels setting unit is used only for H.264/AVC the filter factor of 8 × 8 pixel intra prediction modes defined in carrys out the neighbour to the current block from frame memory 72 Nearly pixel value executes filtering processing, and is supplied into forecast image and generates unit 112.That is, in other frames In the case where prediction mode, the neighborhood pixels value of the current block from frame memory 72 is supplied to forecast image generation as former state Unit 112.
In step s101, forecast image generates unit 112 to all pieces of execution intra predictions for including in current clip Processing.That is, forecast image, which generates unit 112, uses the neighbouring picture of the current block from neighborhood pixels setting unit 111 Plain value to execute intra prediction with each intra prediction mode, and generates forecast image.
To optimum prediction mode determination unit 113 supply from picture rearrange buffer 62 read to carry out frame The image of interior prediction generates the forecast image and its intraprediction mode information that unit 112 generates by forecast image.
In step s 102, optimum prediction mode determination unit 113 is calculated pre- about producing using the information of supply Cost function value in the above-mentioned expression formula (73) or expression formula (74) of all intra prediction modes of altimetric image.
In step s 103, optimum prediction mode determination unit 113 is by the cost in expression formula (73) or expression formula (74) Function is that the smallest intra prediction mode is determined as best intra prediction mode, and by the information of determining intra prediction mode It is supplied to prediction mode buffer 121.
Carrying out in frame in advance for the reading of buffer 62 has been rearranged from picture to the supply of optimum filtering computing unit 122 The neighborhood pixels value of the image of survey and the current block from frame memory 72.Optimum filtering computing unit 122 is slow from prediction mode It rushes device 121 and reads each piece of intra prediction mode for including in current clip.
In step S104, optimum filtering computing unit 122 uses the information to calculate the residual error for making entire current clip The smallest filter factor, the optimum filtering coefficient of each intra prediction mode as current clip.It will be above in reference to figure The filter factor of 15 descriptions is supplied to low-pass filtering setting unit 123.
Low-pass filtering setting unit 123 is arranged corresponding to current block in the filter factor of calculated current clip Filter factor, connects the terminal of switch 101, and the filter factor of setting is supplied to neighborhood pixels setting unit 111.
In step s105, neighborhood pixels setting unit 111 uses the filtering system being arranged by low-pass filtering setting unit 123 Number executes filtering processing to the neighborhood pixels value of the current block from frame memory 72.
Forecast image will be supplied to generate unit 112 by the neighborhood pixels value of filtering processing.In step s 106, it predicts Image generation unit 112 reuses the neighborhood pixels value by filtering processing come all pieces of execution to including in current clip Intra prediction, to generate forecast image.The forecast image of generation is supplied to together with intraprediction mode information best pre- Survey pattern determining unit 113.
Switch 101 in an ON state in the case where, optimum prediction mode determination unit 113 is by best intra prediction mould The forecast image of formula and corresponding cost function value are supplied to forecast image selecting unit 77.
In the step S22 of above-mentioned Figure 16, forecast image selecting unit 77 is by best intra prediction mode and optimum frame Between one of prediction mode be determined as optimum prediction mode, and supply the selection information of forecast image.
In step s 107, optimum prediction mode determination unit 113 determines whether to select according to the selection information of forecast image The forecast image of best intra prediction mode is selected.When the determining prediction for having selected best intra prediction mode in step s 107 In the case where image, processing proceeds to step S108.
In step S108, intraprediction mode information is supplied to lossless volume by optimum prediction mode determination unit 113 Code unit 66.It note that in the case where not yet supplying filter factor for current clip, come from optimum filtering computing unit 122 filter factor is also supplied to lossless coding unit 66.
In the case where determination not yet selects the forecast image of best intra prediction mode in step s 107, intra prediction Processing terminate.
Note that can obtain the filtering system advanced optimized by repeating the processing of above-mentioned step S104 to S106 Number.
The compression image of coding is transmitted via scheduled transmitting path, and is decoded by picture decoding apparatus.
[configuration example of picture decoding apparatus]
Figure 22 indicates the configuration for the embodiment for being used as the picture decoding apparatus for applying image processing apparatus of the invention.
Picture decoding apparatus 151 is by storage buffer 161, lossless decoding unit 162, inverse quantization unit 163, inverse orthogonal Converter unit 164, computing unit 165, de-blocking filter 166, picture rearrange buffer 167, D/A converting unit 168, frame Memory 169, switch 170, intraprediction unit 171, neighborhood pixels interpolation filtering switch unit 172, motion prediction/compensation list Member 173 and switch 174 are constituted.
Storage buffer 161 stores the compression image of transmission.Lossless decoding unit 162 using with it is lossless in Fig. 1 The corresponding format of the coded format of coding unit 66 come to from storage buffer 161 supply and compiled by the lossless coding unit 66 The information of code is decoded.Inverse quantization unit 163 using format corresponding with the quantization format of quantifying unit 65 in Fig. 1 to by The decoded image of lossless decoding unit 162 carries out inverse quantization.Inverse orthogonal transformation unit 164 uses and the orthogonal transformation in Fig. 1 The corresponding format of orthogonal transformation format of unit 64 carrys out the output to inverse quantization unit 163 and carries out inverse orthogonal transformation.
It will be by the forecast image phase Calais for exporting with being supplied from switch 174 of inverse orthogonal transformation by computing unit 165 The output is decoded.The block distortion of the removal decoding image of de-blocking filter 166, then, is supplied into frame memory 169 To store, and also outputs it to picture and rearrange buffer 167.
Picture rearranges buffer 167 and rearranges to image execution.Specifically, by by the picture in Fig. 1 again Arrangement buffer 62 is rearranged for original display sequence to the sequence for the frame that coded sequence rearranges.D/A converting unit 168 pairs rearrange the image that buffer 167 is supplied from picture and carry out numerical simulation conversion, and output it to unshowned Display is to show.
Switch 170 reads the image that carry out interframe processing and the image to be referred to from frame memory 169, outputs it to Motion prediction/compensating unit 173, the image of intra prediction will be used for and be supplied into frame by also reading from frame memory 169 Predicting unit 171.
The information for indicating intra prediction mode obtained will be decoded from lossless decoding unit by correct information 162 are supplied to intraprediction unit 171.Intraprediction unit 171 is based on the information by using by neighborhood pixels interpolation filtering The filter factor that switch unit 172 is arranged executes filtering processing and intra prediction to neighborhood pixels value to generate forecast image, and And the forecast image of generation is output to switch 174.
From lossless decoding unit 162 to neighborhood pixels interpolation filtering switch unit 172, supply according in image by compiling The correct information of coding at code device 51 is decoded and the letter of the information of instruction intra prediction mode and quantization parameter that obtains At least one of breath.In a manner of identical with the neighborhood pixels interpolation filtering switch unit 75 in Fig. 1, the filter of neighborhood pixels interpolation The storage of wave switch unit 172 is obtained and learning at the learning device 251 in the Figure 28 being described later on and quantization parameter The corresponding filter factor at least one of intra prediction mode.
The setting of neighborhood pixels interpolation filtering switch unit 172 and quantization parameter and frame from lossless decoding unit 162 The corresponding filter factor of at least one of inner estimation mode.Each segment neighborhood pixels interpolation filtering switch unit 172 will The filter factor of setting is supplied to intraprediction unit 74.
It note that for neighborhood pixels interpolation filtering switch unit 172, store the filter factor of preparatory off-line learning.But Be, it is noted that, with the neighborhood pixels interpolation filtering switch unit 75 in Fig. 1 in the case where line computation filter factor, for example, These filter factors are transmitted to it for each segment.In this case, neighborhood pixels interpolation filtering switch unit 172 makes With by the decoded filter factor of lossless decoding unit 162.
The information obtained (prediction mode information, motion vector information and reference frame will be decoded by correct information Information) from lossless decoding unit 162 it is supplied to motion prediction/compensating unit 173.Supplying instruction inter-frame forecast mode In the case where information, motion prediction/compensating unit 173 move to image pre- based on motion vector information and reference frame information Survey and compensation deals, to generate forecast image.The prediction that motion prediction/compensating unit 173 will generate in inter-frame forecast mode Image is output to switch 174.
Switch 174 selects the forecast image generated by motion prediction/compensating unit 173 or intraprediction unit 171 and will It is supplied to computing unit 165.
It note that using the picture coding device 51 in Fig. 1, in order to be determined based on the prediction mode of cost function, for All intra prediction modes execute intra-prediction process.On the other hand, using picture decoding apparatus 151, it is based only on coding The intra prediction mode being sent to it information, execute intra-prediction process.
[configuration example of intraprediction unit and neighborhood pixels interpolation filtering switch unit]
Figure 23 is the block diagram for illustrating the detailed configuration example of neighborhood pixels interpolation filtering switch unit and intraprediction unit. It note that the functional block in Figure 23 corresponds in the case where using the processed offline of picture coding device 51 shown in Figure 14 Functional block.
In the case where the example of Figure 23, intraprediction unit 71 generates unit 181 by forecast image and neighborhood pixels are set Set the composition of unit 182.Neighborhood pixels interpolation filtering switch unit 172 is by prediction mode buffer 191, quantization parameter buffer 192 and low-pass filtering setting unit 193 constitute.Low-pass filtering setting unit 193 has built-in filter factor memory 194.
It generates intraprediction mode information of the supply of unit 181 from lossless decoding unit 162 to forecast image and comes From the neighborhood pixels value by filtering processing of neighborhood pixels setting unit 182.Forecast image generates 181 use of unit and supplies to it The neighborhood pixels value answered to generate prognostic chart to execute intra prediction from the intra prediction mode of lossless decoding unit 162 Picture, and the forecast image of generation is supplied to switch 174.
From frame memory 169 to neighborhood pixels setting unit 182, supply will carry out the neighbouring picture of the current block of intra prediction Element value.In the case where Figure 23, the diagram of switch 170 is omitted, still, in fact, by neighborhood pixels value via switch 170 from Frame memory 169 is supplied to neighborhood pixels setting unit 182.
Neighborhood pixels setting unit 182 is using the filter factor being arranged by low-pass filtering setting unit 193 come to from frame The neighborhood pixels value of the current block of memory 169 executes filtering processing, and the neighborhood pixels value that have passed through filtering processing is supplied Unit 181 should be generated to forecast image.
Prediction mode buffer 191 stores the intraprediction mode information from lossless decoding unit 162.Quantization parameter Buffer 192 stores the quantization parameter from lossless decoding unit 162.
Low-pass filtering setting unit 193 reads the intraprediction mode information of current block from prediction mode buffer 191, and And quantization parameter corresponding with current block is read from quantization parameter buffer 192.Low-pass filtering setting unit 193 is from being stored in Filter factor in built-in filter factor memory 194 is arranged filter factor corresponding with the information, also, by the filtering of setting Coefficient is supplied to neighborhood pixels setting unit 182.
In a manner of identical with the filter factor memory 94 in Figure 14, filter factor memory 194 storage with by The corresponding filter factor of intra prediction mode and quantization parameter for learning at the learning device in Figure 28 being described later on and obtaining.
For example, being directed to each segment, calculating and storing filter factor, as described in above referring to Fig.1 5.It note that pair In filter factor memory 194, filter factor is also remained n-bit value according to the register capacity of processor, and (wherein, n is Integer).
[description of the decoding process of picture decoding apparatus]
Next, by the decoding process that picture decoding apparatus 151 executes is described referring to the flow chart of Figure 24.
In step S131, storage buffer 161 stores the image of transmission.In step S132, lossless decoding unit 162 pairs are decoded from the compression image that storage buffer 161 is supplied.Specifically, to by the lossless coding unit in Fig. 1 I picture, P picture and the B picture of 66 codings are decoded.
At this point, also to motion vector information, reference frame information, prediction mode information (instruction intra prediction mode or interframe The information of prediction mode), quantified parameter information, flag information etc. be decoded.
Specifically, prediction mode information is supplied in the case where prediction mode information is intraprediction mode information To intraprediction unit 171 and neighborhood pixels interpolation filtering switch unit 172.In addition, being carried out to quantified parameter information In the case where decoding, it is also supplied into neighborhood pixels interpolation filtering switch unit 172.It is inter-prediction in prediction mode information In the case where pattern information, it is supplied to movement pre- the reference frame information for corresponding to prediction mode information and motion vector information Survey/compensating unit 173.
In step S133, inverse quantization unit 163 is come pair using characteristic corresponding with the characteristic of quantifying unit 65 in Fig. 1 Inverse quantization is carried out by the decoded transformation coefficient of lossless decoding unit 162.In step S134, inverse orthogonal transformation unit 164 makes With characteristic corresponding with the characteristic of orthogonal transform unit 64 in Fig. 1 come to by the inverse-quantized transformation coefficient of inverse quantization unit 163 Carry out inverse orthogonal transformation.It means that the input (output of computing unit 63) with the orthogonal transform unit 64 in Fig. 1 Corresponding difference information is decoded.
In step S135, computing unit 165 will be via the input of switch 174 and the place in the step S141 being described later on The forecast image selected in reason is added with the difference information.Therefore, original image is decoded.In step S136, deblocking Filter 166 is filtered the image exported from computing unit 165.Therefore, block distortion is eliminated.In step S137, frame Memory 169 stores filtered image.
In step S138, intraprediction unit 171 and motion prediction/compensating unit 173 are in response to from lossless decoding The prediction mode information that unit 162 is supplied is handled to execute corresponding image prediction.
Specifically, in the case where supplying intraprediction mode information from lossless decoding unit 162, it is pre- in frame It surveys unit 171 and intra-prediction process is executed with intra prediction mode.At this point, 171 use of intraprediction unit is by neighborhood pixels The filter factor that filtering switch unit 172 is arranged is inserted to execute filtering processing and intra-prediction process to neighborhood pixels.
The details that the prediction processing in step S138 is described later with reference to Figure 25 is still managed according to this, will be by frame The forecast image or switch is supplied to by the forecast image that motion prediction/compensating unit 173 generates that predicting unit 171 generates 174。
In step S139, switch 174 selects forecast image.Specifically, what supply was generated by intraprediction unit 171 Forecast image or the forecast image generated by motion prediction/compensating unit 173.Therefore, the forecast image of supply is selected, quilt It is supplied to computing unit 165, also, in step S134, as described above, by the forecast image and inverse orthogonal transformation unit 164 Output be added.
In step S140, picture rearranges the execution of buffer 167 and rearranges.Specifically, will be encoded by image The picture of device 51 rearranges the sequence that buffer 62 is the frame for encoding and rearranging and is rearranged for original display sequence Column.
In step s 141, D/A converting unit 168 carries out number to the image for rearranging buffer 167 from picture Analog-converted.The image is exported to unshowned display, also, shows the image.
[description of prediction processing]
Next, the prediction in the step S138 described in Figure 24 referring to the flow chart in Figure 25 is handled.
In step S171, forecast image generates unit 181 and determines whether to have carried out intraframe coding to current block.By frame When inner estimation mode information is supplied to forecast image to generate unit 181 from lossless decoding unit 162, in step S171, in advance Altimetric image generates the determination of unit 181 and has carried out intraframe coding to current block, also, present treatment proceeds to step S172.
In step S172, forecast image generates unit 181 and receives and obtain intra prediction from lossless decoding unit 162 Pattern information.At this point, intraprediction mode information is also supplied to prediction mode buffer 191 and is stored.
In addition, the quantified parameter information from lossless decoding unit 162 is supplied to quantization parameter buffer 192 When, in step S173, quantization parameter buffer 192 obtains and stores quantization parameter.
Low-pass filtering setting unit 193 reads the intraprediction mode information of current block from prediction mode buffer 191, and And the quantization parameter about current block is read from quantization parameter buffer 192.In step S174, low-pass filtering setting unit 193 are arranged in the filter factor for being stored in each of built-in filter factor memory 194 segment corresponding to the letter The filter factor of the neighborhood pixels of breath.The filter factor of setting is supplied to neighborhood pixels setting unit 182.
In step S175, neighborhood pixels setting unit 182 uses the filtering system being arranged by low-pass filtering setting unit 193 Number executes filtering processing to the neighborhood pixels value of the current block from frame memory 169, and will have passed through filtering processing Neighborhood pixels value is supplied to forecast image to generate unit 181.
Forecast image generates unit 181 using the neighborhood pixels value supplied from neighborhood pixels setting unit 182 come in step The intra prediction mode obtained in rapid S172 executes intra prediction, and generates forecast image.The forecast image of generation is supplied To switch 174.
On the other hand, in the case where determination has not carried out intraframe coding in step S171, present treatment proceeds to step S177。
In the case where image to be processed is the image handled in frame to be carried out, by inter-frame forecast mode information, reference Frame information and motion vector information are supplied to motion prediction/compensating unit 173 from lossless decoding unit 162.In step S177 In, motion prediction/compensating unit 173 obtains inter-frame forecast mode information, reference frame information, fortune from lossless decoding unit 162 Moving vector information etc..
In step S178, then motion prediction/compensating unit 173 executes interframe movement prediction.Specifically, to locate In the case where the image of reason is the image of inter-prediction processing to be carried out, necessary image is read from frame memory 169, also, Necessary image is supplied to motion prediction/compensating unit 173 via switch 170.In step S177, motion prediction/compensation Unit 173 executes motion prediction based on the motion vector obtained in step S176 with inter-frame forecast mode, to generate prognostic chart Picture.The forecast image of generation is exported to switch 174.
[the other configurations example of intraprediction unit and neighborhood pixels interpolation filtering switch unit]
Figure 26 is the block diagram for illustrating the detailed configuration example of neighborhood pixels interpolation filtering switch unit and intraprediction unit. It note that the functional block in Figure 26 corresponds in the case where using the online processing of picture coding device 51 shown in Figure 20 Functional block.
In the case where the example of Figure 26, intraprediction unit 71 is by the neighborhood pixels setting unit 182 in Figure 23 and in advance Altimetric image generates unit 181 and constitutes.Neighborhood pixels interpolation filtering switch unit 172 by Figure 23 prediction mode buffer 191, Interpolation filtering buffer 201 and low-pass filtering setting unit 202 are constituted.It note that in the example in Figure 26, correspond to Figure 23 In the part of situation indicated with corresponding appended drawing reference, also, identical processing is essentially performed, to retouch it is omitted It states.
In the case where Figure 26, it is encoded for the filter factor that current clip calculates and is sent out from picture coding device 51 It send.Therefore, lossless decoding unit 162 decodes it together with other information, and neighborhood pixels interpolation filtering is supplied to switch The interpolation filtering buffer 201 of unit 172.
Interpolation filtering buffer 201 obtains filter factor and the storage for current clip from lossless decoding unit 162 It.
Low-pass filtering setting unit 202 reads the intraprediction mode information of current block from prediction mode buffer 191.It is low Pass filter setting unit 202 is read from the filter factor for the current clip being stored in interpolation filtering buffer 201 to be corresponded to Filter factor through the intra prediction mode read, and it is set to the filter factor for current block.By the filter of setting Wave system number is supplied to neighborhood pixels setting unit 182.
[other descriptions of prediction processing]
Next, intraprediction unit 171 and the switching of neighborhood pixels interpolation filtering will be described referring to the flow chart in Figure 27 Prediction processing in the case where unit 172.It note that the intra-prediction process is the prediction processing in step S138 in Figure 24 Another example.In addition, the processing of step S181, S182 and S185 to S188 essentially perform and step in Figure 25 in Figure 27 S171, S172 and the identical processing of S175 to S178, therefore omit detailed description.
In step S181, forecast image generates unit 181 and determines whether to have carried out intraframe coding to current block.By frame When inner estimation mode information is supplied to forecast image to generate unit 181 from lossless decoding unit 162, in step S181, in advance Altimetric image generates the determination of unit 181 and has carried out intraframe coding to current block, also, present treatment proceeds to step S182.
In step S182, forecast image generates unit 181 and receives and obtain intra prediction from lossless decoding unit 162 Pattern information.At this point, the intraprediction mode information is also supplied to prediction mode buffer 191 and is stored.
In addition, the information in the filter factor for current clip is supplied to interpolation to filter from lossless decoding unit 162 When wave buffer 201, interpolation filtering buffer 201 obtains the filter factor for being used for current clip in step S183, and stores.Please Note that supplying filter factor for each segment.
Low-pass filtering setting unit 202 is read from prediction mode buffer 191 to be believed for the intra prediction mode of current block Breath.In step S184, in addition to the filter factor for the current clip being stored in interpolation filtering buffer 201, low-pass filtering is set It sets unit 202 and also corresponds to filter factor of the intra prediction mode setting of current block for neighborhood pixels.By the filtering of setting Coefficient is supplied to neighborhood pixels setting unit 182.
In step S185, neighborhood pixels setting unit 182 uses the filtering system being arranged by low-pass filtering setting unit 202 Number executes filtering processing to the neighborhood pixels value of the current block from frame memory 169, and will have passed through filtering processing Neighborhood pixels value is supplied to forecast image to generate unit 181.
In step S186, forecast image generates unit 181 and uses the neighbouring picture supplied from neighborhood pixels setting unit 182 Plain value uses the intra prediction mode obtained in step S172 execution intra prediction, and generates forecast image.It will generate Forecast image be supplied to switch 174.
On the other hand, in the case where determining that it is not intraframe coding in step S181, present treatment proceeds to step S187。
In step S187, motion prediction/compensating unit 173 obtains inter-frame forecast mode from lossless decoding unit 162 Information, reference frame information, motion vector information etc..
In step S188, motion prediction/compensating unit 173 executes interframe movement prediction.Due to this processing, generation Forecast image is output to switch 174.
Therefore, prior to being handled in frame, by the picture coding device 51 in Fig. 1 and the picture decoding apparatus in Figure 22 151, Filtering processing is executed to the neighborhood pixels for intra prediction using the filter factor for image adaptive being arranged.For example, filter Wave system number is arranged according to intra prediction mode or quantization parameter.
Therefore, the noise remove corresponding to image and bit rate can be executed.As a result, it is possible to increase forecasting efficiency.
Figure 28 indicates to apply the configuration of one embodiment of learning device of the invention.In example in Figure 28, learn Device 251 is practised to handle using the study that training image signal executes filter factor.
It note that training image signal is the test image for obtaining filter factor, and can be used, for example, can With the standardized standard sequence for image compression encoding obtained in www.vqeg.org.It is corresponded to alternatively, can also use In the input picture of each application.For example, can be used in the case where input is camera signal using CCD or CMOS The baseband signal of sensor imaging executes study.
The picture coding device 51 in learning device 251 and Fig. 1 in Figure 28, which is had in common that, converts list with A/D Member 61, picture rearrange buffer 62, computing unit 63, orthogonal transform unit 64, quantifying unit 65, lossless coding unit 66, storage buffer 67, inverse quantization unit 68, inverse orthogonal transformation unit 69, computing unit 70, de-blocking filter 71, frame storage Device 72, switch 73, intraprediction unit 74, motion prediction/compensating unit 76, forecast image selecting unit 77 and rate control list Member 78.
In addition, picture coding device 51 in learning device 251 and Fig. 1 the difference is that: for the signal used, Using training image signal, also, including neighborhood pixels interpolation filtering computing unit 261, rather than neighborhood pixels interpolation filtering Switch unit 75.
Specifically, only executing study using only including the block in I picture for learning device 251.Alternatively, right In learning device 251, study only is executed using only including the block in intra-frame macro block in B picture and P picture.The former is than rear Person needs the less calculation amount for study.In addition, in the case where the former, for including that the block in I picture obtains Coefficient can be suitable only for the block for including in I picture, or can be adapted in the frame for including in B picture and P picture Macro block.
That is, executing the study only by the intra prediction using intraprediction unit 74 for learning device 251.Cause This, we will say, motion prediction/compensating unit 76 does not work actually.
In addition, the neighborhood pixels interpolation filtering computing unit 261 in Figure 29 is cut with the neighborhood pixels interpolation filtering in Figure 20 It changes unit 75 to have in common that, there is prediction mode buffer 121, optimum filtering computing unit 122 and low-pass filtering to set Set unit 123.
On the other hand, the neighborhood pixels interpolation filtering computing unit 261 in Figure 29 and the neighborhood pixels interpolation in Figure 20 are filtered Wave switch unit 75 the difference is that, be added to filter factor storage unit 271, also, Rate control unit will be come from 78 quantization parameter is supplied to optimum filtering computing unit 122.
Specifically, in example in Figure 29, it is identical by with the example in Figure 20 the case where in a manner of, in intra prediction Switch 101 is set between unit 74 and neighborhood pixels interpolation filtering switch unit 75, also, intraprediction unit 74 is according to switch 101 open/close executes intra prediction twice.
That is, for intraprediction unit 74, in the closed state of switch 101, it is fixed in H.264/AVC to execute The intra prediction of justice, also, it is directed to each fragment computations filter factor optimal for intra prediction mode and quantization parameter. The filter factor gone out for each fragment computations is stored in filter factor storage unit 271.Then, in switch 101 In opening state, use in the filter factor of each calculated segment by neighborhood pixels interpolation filtering switch unit 75 The filtering condition of setting executes intra prediction.
Via storage medium or network etc., the filter factor stored in the filter factor storage unit 271 is stored in The filtering of picture decoding apparatus 151 in the filter factor memory 94 (Figure 14) and Figure 22 of picture coding device 51 in Fig. 1 Coefficient memory 194 (Figure 23).
[description of the intra-prediction process in study processing]
Next, by being filled referring to the flow chart description in Figure 30 as the study in Figure 28 of a process of study processing Set the intra-prediction process of 251 execution.It note that and study is handled, in addition to the prediction in step S21 is handled by step S30 In intra-prediction process substitution other than, learning device 251 essentially performs processing identical with the coded treatment in Figure 17.
In addition, the step S201 to S203 and S206 to S209 in Figure 30 essentially perform with step S101 to S103 and The identical processing of S105 to S108, thus, it will be omitted and repeated.That is, in step S204 in Figure 30, optimum filtering Computing unit 122 calculate make the smallest current clip of the residual error of whole fragment each intra prediction mode filter factor and Corresponding quantization parameter, as optimum filtering coefficient.Calculated filter factor is supplied to filter factor storage unit 271.
In step S205, filter factor storage unit 271 stores the filtering system supplied from optimum filtering computing unit 122 Number.
Filter factor of the low-pass filtering setting unit 123 from the current clip being stored in filter factor storage unit 271 Middle setting corresponds to the filter factor of current block, connects the terminal of switch 101, and the filter factor of setting is supplied to neighbouring Pixel setting unit 111.
Therefore, in step S206, using the filter factor of setting come to the neighbouring of the current block from frame memory 72 Pixel value executes filtering processing.
Certainly, in a manner of identical with the example in Figure 21, the processing for repeating above-mentioned step S204 to S207 can be obtained The filter factor that must be advanced optimized.
As described previously for learning device 251, coded treatment phase training image signal execution and actually used is used Same processing, also, will thus calculated filter factor store into filter factor storage unit 271.Thus, it is possible to obtain Optimum filtering coefficient.
Via storage medium or network etc., the filter factor stored in the filter factor storage unit 271 is stored in The filtering of picture decoding apparatus 151 in the filter factor memory 94 (Figure 14) and Figure 22 of picture coding device 51 in Fig. 1 Coefficient memory 194 (Figure 23).
In addition, using picture coding device 251, as described previously for the block for including in I picture (or in B picture and P The intra-frame macro block for including in picture) obtain coefficient can be suitable only for the block in I picture included.Alternatively, the coefficient is not only It can be adapted for the block for including in I picture, can be applicable to the intra-frame macro block for including in B picture and P picture.
Alternatively, can realize high coding effect with the picture coding device 51 in Fig. 1 and the picture decoding apparatus in Figure 22 151 Rate.
It note that the above-mentioned orthogonal transform unit 64 and inverse orthogonal transformation list using the picture coding device 51 in Fig. 1 The inverse orthogonal transformation unit 164 of picture decoding apparatus 151 in 69 and Figure 22 of member executes defined in H.264/AVC just Alternation changes/inverse orthogonal transformation.Alternatively, such arrangement can be carried out: where the positive alternation of the picture coding device 51 in Fig. 1 The inverse orthogonal transformation unit 164 for changing the picture decoding apparatus 151 in unit 64 and inverse orthogonal transformation unit 69 and Figure 22, holds Orthogonal transformation/inverse orthogonal transformation that row proposes in non-patent literature 1.
Therefore, it can be further improved the code efficiency of the format proposed in non-patent literature 1.
Although note that the example for having been described execute intra prediction in the above description, the present invention can also With the intra prediction in the re prediction that is proposed suitable for non-patent literature 2.
<2. second embodiment>
[the other configurations example of picture coding device]
Figure 31 shows matching for another embodiment for being used as the picture coding device for applying image processing apparatus of the invention It sets.
Picture coding device 51 in picture coding device 351 and Fig. 1 is had in common that with A/D converting unit 61, picture rearranges buffer 62, computing unit 63, orthogonal transform unit 64, quantifying unit 65, lossless coding unit 66, storage buffer 67, inverse quantization unit 68, inverse orthogonal transformation unit 69, computing unit 70, de-blocking filter 71, frame storage Device 72, switch 73, intraprediction unit 74, motion prediction/compensating unit 76, forecast image selecting unit 77 and rate control list Member 78.
In addition, picture coding device 51 in picture coding device 351 and Fig. 1 the difference is that, be omitted neighbouring Pixel interpolating filters switch unit 75, also, is added to re prediction unit 361 and neighborhood pixels interpolation filtering switch unit 362。
That is, intraprediction unit 74 executes H.264/AVC intra prediction for the example in Figure 31.
On the other hand, motion prediction/compensating unit 76 is all based on the image handled for interframe and reference picture detection Candidate inter-frame forecast mode motion vector, processing is compensated to reference picture based on motion vector, and generate prediction Image.
Motion prediction/compensating unit 76 supplies the motion vector information detected, for interframe to re prediction unit 361 Information (address etc.) He Yici residual error of the image of processing, which is the prediction of the image for inter-prediction and generation Difference between image.
Motion prediction/compensating unit 76 is secondary pre- to determine by comparing the quadratic residue from re prediction unit 361 Best intra prediction mode in survey.In addition, motion prediction/compensating unit 76 comes really by comparing quadratic residue and a residual error It is fixed that quadratic residue encode or a residual error is encoded.It note that and all candidate inter-frame forecast modes are held The row processing.
Motion prediction/compensating unit 76 calculates cost function value for all candidate inter-frame forecast modes.In addition, using The residual error determined for each inter-frame forecast mode in residual sum quadratic residue calculates cost function value.Movement is pre- The prediction mode of the minimum value generated in calculated cost function value is determined as best inter-prediction mould by survey/compensating unit 76 Formula.
Motion prediction/compensating unit 76 by the forecast image generated in best inter-frame forecast mode (or be used for interframe Difference between the image and quadratic residue of prediction) and its cost function value be supplied to forecast image selecting unit 77.In prognostic chart In the case where having selected the forecast image generated in best inter-frame forecast mode as selecting unit 77, motion prediction/compensation list Member 76 exports the information for indicating best inter-frame forecast mode to lossless coding unit 66.At this point, motion vector information, reference Frame information, instruction to execute the re prediction mark of re prediction, information of intra prediction mode in re prediction etc. also by It is output to lossless coding unit 66.
Lossless coding unit 66 also carries out such as variable-length to the information from motion prediction/compensating unit 76 and compiles The lossless coded treatment of code, arithmetic coding etc., and be inserted into the head of compression image.
Letter based on the motion vector information from motion prediction/compensating unit 76 and the image that carry out interframe processing Breath, re prediction unit 361 are read and the neighbouring current neighborhood pixels of current block that carry out interframe processing from frame memory 72. In addition, re prediction unit 361 reads associated with current block with by motion vector information and refers to from frame memory 72 The neighbouring reference neighborhood pixels of block.
Re prediction unit 361 executes re prediction processing.Re prediction processing is such processing: where primary Intra prediction is executed between difference between the current neighborhood pixels of residual sum and reference neighborhood pixels, so that it is (secondary residual to generate second difference Difference) information.
Now, re prediction processing will be described referring to Figure 32.
For the example in Figure 32, present frame and reference frame are shown, wherein show current block A in the current frame.
In the case where obtaining motion vector mv (mv_x, mv_y) in the reference frame and present frame for current block A, calculate Current block A and pass through the difference information (residual error) between associated with current block A piece of motion vector mv.
For re prediction system, difference information related with current block A is not only calculated, but also calculates neighbouring current block The neighborhood pixels group R of A and believed by difference between motion vector mv neighborhood pixels group R1 associated with neighborhood pixels group R Breath.
That is, obtaining the coordinate of neighborhood pixels group R from the top-left coordinates (x, y) of current block A.In addition, from fortune is passed through Associated with current block A piece of moving vector mv of top-left coordinates (x+mv_x, y+mv_y) obtain the coordinate of neighborhood pixels group R1.From The difference information of these coordinate values calculating neighborhood pixels group.
For re prediction system, in the difference information about current block calculated in this way and the difference about reference pixel Divide between information the intra prediction executed according to H.264/AVC format, to generate second order difference information.To the second difference of generation Divide information to carry out orthogonal transformation and quantization, be encoded together with compression image, and is sent to decoding side.
Before the re prediction, 361 use of re prediction unit is arranged by neighborhood pixels interpolation filtering switch unit 362 Filter factor to execute filtering processing to the current neighborhood pixels for intra prediction and with reference to the difference between neighborhood pixels.Two Then secondary predicting unit 361 is come using the current neighborhood pixels by filtering processing and with reference to the difference of the filtering between neighborhood pixels Re prediction processing is executed, and second order difference information (quadratic residue) is output to motion prediction/compensating unit 76.
That is, re prediction unit 361 is configured to include intraprediction unit 74 etc. shown in Figure 14.
Neighborhood pixels interpolation filtering switch unit 362 substantially with the neighborhood pixels interpolation filtering switch unit in Fig. 1 75 identical modes configure, and execute identical processing.That is, 362 basis of neighborhood pixels interpolation filtering switch unit Intraprediction mode information from re prediction unit 361 and the quantization parameter from Rate control unit 78 filter to be arranged Coefficient, and the filter factor of setting is supplied to re prediction unit 361.
It note that the coded treatment of the picture coding device 351 in Figure 31 is executed with by the picture coding device 51 in Fig. 1 Figure 16 in the difference of coded treatment be only in that processing and motion prediction process in following frames, also, it is other Handle it is substantially the same, therefore, will the descriptions thereof are omitted.
That is, as being handled in frame, being executed according to H.264/AVC lattice using the picture coding device 351 in Figure 31 The intra prediction of formula.In addition, in motion prediction process, use is switched by neighborhood pixels interpolation filtering as motion prediction process The filter factor that unit 362 is arranged, to generate second order difference information.It is selected from first difference information and second order difference information Better second order difference information, also, best intra prediction mode is determined by comparing cost function value.
Picture decoding apparatus will be described referring to Figure 33, picture decoding apparatus reception is encoded by the picture coding device 351 Compression image and decode it.
[other examples of picture decoding apparatus]
Figure 33 indicates matching for another embodiment for being used as the picture decoding apparatus for applying image processing apparatus of the invention It sets.
Picture decoding apparatus 151 in picture decoding apparatus 401 and Figure 22 is had in common that comprising storage buffer 161, lossless decoding unit 162, inverse quantization unit 163, inverse orthogonal transformation unit 164, computing unit 165, de-blocking filter 166, picture rearrange buffer 167, D/A converting unit 168, frame memory 169, switch 170, intraprediction unit 171, Motion prediction/compensating unit 173 and switch 174.
In addition, picture decoding apparatus 151 in picture decoding apparatus 401 and Figure 22 the difference is that, neighbour is omitted Nearly pixel interpolating filters switch unit 172, also, is added to re prediction unit 411 and neighborhood pixels interpolation filtering switching list Member 412.
That is, the information of the instruction intra prediction mode obtained will be decoded by correct information from lossless Decoding unit 162 is supplied to intraprediction unit 171.Based on the information, intraprediction unit 171 generates forecast image and will produce Raw forecast image is output to switch 174.
Prediction mode information, motion vector information and reference in the information obtained will be decoded by correct information Frame information etc. is supplied to motion prediction/compensating unit 173 from lossless decoding unit 162.In addition, applying two to current block In the case where secondary prediction processing, the frame mode of re prediction mark and re prediction that instruction will execute the re prediction is believed Breath is also supplied to motion prediction/compensating unit 173 from lossless decoding unit 162.
In the case where determining using re prediction processing, motion prediction/compensating unit 173 controls re prediction unit 411, so that executing re prediction with the intra prediction mode that the intraprediction mode information of re prediction indicates.
Motion prediction/compensating unit 173 is based on motion vector information and reference frame information and carries out motion prediction to the image And compensation deals, and generate forecast image.That is, using reference block associated with current block in reference block Pixel value generates the forecast image of current block.Motion prediction/compensating unit 173 then will be from the pre- of re prediction unit 411 Error of measurement score value is added with the forecast image of generation, and these are output to switch 174.
Re prediction unit 411 is used from the current neighborhood pixels that frame memory 169 is read and with reference between neighborhood pixels Difference execute re prediction.Before the re prediction, 411 use of re prediction unit is switched by neighborhood pixels interpolation filtering The filter factor that unit 412 is arranged to execute filtering processing to current neighborhood pixels and with reference to the difference between neighborhood pixels.It is secondary Predicting unit 411 then using by filtering processing current neighborhood pixels and with reference to the difference between neighborhood pixels it is secondary to execute Prediction processing, and the second order difference information (quadratic residue) of acquisition is output to motion prediction/compensating unit 173.
That is, re prediction unit 411 is configured to include intraprediction unit 171 etc. shown in Figure 26.
Neighborhood pixels interpolation filtering switch unit 412 is substantially with identical as neighborhood pixels interpolation filtering switch unit 172 Mode configure.That is, neighborhood pixels interpolation filtering switch unit 412 is arranged and from lossless decoding unit 162 The corresponding filter factor of at least one of quantization parameter and intra prediction mode.Neighborhood pixels interpolation filtering switch unit 412 The filter factor of setting is supplied to re prediction unit 411.
It note that the decoding process of the picture decoding apparatus 401 in Figure 33 is held with by the picture decoding apparatus 151 in Figure 22 The difference of decoding process in capable Figure 24 is only in that processing and motion prediction process in following frames, also, other Processing be substantially the same processing, therefore, will the descriptions thereof are omitted.
That is, as being handled in frame, being executed according to H.264/AVC lattice using the picture decoding apparatus 401 in Figure 33 The intra prediction of formula.In addition, using the filter being arranged by neighborhood pixels interpolation filtering switch unit 412 as motion prediction process Wave system number executes re prediction (intra prediction) in motion prediction process, and generates second order difference information.
It it may also be possible to apply the invention for the intra prediction in re prediction processing as described above.
It note that in the above description, describe such example: before intra prediction, being adapted to using to image Neighborhood pixels for intra prediction are executed filtering processing by the filter factor of ground setting.
Now, the noise for including in the neighborhood pixels for intra prediction system is joined according to the content of such as image, quantization The encoding condition of number etc. is different.Accordingly, there exist improved by executing the filtering processing for example executed in H.264/AVC format The block of code efficiency and be not such block.
In any case, when executing processing in the frame based on 8 × 8 pieces to macro block with H.264/AVC format, to all blocks Perform blank filtering processing (blanket filtering processing), thus will appear code efficiency thus reduce Block.
Therefore, the feelings opened/closed that the neighborhood pixels for intra prediction are executed with filtering processing are described below The example of condition.
<3. 3rd embodiment>
[the other configurations example of picture coding device]
Figure 34 shows matching for another embodiment for being used as the picture coding device for applying image processing apparatus of the invention It sets.
Picture coding device 51 in picture coding device 451 and Fig. 1 is had in common that with A/D converting unit 61, picture rearranges buffer 62, computing unit 63, orthogonal transform unit 64, quantifying unit 65, lossless coding unit 66, storage buffer 67, inverse quantization unit 68, inverse orthogonal transformation unit 69, computing unit 70, de-blocking filter 71, frame storage Device 72, switch 73, intraprediction unit 74, motion prediction/compensating unit 76, forecast image selecting unit 77 and rate control list Member 78.
In addition, picture coding device 51 in picture coding device 451 and Fig. 1 the difference is that, use neighborhood pixels Interpolation filtering control unit 461 is instead of neighborhood pixels interpolation filtering switch unit 75.
That is, neighborhood pixels interpolation filtering control unit 461 is executed to opening or closing of being filtered of blank Control executes all pieces of neighborhood pixels when executing processing in the frame based on 8 × 8 pieces to macro block with H.264/AVC format Blank filtering processing.It is performed although note that and be only used to processing in the H.264/AVC frame based on 8 × 8 pieces of format Filtering processing still also executes the filtering in intra-frame 4 * 4 and frame 16 × 16 with neighborhood pixels interpolation filtering control unit 461 Processing.
The control signal that opens/closes from neighborhood pixels interpolation filtering control unit 461 is supplied to intra prediction list Member 74.
Intraprediction unit 74 is based on the figure that carry out intra prediction for rearranging the reading of buffer 62 from picture Picture and the intra-prediction process that all candidate intra prediction modes are executed from the reference picture that frame memory 72 is supplied.At this point, At intraprediction unit 74, before intra prediction, believed according to the control from neighborhood pixels interpolation filtering control unit 461 Number be switched on or switched off filtering processing to execute intra prediction, also, using as a result there is lesser calculated cost The intra prediction mode of functional value.
In addition, intraprediction unit 74 generates the mark of instruction filtering processing opened or closed.It is selected in forecast image In the case where the forecast image that the selection of unit 77 is generated with best intra prediction mode, will the flag information with indicate in optimum frame The information of prediction mode is supplied to lossless coding unit 66 together.
[configuration example of intraprediction unit]
Figure 35 is the block diagram of the detailed configuration example of the intraprediction unit 74 in pictorial image 34.Example in Figure 35 In the case where, intraprediction unit 74 generates unit 471 by forecast image, cost function value generates unit 472 and mode and beats Opening/closing flag generating unit 473 is constituted.
The neighborhood pixels value that unit 471 supplies the current block of intra prediction is generated from frame memory 72 to forecast image.? In the case where in Figure 35, switch 73 is omitted in Cong Tuzhong, still, in fact, neighborhood pixels value is stored via switch 73 from frame Device 72 is supplied to forecast image to generate unit 471.It note that in the case where intra prediction, without de-blocking filter 71 The pixel value of deblocking filtering is used as neighborhood pixels value.
Based on the control signal from neighborhood pixels interpolation filtering control unit 461, forecast image generates unit 471 with institute The candidate intra prediction mode having executes intra prediction, executes filtering processing to neighborhood pixels value, or do not execute filtering processing, To generate forecast image.As described in later in reference to Figure 36 to Figure 38, execute by increment of block or by increment of macro block by controlling The neighborhood pixels interpolation filtering control unit 461 of signal designation opens/closes control.
Forecast image generates unit 471 and is supplied into the forecast image pixel value and its intraprediction mode information of generation This functional value generates unit 472.
Buffer 62, which is rearranged, from picture generates the supply original image pixels value of unit 472 to cost function value.For Filtering processing is the case where opening and closing, and cost function value generates unit 472 and uses original image pixels value and forecast image Pixel value calculates the cost function value of each intra prediction mode.Cost function value generate unit 472 by it is calculated at This functional value, forecast image pixel value and intraprediction mode information are supplied to mode and open/close flag generating unit 473。
Mode uses the cost function that unit 472 is generated from cost function value with flag generating unit 473 is opened/closed Value come determine best intra prediction mode and its setting filtering processing opening and closing, and generate instruction filtering processing be beat What on or off was closed opens/closes flag information.
It mode and opens/closes flag generating unit 473 and supplies the forecast image pixel value of best intra prediction mode To forecast image selecting unit 77.The case where forecast image selecting unit 77 selects the forecast image of best intra prediction mode In, mode beats ON/OFF by the information for indicating best intra prediction mode and accordingly with flag generating unit 473 is opened/closed It closes flag information and is supplied to lossless coding unit 66.
It note that other than the intra prediction next described, the processing executed by picture coding device 451 is substantially It is identical as the processing of picture coding device 51 in Fig. 1, thus, it will be omitted and repeated.
Next, by being executed in the case where opening or closing control referring to the description of the flow chart of Figure 36 by increment of block The intra-prediction process executed by the intraprediction unit 74 in Figure 34.It note that the processing is in step S31 in Figure 17 Another example of intra-prediction process, also, in the example of Figure 36, the example of intra-frame 4 * 4 will be described.In addition, below In, filtering processing, which opens/closes, to be also briefly termed as filter and opens/closes.
In step S401, for each in nine kinds of intra prediction modes shown in Fig. 3 or Fig. 4, cost function Value generates the cost function value that unit 472 generates current block.
That is, generating the neighbouring of the current block of the supply intra prediction of unit 471 from frame memory 72 to forecast image Pixel value.Forecast image generates unit 471 and executes frame to each in nine kinds of intra prediction modes shown in Fig. 3 or Fig. 4 Interior prediction, also, generate the forecast image of current block.
At this point, from neighborhood pixels interpolation filtering control unit 461 for being applied to not execute filtering processing to neighborhood pixels Signal is controlled, also, forecast image generates unit 471 and neighborhood pixels are executed with the intra prediction for being not carried out filtering processing.Please Note that here it is possible to carry out such arrangement: where the control signal that neighborhood pixels are executed with filtering processing is realized in supply. It is noted, however, that for nine kinds of modes without different control, for example, not executed to vertical execute to level;Definitely It says, for nine kinds of modes, carries out about the identical control for executing or not executing.Further, it is noted that here, not to all moulds Formula, which executes filtering processing, allows less calculation amount.
Forecast image generates unit 471 and the forecast image pixel value and intraprediction mode information of generation is supplied to cost Functional value generates unit 472.In the case where filtering is closing, for each intra prediction mode, cost function value is generated Unit 471 is above-mentioned to calculate using the original image pixels value and forecast image pixel value for rearranging buffer 62 from picture Expression formula (73) or expression formula (74) shown in cost function value.Cost function value generate unit 472 by it is calculated at This functional value, forecast image pixel value and intraprediction mode information are supplied to mode and open/close flag generating unit 473。
In step S402, mode is used with flag generating unit 473 is opened/closed from cost function value generation unit 472 cost function value selects the best intra prediction mode for current block.The intraprediction mode information of selection is supplied Neighborhood pixels interpolation filtering control unit 461 should be given.
In step S403, neighborhood pixels interpolation filtering control unit 461 is filtering cost function value generation unit 472 Wave generates the cost function value of the intra prediction mode of selection in the case where opening and closing.It note that in step S401, produce Raw filtering is the cost function value closed, therefore, in fact, generating the cost function value that filtering is opened in step S403.
That is, neighborhood pixels interpolation filtering control unit 461 opens filtering in control signal and the frame of selection in advance Surveying pattern information is supplied to forecast image to generate unit 471.Forecast image generates unit 471 to the intra prediction mould for selection The neighborhood pixels of formula execute filtering processing, execute intra prediction with the intra prediction mode of selection, and generate the pre- of current block Altimetric image.
Forecast image generates unit 471 and supplies the forecast image pixel value of generation and the intraprediction mode information of selection Unit 472 is generated to cost function value.It is open in situation in filtering, for the intra prediction mode of selection, cost function Value is generated unit 471 and is counted using the original image pixels value and forecast image pixel value that rearrange buffer 62 from picture Count cost function value shown in the expression formula (73) stated or expression formula (74) in.Cost function value generates unit 472 and will calculate Cost function value, forecast image pixel value out is supplied to mode and opens/closes flag generating unit 473.
In step s 404, mode and open/close flag generating unit 473 by comparing selection intra prediction mode Filtering be that the cost function value opened and closed determines that the filtering of current block opens/closes.That is, being to beat in filtering It in the lesser situation of the cost function value opened, determines that filtering is opened for current block, also, is the cost function closed in filtering It is worth in lesser situation, determines that filtering is closed for current block.Mode and open/close flag generating unit 473 and then will be true Fixed prognostic chart picture value is supplied to forecast image selecting unit 77.
In step S405, mode and open/close flag generating unit 473 generate instruction current block in step S404 Middle decision open or close open/close mark.For example, opening/closing filter value is 1 in the case where filtering opening. In the case where filtering closing, opening/closing filter value is 0.
It, will in the case where having selected the forecast image in intra prediction mode in the step S22 in above-mentioned Figure 16 The flag information that opens/closes generated is supplied to lossless coding unit together with the information for indicating best intra prediction mode 66.The information of supply is encoded in the step S23 in Figure 16, is added with the head of compression image, and decoding side is sent to.
Next, by being executed in the case where opening or closing control referring to the description of the flow chart of Figure 37 by increment of block By another example for the intra-prediction process that the intraprediction unit 74 in Figure 34 executes.In the case where the example of Figure 37, The example of intra-frame 4 * 4 will be described.
In step S421, for each intra prediction mode, it is to beat that cost function value, which generates unit 472 according to filtering, The open and close cost function value to generate current block.
That is, generating the neighbouring of the current block of the supply intra prediction of unit 471 from frame memory 72 to forecast image Pixel value.Forecast image generates unit 471 and executes frame to each in nine kinds of intra prediction modes shown in Fig. 3 or Fig. 4 Interior prediction, also, generate the forecast image of current block.
At this point, firstly, from neighborhood pixels interpolation filtering control unit 461 for being applied to not execute filtering place to neighborhood pixels The control signal of reason, also, forecast image generates unit 471 and is not carried out in each frame of filtering processing to neighborhood pixels Prediction mode executes intra prediction, and generates forecast image.In addition, supplying reality from neighborhood pixels interpolation filtering control unit 461 Now neighborhood pixels are executed with the control signal of filtering processing, also, forecast image generates unit 471 to neighborhood pixels to perform Each intra prediction mode of filtering processing executes intra prediction, and generates forecast image.
Forecast image generates unit 471 for each intra prediction mode in the case where filtering is to open and close Information and corresponding forecast image pixel value are supplied to cost function value to generate unit 472.It is to close and open every in filtering In a kind of situation, for each intra prediction mode, it is slow using rearranging from picture that cost function value generates unit 471 Original image pixels value and the forecast image pixel value of device 62 is rushed to show to calculate in above-mentioned expression formula (73) or expression formula (74) Cost function value out.It is in the every case closed and opened in filtering, cost function value generates unit 472 and will calculate Cost function value, forecast image pixel value and intraprediction mode information be supplied to mode and open/close mark generate it is single Member 473.
In step S422, mode is used with flag generating unit 473 is opened/closed from cost function value generation unit To determine, filtering should be opened to 472 cost function value for current block in the case where each intra prediction mode Or it closes.
In step S423, mode and open/close flag generating unit 473 from about determine filtering be on or off Intra prediction mode in select current block best intra prediction mode.
In step S424, mode generates the intra prediction that instruction is directed to selection with flag generating unit 473 is opened/closed The state (opening or closing) of the filter of mode opens/closes flag information.When in the step S22 in above-mentioned Figure 16 In the case where having selected the forecast image in intra prediction mode, generation is opened/closed into flag information and instruction optimum frame The information of inner estimation mode is supplied to lossless coding unit 66 together.The information of supply is compiled in the step S23 in Figure 16 Code is added with the head of compression image, is sent to decoding side.
Next, by being described in the case where opening or closing control as increment execution using macro block referring to the flow chart of Figure 38 The intra-prediction process executed by the intraprediction unit 74 in Figure 34.
It note that the processing is another example of the intra-prediction process in step S31 in Figure 17, also, in Figure 38 Example in, the example of intra-frame 4 * 4 will also be described.
In step S451, the filtering about entire macro block is fixed as closing by neighborhood pixels interpolation filtering control unit 461 It closes or opens.In this case, filtering is fixed as closing by neighborhood pixels interpolation filtering control unit 461, also, will filtering The control signal of closing is supplied to forecast image to generate unit 471.The fixation of filtering can be on or off, however, it is possible to The fixation of closing is realized with less calculation amount.
In step S452, intraprediction unit 74 determines each piece of intra prediction mode.That is, being stored from frame Device 72 generates the neighborhood pixels value that unit 471 supplies the current block of intra prediction to forecast image.Forecast image generates unit 471 Intra prediction is executed to each in nine kinds of intra prediction modes shown in Fig. 3 or Fig. 4, also, generates the pre- of current block Altimetric image.
At this point, not executing filtering place to neighborhood pixels firstly, realizing from the supply of neighborhood pixels interpolation filtering control unit 461 The control signal of reason, also, forecast image generates unit 471 and is executed in the frame on each intra prediction mode to neighborhood pixels Prediction generates forecast image without executing filtering processing.Forecast image generates unit 471 for the forecast image pixel of generation Value and its intraprediction mode information are supplied to cost function value to generate unit 472.
It is in the every case closed in filtering, for each intra prediction mode, cost function value generates unit 471 calculate above-mentioned table using the original image pixels value and forecast image pixel value that rearrange buffer 62 from picture Up to cost function value shown in formula (73) or expression formula (74).It is cost function value in the every case closed in filtering Generate unit 472 by calculated cost function value, forecast image pixel value and intraprediction mode information be supplied to mode and Open/close flag generating unit 473.
Mode uses the cost function that unit 472 is generated from cost function value with flag generating unit 473 is opened/closed Value determines each piece of best intra prediction mode.It is supplied to neighborhood pixels interpolation to filter decoded intraprediction mode information Wave control unit 461.
In step S453, neighborhood pixels interpolation filtering control unit 461 is directed to cost function value generation unit 472 Entire macro block generates the cost function value that filtering opens and closes.It note that being produced in step S452 in filtering is to close In the case where for each piece (that is, entire macro block) in macro block best intra prediction mode cost function value.Therefore, real On border, in step S453, the cost function value for the entire macro block being open in situation in filtering is generated.
That is, filtering is opened control signal and is directed to each piece really by neighborhood pixels interpolation filtering control unit 461 The information of fixed intra prediction mode is supplied to forecast image to generate unit 471.Forecast image generates unit 471 to for determining The neighborhood pixels value of intra prediction mode execute filtering processing, intra prediction is executed with determining intra prediction mode, and Generate the forecast image of current block.
Forecast image generates unit 471 and supplies the forecast image pixel value of generation and the intraprediction mode information of determination Unit 472 is generated to cost function value.It is open in every case in filtering, for determining intra prediction mode, at This functional value generates unit 471 and uses the original image pixels value and forecast image pixel for rearranging buffer 62 from picture Value calculates cost function value shown in above-mentioned expression formula (73) or expression formula (74).It is to close and open every in filtering In a kind of situation, cost function value generates unit 472 for calculated cost function value, forecast image pixel value and intra prediction Pattern information is supplied to mode and opens/closes flag generating unit 473.
In step S454, mode compares with flag generating unit 473 is opened/closed from cost function value generation unit All pieces of cost function value in 472 macro block in the case where filtering is to open and close, and determine to entire macro Block which of is opened/closed using filtering.
In step S455, for entire macro block, mode generates instruction in step with flag generating unit 473 is opened/closed Determined in rapid S454 open or close open/close mark.For each macro block, generation is opened/closed into mark Information is supplied to lossless coding unit 66.The information of supply is encoded in the step S23 in Figure 16, the head with compression image It is added, is sent to decoding side.
As described above, the control that filtering is opened/closed and (turned on or off) can be executed by increment of block, alternatively, can To be executed by increment of macro block.Although note that can be improved at intra prediction by opening/closing using block as increment control algorithm The precision of prediction of reason, but information content needed for the flag information for transmitting each piece increases.On the contrary, using macro block as increment In the case where being controlled, the raising of precision of prediction is lower than the raising of the precision of prediction executed using block as increment, still, for every One flag information of one macro block is enough, it is thus possible to reduce the increase of flag information amount.
Although describing the example of luminance signal in the above description, this be can be used for about color difference signal Intra prediction.In addition, the filter factor for the filtering processing to be controlled be not limited in H.264/AVC format three taps 1,2, // 4, also, this any coefficient for can be adapted for any tap length being arranged with the picture coding device 51 in Fig. 11 }.
That is, filter open in the case where, can also be performed with being arranged by the picture coding device 51 in Fig. 1 The filtering processing of filter factor.
Picture decoding apparatus will be described referring to Figure 39, picture decoding apparatus reception is encoded by the picture coding device 451 Compression image and decode it.
[other examples of picture decoding apparatus]
Figure 39 diagram is used as matching for another embodiment for applying the picture decoding apparatus of image processing apparatus of the invention It sets.
Picture decoding apparatus 151 in picture decoding apparatus 501 and Figure 22 is had in common that comprising storage buffer 161, lossless decoding unit 162, inverse quantization unit 163, inverse orthogonal transformation unit 164, computing unit 165, de-blocking filter 166, picture rearrange buffer 167, D/A converting unit 168, frame memory 169, switch 170, intraprediction unit 171, Motion prediction/compensating unit 173 and switch 174.
In addition, picture decoding apparatus 151 in picture decoding apparatus 501 and Figure 22 the difference is that, with neighbouring picture Plain interpolation filtering control unit 511 is instead of neighborhood pixels interpolation filtering switch unit 172.
That is, the information of the instruction intra prediction mode obtained will be decoded by correct information from lossless Decoding unit 162 is supplied to intraprediction unit 171.Based on the information, intraprediction unit 171 generates forecast image and will produce Raw forecast image is output to switch 174.At this time.Before intra prediction, intraprediction unit 171 is according to from neighborhood pixels The control signal of interpolation filtering control unit 511 executes (or not executing) filtering processing to neighborhood pixels value.
According to the coding of picture coding device 451, controlled from lossless decoding unit 162 to neighborhood pixels interpolation filtering single What member 511 supplied each macro block or each piece opens/closes flag information.
Neighborhood pixels interpolation filtering control unit 511 opens/closes flag information to intraprediction unit according to supply The control signal for executing or not executing filtering processing is realized in 171 supplies.
It note that using the picture coding device 451 in Figure 34, be that the two kinds of situations opened and closed are surveyed to filtering Examination, also, after a kind of situation for having selected the higher code efficiency of generation by cost function value, it executes at intra prediction Reason.On the other hand, using picture decoding apparatus 501, the flag information control filtering that opens/closes of the transmission based on coding is opened Or it closes, also, execute intra-prediction process.
[configuration example of intraprediction unit and neighborhood pixels interpolation filtering control unit]
Figure 40 is the block diagram for illustrating the detailed configuration example of neighborhood pixels interpolation filtering control unit and intraprediction unit.
In the case where the example of Figure 40, intraprediction unit 171 is generated by prediction mode buffer 521 and forecast image Unit 522 is constituted.Neighborhood pixels interpolation filtering control unit 511 is by mark buffer 531 and control 532 structure of signal generation unit At.
The intraprediction mode information from lossless decoding unit 162 is supplied to prediction mode buffer 521.It is deposited from frame Reservoir 169 generates the neighborhood pixels value that unit 522 supplies the current block of intra prediction to forecast image.Situation in Figure 40 In, switch 170 is also omitted from figure, still, in fact, neighborhood pixels value is supplied via switch 170 from frame memory 169 Unit 522 is generated to forecast image.
Forecast image generates the intraprediction mode information that unit 522 reads current block from prediction mode buffer 521, with The intra prediction mode of reading executes intra prediction to current block, and generates forecast image.Before the intra prediction, prediction Image generation unit 522 is according to the control signal from control signal generation unit 532 to the prognostic chart from frame memory 169 As pixel value executes filtering processing.
For each macro block or each piece, ON/OFF is beaten from lossless decoding unit 162 to the mark supply of buffer 531 Close flag information.Control signal generation unit 532 reads from mark buffer 531 and opens/closes mark accordingly, generates instruction Filtering processing is executed for each piece and does not still execute the control signal of filtering processing, and the control signal of generation is supplied to Forecast image generates unit 522.
It note that the processing executed by picture decoding apparatus 501 is substantially other than the prediction processing next described It is identical as the processing of picture decoding apparatus 151 in Figure 22, thus, it will be omitted and repeated.
[description of prediction processing]
Next, the prediction of the picture decoding apparatus 501 described in Figure 39 referring to the flow chart of Figure 41 is handled.It please infuse Meaning, the intra-prediction process are another examples of the prediction processing in the step S138 in Figure 24.
In step S501, forecast image generates unit 522 and determines whether to have carried out intraframe coding to current block.It will be in frame Prediction mode information is supplied to prediction mode buffer 521 from lossless decoding unit 162, to generate unit by forecast image 522 read.Therefore, in step S501, forecast image generates unit 522 and determines to current block progress intraframe coding, also, this Processing proceeds to step S502.
In step S502, forecast image generates the intra prediction mode letter that unit 522 obtains prediction mode buffer 521 Breath.
In addition, mark buffer 531 will be supplied to from the flag information that opens/closes of lossless decoding unit 162 When, mark buffer 531 obtains forecast image pixel value mark in step S503 and stores it.
Control signal generation unit 532 reads from mark buffer 531 and opens/closes mark corresponding to current flag, Also, determination opens/closes whether mark is 1 in step S504.It is 1 when determination opens/closes mark in step S504, That is, filtering processing is open in situation, control signal generation unit 532 is supplied to forecast image to generate unit for signal is controlled 522, so that executing filtering processing.
According to control signal, in step S505, forecast image generate unit 522 using filter factor to neighborhood pixels into Row filtering processing.In step S506, forecast image is generated unit 522 and is held using the neighborhood pixels value that have passed through filtering processing Row intra prediction, and generate forecast image.
It on the other hand, is not 1 when determination opens/closes mark in step S504, that is, filtering processing is the case where closing In, the filtering processing of step S505 is skipped, also, present treatment proceeds to step S506.
In step S506, forecast image is generated unit 522 and is come using the forecast image pixel value from frame memory 169 Intra prediction is executed, and generates forecast image.
The forecast image generated in step S506 is supplied to switch 174.
On the other hand, in the case where determination does not execute intraframe coding in step S501, present treatment proceeds to step S507。
In step s 507, motion prediction/compensating unit 173 executes interframe movement prediction.That is, to be processed In the case where image is the image for inter-prediction processing, necessary image is read from frame memory 169, also, will be necessary Image be supplied to motion prediction/compensating unit 173 via switch 170.In step S508, motion prediction/compensating unit 173 Motion prediction is executed with inter-frame forecast mode based on the motion vector obtained in step s 507, and generates forecast image.It will The forecast image of generation is exported to switch 174.
As described above, being controlled using picture coding device 451 and picture decoding apparatus 501 for for intra prediction The opening and closing of the filtering processing of neighborhood pixels, also, filtering processing is not executed for the block of code efficiency deterioration.Therefore, Code efficiency can be improved.
Although note that the example for describing execute intra prediction in the above description, filtering processing is opened Control with closing can be adapted for the intra prediction in the re prediction described above in reference to Figure 32.
<4. fourth embodiment>
[the other configurations example of picture coding device]
Figure 42 shows matching for another embodiment for being used as the picture coding device for applying image processing apparatus of the invention It sets.
Picture coding device 451 in picture coding device 551 and Figure 34 is had in common that with A/D converting unit 61, picture rearranges buffer 62, computing unit 63, orthogonal transform unit 64, quantifying unit 65, lossless coding unit 66, storage buffer 67, inverse quantization unit 68, inverse orthogonal transformation unit 69, computing unit 70, de-blocking filter 71, frame storage Device 72, switch 73, intraprediction unit 74, motion prediction/compensating unit 76, forecast image selecting unit 77 and rate control list Member 78.
In addition, picture coding device 451 in picture coding device 551 and Figure 34 the difference is that, neighbour is omitted Nearly pixel interpolating filtering control unit 461, also, re prediction unit 361 and neighborhood pixels the interpolation filter being added in Figure 31 Wave control unit 561.
That is, for the example in Figure 42, H.264/AVC intraprediction unit 74 according to executing intra prediction.
On the other hand, motion prediction/compensating unit 76 is all based on the image handled for interframe and reference picture detection Candidate inter-frame forecast mode motion vector, processing is compensated to reference picture based on motion vector, and generate prediction Image.
Motion prediction/compensating unit 76 supplies the motion vector information detected, for interframe to re prediction unit 361 Information (address etc.) He Yici residual error of the image of processing, which is the prediction of the image for inter-prediction and generation Difference between image.
Motion prediction/compensating unit 76 is secondary pre- to determine by comparing the quadratic residue from re prediction unit 361 Best intra prediction mode in survey.In addition, motion prediction/compensating unit 76 comes really by comparing quadratic residue and a residual error It is fixed that quadratic residue encode or a residual error is encoded.It note that and all candidate inter-frame forecast modes are held The row processing.
Motion prediction/compensating unit 76 calculates cost function value for all candidate inter-frame forecast modes.At this point, using The residual error determined for each inter-frame forecast mode in residual sum quadratic residue calculates cost function value.Movement is pre- The prediction mode of the minimum value generated in calculated cost function value is determined as best inter-prediction mould by survey/compensating unit 76 Formula.
Motion prediction/compensating unit 76 by the forecast image generated in best inter-frame forecast mode (or be used for interframe Difference between the image and quadratic residue of prediction) and its cost function value be supplied to forecast image selecting unit 77.In prognostic chart In the case where having selected the forecast image generated in best inter-frame forecast mode as selecting unit 77, motion prediction/compensation list Member 76 exports the information for indicating best inter-frame forecast mode to lossless coding unit 66.
At this point, motion vector information, reference frame information, instruction will execute the re prediction mark, secondary pre- of re prediction The information etc. of intra prediction mode in survey is also output to lossless coding unit 66.Lossless coding unit 66 is also to coming from The information of motion prediction/compensating unit 76 carries out the lossless coded treatment of variable length code, arithmetic coding etc., and It is inserted into the head of compression image.
Letter based on the motion vector information from motion prediction/compensating unit 76 and the image that carry out interframe processing Breath, re prediction unit 361 are read and the neighbouring current neighborhood pixels of current block that carry out interframe processing from frame memory 72. In addition, re prediction unit 361 reads from frame memory 72 and passes through motion vector information with the associated reference block of current block Neighbouring reference neighborhood pixels.
Re prediction unit 361 is executed to be handled above with reference to Figure 32 re prediction described.Re prediction processing is in this way Processing: where execute intra prediction in the current neighborhood pixels of a residual sum and with reference between the difference between neighborhood pixels, thus Generate the information of second difference (quadratic residue).
It is noted, however, that the re prediction unit 361 in Figure 42 is according to from neighborhood pixels before the re prediction The control signal of interpolation filtering control unit 561 is between the reference neighborhood pixels and current neighborhood pixels for inter-prediction Difference executes (or not executing) filtering processing.Then re prediction unit 361 using current neighborhood pixels and refers to neighborhood pixels Between (or unfiltered) difference of filtering execute re prediction processing, and it is the second order difference information of acquisition is (secondary residual Difference) it is output to motion prediction/compensating unit 76.At this point, re prediction unit 361 will also indicate whether to execute beating for filtering processing Opening/closing flag information is output to motion prediction/compensating unit 76.
That is, re prediction unit 361 includes intraprediction unit 74 shown in Figure 35.
Neighborhood pixels interpolation filtering control unit 561 is substantially with identical as neighborhood pixels interpolation filtering control unit 461 Mode configure, and execute identical processing.That is, neighborhood pixels interpolation filtering control unit 561 will realize whether The control signal that the control of filtering processing is executed by increment of block or by increment of macro block is supplied to re prediction unit 361.
It note that other than processing in following frames and motion prediction process, by the picture coding device 551 in Figure 42 The processing of execution is substantially identical as the processing of the picture coding device 451 in Figure 34 (that is, coded treatment in Figure 16), from And by the descriptions thereof are omitted.
That is, as being handled in frame, being executed according to H.264/AVC lattice using the picture coding device 551 in Figure 42 The intra prediction of formula.In addition, as motion prediction process, in motion prediction process, according to from neighborhood pixels interpolation filtering The control signal of control unit 561, control filtering processing, to generate (or unfiltered) second order difference information of filtering.Choosing Select the difference information with preferable code efficiency in first difference information and second order difference information, and comparative costs function Value, so that it is determined that best inter-frame forecast mode.
Picture decoding apparatus will be described referring to Figure 43, picture decoding apparatus reception is encoded by the picture coding device 551 Compression image and decode it.
[the other configurations example of picture decoding apparatus]
Figure 43 diagram is used as matching for another embodiment for applying the picture decoding apparatus of image processing apparatus of the invention It sets.
Picture decoding apparatus 501 in picture decoding apparatus 601 and Figure 39 is had in common that comprising storage buffer 161, lossless decoding unit 162, inverse quantization unit 163, inverse orthogonal transformation unit 164, computing unit 165, de-blocking filter 166, picture rearrange buffer 167, D/A converting unit 168, frame memory 169, switch 170, intraprediction unit 171, Motion prediction/compensating unit 173 and switch 174
In addition, picture decoding apparatus 501 in picture decoding apparatus 601 and Figure 39 the difference is that, neighbour is omitted Nearly pixel interpolating filtering control unit 511, also, re prediction unit 411 and neighborhood pixels the interpolation filter being added in Figure 33 Wave control unit 611.
That is, the information of the instruction intra prediction mode obtained will be decoded by correct information from lossless Decoding unit 162 is supplied to intraprediction unit 171.Based on the information, intraprediction unit 171 generates forecast image and will produce Raw forecast image is output to switch 174.
Prediction mode information, motion vector information and reference in the information obtained will be decoded by correct information Frame information etc. is supplied to motion prediction/compensating unit 173 from lossless decoding unit 162.In addition, applying two to current block In the case where secondary prediction processing, by the frame mode information of re prediction mark and re prediction from lossless decoding unit 162 It is supplied to motion prediction/compensating unit 173.
In the case where determining using re prediction processing, motion prediction/compensating unit 173 controls re prediction unit 411, to execute re prediction with the intra prediction mode that the intraprediction mode information of re prediction indicates.
Motion prediction/compensating unit 173 is based on motion vector information and reference frame information and executes motion prediction to the image And compensation deals, and generate forecast image.That is, using reference block associated with current block in reference block Pixel value generates the forecast image of current block.Motion prediction/compensating unit 173 then will be from the pre- of re prediction unit 411 Error of measurement score value is added with the forecast image of generation, and these are output to switch 174.
Re prediction unit 411 is used from the current neighborhood pixels that frame memory 169 is read and with reference between neighborhood pixels Difference execute re prediction.It is noted, however, that being executed having received realization from neighborhood pixels interpolation filtering control unit 611 In the case where the control signal of the control of filtering processing, re prediction unit 411 is before the re prediction to current neighbouring picture The difference of element and reference neighborhood pixels executes filtering processing.Re prediction unit 411 is then using the current neighbour by filtering processing Nearly pixel and re prediction processing is executed with reference to the difference between neighborhood pixels, and the second order difference information of acquisition is (secondary Residual error) it is output to motion prediction/compensating unit 173.
It note that and having received the control that realization does not execute filtering processing from neighborhood pixels interpolation filtering control unit 611 Control signal in the case where, re prediction unit 411 does not execute filtering processing, and using current neighborhood pixels and with reference to adjacent The difference of nearly pixel executes secondary treatment.
That is, re prediction unit 411 is configured to include intraprediction unit 171 shown in Figure 40.
Neighborhood pixels interpolation filtering control unit 611 substantially with the neighborhood pixels interpolation filtering control unit in Figure 39 511 identical modes configure, and essentially perform identical processing.That is, from lossless decoding unit 162 to neighbouring Pixel interpolating filtering control unit 611 is supplied in the information obtained and to header information decoder and opens/closes flag information. Neighborhood pixels interpolation filtering control unit 611 is according to flag information supply control signal is opened/closed, so that re prediction list 411 pairs of neighborhood pixels of member execute or do not execute filtering processing.
It note that other than processing in following frames and motion prediction process, by the picture decoding apparatus 601 in Figure 43 The processing of execution is substantially identical as the processing of the picture decoding apparatus 501 in Figure 39 (that is, decoding process in Figure 24), from And by the descriptions thereof are omitted.
That is, as being handled in frame, being executed according to H.264/AVC lattice using the picture decoding apparatus 601 in Figure 43 The intra prediction of formula.In addition, as motion prediction process, in motion prediction process, according to from neighborhood pixels interpolation filtering The control signal of control unit 611 controls filtering processing, executes re prediction (intra prediction), and generates second order difference letter Breath.
The opening and closing control of filtering processing is readily applicable in the frame as described above with re prediction processing Prediction.
The case where size of macro block is 16 × 16 pixel is described in the above description although note that, the present invention It is readily applicable to the macroblock size of extension described in non-patent literature 3.
Figure 44 is the diagram for illustrating the example of extended macroblock size.For non-patent literature 3, macroblock size extended until 32 × 32 pixels.
Upper layer in Figure 44 in turn shown from left side by be divided into 32 × 32 pixels, 32 × 16 pixels, 16 × The macro block that 32 × 32 pixels of the block (subregion) of 32 pixels and 16 × 16 pixels are constituted.Middle layer in Figure 44 from left side successively Ground shows 16 × 16 pictures of the block (subregion) by being divided into 16 × 16 pixels, 16 × 8 pixels, 8 × 16 pixels and 8 × 8 pixels The block that element is constituted.Lower layer in Figure 44 is in turn shown from left side by being divided into 8 × 8 pixels, 8 × 4 pixels, 4 × 8 pictures The block that 8 × 8 pixels of element and the block (subregion) of 4 × 4 pixels are constituted.
In other words, can be used in Figure 44 in upper layer shown in 32 × 32 pixels, 32 × 16 pixels, 16 × 32 pixels and The block of 16 × 16 pixels handles the macro blocks of 32 × 32 pixels.
In addition, in a manner of identical with H.264/AVC format, can with show in middle level 16 × 16 pixels, 16 × 8 pictures Element, 8 × 16 pixels and 8 × 8 pixels block handle the blocks of 16 × 16 pixels shown on the right of upper layer.
In addition, in a manner of identical with H.264/AVC format, can be used in 8 × 8 pixels shown in lower layer, 8 × 4 pixels, The block of 4 × 8 pixels and 4 × 4 pixels handles the blocks of 8 × 8 pixels shown in the right in middle level.
Using the macroblock size of extension, by utilizing such layer structure, about 16 × 16 block of pixels or smaller, protecting While holding with H.264/AVC format compatible, its superset is defined as by biggish piece.
Filter factor setting, calculating and filtering processing according to the present invention open/close control and are readily applicable to as above The macroblock size of the proposition extended describedly.
It is described so far in the case where H.264/AVC format is used as coded format, but the present invention is not It is limited to this, and it is possible to utilize the another kind for executing prediction (for example, intra prediction or re prediction) using neighborhood pixels Coded format/codec format.
Note that for example, with MPEG, H.26X etc., present invention may apply to via such as satellite broadcasting, have The network media of line TV, internet, cellular phone etc. receives the orthogonal transformation and movement by discrete cosine transform etc. The picture coding device and picture decoding apparatus used when compensating image information (bit stream) of compression.In addition, the present invention can be with Suitable for the image coding dress used when handling the image information on such as storage medium of CD, disk and flash memory It sets and picture decoding apparatus.In addition, present invention may apply to wrap in such picture coding device and picture decoding apparatus The motion predicted compensation device contained.
A series of above-mentioned processing can be executed by hardware, or can be executed by software.Passing through software In the case where executing this series of processing, the program for constituting the software is mounted in a computer.Here, the example packet of computer The computer and general purpose personal computer being placed in specialized hardware are included, thus, it is possible to hold by installing various programs wherein The various functions of row.
Figure 45 is the block diagram of the configuration example of the hardware for the computer that diagram executes a series of above-mentioned processing using program.
For computer, CPU (central processing unit) 801, ROM (read-only memory) 802 and RAM (random access memory Device) it 803 is connected with each other by bus 804.
In addition, input/output interface 805 is connected to bus 804.Input unit 86, output unit 807, storage unit 808, communication unit 809 and driver 810 are connect with input/output interface 805.
Input unit 806 is made of keyboard, mouse, microphone etc..Output unit 807 is made of display, loudspeaker etc.. Storage unit 807 is made of hard disk, nonvolatile memory etc..Communication unit 809 is made of network interface etc..Driver 810 Drive the removable medium 811 of disk, CD, magneto-optic disk, semiconductor memory etc..
For the computer being configured so that, for example, CPU 801 will be stored in the program in storage unit 808 via input/ Output interface 805 and bus 804 are loaded into RAM 803, execute the program, thereby executing a series of above-mentioned processing.
The program that computer (CPU 801) executes can be used as the removable medium of encapsulation medium etc. by being recorded in It is provided in 811.In addition, program can be via the wired or wireless biography of such as local area network, internet or digital satellite broadcasting Defeated media are provided.
For computer, by the way that removable medium 808 is installed on driver 810, can by program via input/it is defeated Outgoing interface 805 is installed in storage unit 808.In addition, the program can be via wired or wireless transmission media in communication unit It is received, and is installed in storage unit 808 at 809.
Furthermore it is possible to which program is pre-installed in ROM 802 or storage unit 808.
The program that note that computer executes can be wherein along sequence described in this specification according to time series The program of processing is executed, or can be and wherein execute processing concurrently or at the necessary timing for such as executing calling Program.
The embodiment of the present invention is not limited to the above embodiments, also, in the case where not departing from essence of the invention Various modifications can be carried out.
For example, above-mentioned 151,401,501 and of picture coding device 51,351,451 and 551 and picture decoding apparatus 601 can be adapted for optional electronic device.Hereinafter, its example will be described.
Figure 46 is main configuration example of the diagram using the television receiver for applying picture decoding apparatus of the invention Block diagram.
Television receiver 1300 shown in Figure 46 includes terrestrial tuner 1313, Video Decoder 1315, vision signal Processing circuit 1318, graphics generation circuit 1319, panel drive circuit 1320 and display panel 1321.
Terrestrial tuner 1313 receives the broadcast wave signal of terrestrial analog broadcast via antenna, carries out to the broadcast wave signal Demodulation obtains vision signal, and these vision signals is supplied to Video Decoder 1315.
Video Decoder 1315 is decoded processing to the vision signal supplied from terrestrial tuner 1313, and will obtain Digital component signal be supplied to video processing circuit 1318.
Video processing circuit 1318 carries out such as noise remove to the video data supplied from Video Decoder 1315 Deng predetermined process, also, the video data of acquisition is supplied to graphics generation circuit 1319.
Graphics generation circuit 1319 generates the video data for the program that be shown on display panel 1321, or due to base In will via the processing of the application of network provisioning caused by image data, also, by the video data or image data of generation It is supplied to panel drive circuit 1320.In addition, graphics generation circuit 1319 also execute such as will by for user generate video counts The video data obtained according to (figure) is supplied to the processing of panel drive circuit 1320, the video data in due course Picture of the display for the selection of project etc., and overlapped on the video data of program.
Panel drive circuit 1320 drives display panel 1321 based on the data supplied from graphics generation circuit 1319, with The video of display program or above-mentioned various pictures on display panel 1321.
Display panel 1321 is made of LCD (liquid crystal display) etc., and aobvious according to the control of panel drive circuit 1320 Show the video etc. of program.
In addition, television receiver 1300 further includes audio A/D (analog/digital) converting unit 1314, Audio Signal Processing Circuit 1322, echo cancellor/audio synthesizing circuitry 1323 and audio-frequency amplifier circuit 1324 and loudspeaker 1325.
Terrestrial tuner 1313 demodulates the broadcast wave signal received, also obtains to not only obtain vision signal Audio signal.The audio signal of acquisition is supplied to audio A/D conversion circuit 1314 by terrestrial tuner 1313.
The audio A/D conversion circuit 1314 carries out A/D conversion process to the audio signal supplied from terrestrial tuner 1313, Also, the digital audio and video signals of acquisition are supplied to audio signal processing circuit 1322.
Audio signal processing circuit 1322 carries out such as noise to the audio data supplied from audio A/D conversion circuit 1314 The predetermined process of removal etc., also, the audio data of acquisition is supplied to echo cancellor/audio synthesizing circuitry 1323.
The audio data supplied from audio signal processing circuit 1322 is supplied to by echo cancellor/audio synthesizing circuitry 1323 Audio-frequency amplifier circuit 1324.
Audio-frequency amplifier circuit 1324 carries out D/A to the audio data supplied from echo cancellor/audio synthesizing circuitry 1323 Conversion process amplifies device processing to adjust scheduled volume, then, will export audio from loudspeaker 1325.
In addition, television receiver 1300 further includes digital tuner 1316 and mpeg decoder 1317.
Digital tuner 1316 receives digital broadcasting (received terrestrial digital broadcasting, BS (broadcasting satellite)/CS (communication via antenna Satellite) digital broadcasting) broadcast wave signal, it is demodulated to obtain MPEG-TS (motion characteristics planning-transport stream), And it is supplied into mpeg decoder 1317.
Mpeg decoder 1317 descrambles the scrambling for giving the MPEG-TS supplied from digital tuner 1316, and Extract the stream of the data comprising being used as the program for resetting target (viewing target).1317 pairs of mpeg decoder constitute the stream extracted Audio packet is decoded, and the audio data of acquisition is supplied to video processing circuit 1322, also, also to composition stream Video packets are decoded, and the video data of acquisition is supplied to video processing circuit 1318.In addition, mpeg decoder EPG (electronic program guides) data extracted from MPEG-TS are supplied to CPU 1332 via unshowned path by 1317.
Television receiver 1300 uses above-mentioned picture decoding apparatus 151,401,501 or 601, as with such The mpeg decoder 1317 that mode is decoded video packets.Therefore, with side identical with picture coding device 151 and 401 Formula, mpeg decoder 1317 switch filter factor according to quantization parameter and prediction mode, also, before intra prediction, to neighbour Nearly pixel executes filtering processing.Alternatively, mpeg decoder 1317 is based in a manner of identical with picture coding device 501 and 601 Open/close whether mark control executes filtering processing to neighborhood pixels before intra prediction.It is thus possible to improve coding effect Rate.
By with the video data supplied from Video Decoder 1315 the case where it is identical in a manner of, in video processing circuit Scheduled processing is carried out to the video data supplied from mpeg decoder 1317 at 1318.Then, in due course, in figure At generation circuit 1319, the video data Jing Guo predetermined process is overlapped on video data of generation etc., via panel driving The video data is supplied to display panel 1321 by circuit 1320, also, its image is shown on display panel 1321.
By with the audio data supplied from audio A/D conversion circuit 1314 the case where it is identical in a manner of, at audio signal Scheduled processing is carried out to the audio data supplied from mpeg decoder 1317 at reason circuit 1322.By the audio of predetermined process Then data are supplied to audio-frequency amplifier circuit 1324 via echo cancellor/audio synthesizing circuitry 1323, and turn by D/A Change processing and amplifier processing.As a result, by being exported with the audio of predetermined volume adjustment from loudspeaker 1325.
In addition, television receiver 1300 also includes microphone 1326 and A/D conversion circuit 1327.
A/D conversion circuit 1327 receives the audio of the user collected by the microphone 1326 being arranged to television receiver 1300 Signal is converted for audio.A/D conversion circuit 1327 carries out A/D conversion process to the audio signal received, also, will obtain The digital audio-frequency data obtained is supplied to echo cancellor/audio synthesizing circuitry 1323.
The audio data of user (user A) for having supplied television receiver 1300 from A/D conversion circuit 1327 the case where In, echo cancellor/audio synthesizing circuitry 1323 executes echo cancellor using the audio data of user A as target. After echo cancellor, echo cancellor/audio synthesizing circuitry 1323 will pass through the audio data and other audio numbers of synthesis user A The audio data obtained according to equal is exported via audio-frequency amplifier circuit 1324 from loudspeaker 1325.
In addition, television receiver 1300 further includes audio codec 1328, internal bus 1329, SDRAM (synchronous dynamic Random access memory) 1330, flash memory 1331, CPU 1332, USB (universal serial bus) I/F 1333 and network I/ F 1334。
A/D conversion circuit 1327 receives the audio of the user collected by the microphone 1326 being arranged to television receiver 1300 Signal is converted for audio.A/D conversion circuit 1327 carries out A/D conversion process to the audio signal received, also, will obtain The digital audio-frequency data obtained is supplied to audio codec 1328.
The audio data supplied from A/D conversion circuit 1327 is converted to the data of predetermined format by audio codec 1328 To be supplied to network I/F 1334 via internal bus 1329 via transmission of network, also, by it.
Network I/F 1334 is via the cable and network connection being mounted in network terminal 1335.For example, network I/F The audio data supplied from audio codec 1328 is transmitted to another device being connected to the network with it by 1334.In addition, for example, Network I/F 1334 is received via network terminal 1335 from the audio data transmitted via network another device connected to it, and And it is supplied to audio codec 1328 via internal bus 1329.
The audio data supplied from network I/F 1334 is converted to the data of predetermined format by audio codec 1328, and And it is supplied into echo cancellor/audio synthesizing circuitry 1323.
Echo cancellor/audio synthesizing circuitry 1323 is being taken as target with the audio data supplied from audio codec 1328 In the case where execute echo cancellor, also, via audio-frequency amplifier circuit 1324 from the output of loudspeaker 1325 by synthesizing the sound The data of the audio of the acquisitions such as frequency evidence and other audio datas.
SDRAM 1330 stores various data needed for CPU 1332 executes processing.
Flash memory 1331 stores the program to be executed by CPU 1332.By such as activating television receiver 1300 Deng predetermined timing at the program that stores in flash memory 1331 read by CPU 1332.It is obtained via digital broadcasting EPG data is also stored in flash memory 1331 via network from data that book server obtains etc..
For example, the MPEG- of the content-data obtained via network from book server comprising the control by CPU 1331 TS is stored in flash memory 1331.For example, passing through the control of CPU 1332, via internal bus 1329, flash storage Its MPEG-TS is supplied to mpeg decoder 1317 by device 1331.
It is handled in a manner of mpeg decoder 1317 is identical by with the MPEG-TS supplied from digital tuner 1316 the case where Its MPEG-TS.In this way, television receiver 1300 receives the content-data being made of video, audio etc. via network, It is decoded using mpeg decoder 1317, so as to show its video, and it is possible to export its audio.
In addition, television receiver 1300 is also comprising the light-receiving list for receiving the infrared signal emitted from remote controler 1351 Member 1337.
Light receiving unit 1337 indicates user's from 1351 receiving infrared-ray of remote controler, also, by what is obtained by demodulation The control routine of the content of operation is exported to CPU 1332.
CPU 1332 is executed according to control routine supplied from light receiving unit 1337 etc. and is deposited in flash memory 1331 The program of storage, to control the whole operation of television receiver 1300.Each unit of CPU 1332 and television receiver 1300 warp It is connected by unshowned path.
USB I/F 1333 is to the television receiver 1300 via the USB cable connection being mounted on USB terminal 1336 Transmission/reception of external device (ED) execution data.Network I/F 1334 is via the cable and network being mounted in network terminal 1335 Connection also executes transmission/reception of the data in addition to audio data to various devices connected to the network.
Television receiver 1300 uses picture decoding apparatus 151,401,501 or 601 as mpeg decoder 1317, thus Code efficiency can be improved.As a result, television receiver 1300 can be with higher speed from receiving via antenna The decoding image with higher precision is obtained in broadcast wave signal or the content-data obtained via network and shows it.
Figure 47 is master of the diagram using the cellular phone for applying picture coding device and picture decoding apparatus of the invention Want the block diagram of configuration example.
Cellular phone 1400 shown in Figure 47 includes the main control unit for being configured as integrally controlling each unit 1450, power circuit unit 1451, operation input control unit 1452, image encoder 1453, Camera IO/F cell 1454, LCD control unit 1455, image decoder 1456, multiplexing/separative unit 1457, record/playback unit 1462, modulation/ Demodulator circuit unit 1458 and audio codec 1459.These units are connected with each other via bus 1460.
In addition, cellular phone 1400 includes operation key 1419, CCD (charge-coupled device) camera 1416, liquid crystal display Device 1418, storage unit 1423, transmission/reception circuit unit 1463, antenna 1414, microphone (MIC) 1421 and loudspeaker 1417。
In end of calling and when powering on key by the operation of user, power circuit unit 1451 by from battery pack to Each unit is powered to activate cellular phone 1400 in operational state.
Based on the control for the main control unit 1450 being made of CPU, ROM, RAM etc., in such as voice call mode, data In the various modes of communication pattern etc., cellular phone 1400 executes various operations, for example, transmission/reception of audio signal, electronics Transmission/reception of mail and image data, image taking, data record etc..
For example, cellular phone 1400 will be by microphone (words by audio codec 1459 in voice call mode Cylinder) 1421 collect audio signals be converted to digital audio-frequency data, spectrum is carried out to it at modulation/demodulation circuit unit 1458 Extension process, also, it is carried out at digital-to-analog conversion process and frequency conversion at transmission/reception circuit unit 1463 Reason.Cellular phone 1400 sends the signal for being used to send obtained by its conversion process to via antenna 1414 unshowned Base station.The signal (audio signal) for being used to send for being sent to base station is supplied to the bee of communication parter via public phone network Cellular telephone.
In addition, for example, in voice call mode, cellular phone 1400 at transmission/reception circuit unit 1463 to Received reception signal amplifies at antenna 1414, further carries out at frequency conversion process and analog/digital conversion to it Reason, carries out spectrum against extension process to it at modulation/demodulation circuit unit 1458, and passes through audio codec for its turn It is changed to analog audio signal.Cellular phone 1400 exports the analog audio signal that the sum that it is converted obtains from loudspeaker 1417.
In addition, for example, cellular phone 1400 is defeated in operation in the case where sending Email in data communication mode Enter the text data for the Email that receiving is inputted by the operation of operation key 1419 at control unit 1452.Cellular phone 1400 handle its text data at main control unit 1450, and show via LCD control unit 1455 as image On liquid crystal display 1418.
In addition, cellular phone 1400 is based on the text data received by operation input control unit 1452, the instruction of user Deng the generation e-mail data at main control unit 1450.Cellular phone 1400 is at modulation/demodulation circuit unit 1458 pair Its e-mail data carries out spectrum extension process, and carries out digital-to-analog to it at transmission/reception circuit unit 1463 Conversion process and frequency conversion process.Cellular phone 1400 is by the signal for being used to send obtained by its conversion process via day Line 1414 is sent to unshowned base station.The signal (Email) for being used to send of base station will be sent to via network, mail Server etc. is supplied to scheduled destination.
In addition, for example, when in data communication mode receive Email in the case where, cellular phone 1400 with send/ Receiving circuit unit 1463 receives the signal that sends via antenna 1414 from base station, amplifies to the signal, also, to its into One stepping line frequency conversion process and analog/digital conversion processing.Cellular phone 1400 is at modulation/demodulation circuit unit 1458 Signal is received to it carries out spectrum against extension process, to restore original electronic mail data.Cellular phone 1400 is controlled via LCD Unit 1455 shows the e-mail data of recovery on liquid crystal display 1418.
It note that cellular phone 1400 can record in storage unit 1423 via record/playback unit 1462 (to deposit Storage) e-mail data that receives.
The storage unit 1423 is optional rewritable recording medium.Storage unit 1423 can be such as RAM, interior The semiconductor memory for setting flash memory etc. can be hard disk, or can be such as disk, magneto-optic disk, CD, USB and deposit The removable medium of reservoir, memory card etc..Much less, storage unit 1423 can be storage unit in addition to these.
In addition, for example, when in data communication mode send image data in the case where, cellular phone 1400 by Imaging is at CCD camera 1416 to generate image data.CCD camera 1416 includes the optics as lens, aperture etc. The intensity-conversion of the light received is electric signal by device and the CCD that subject is imaged for being used as photoelectric conversion device, And generate the image data of the image of subject.Via Camera IO/F cell 1451, MPEG2, MPEG4 etc. are used Scheduled coded format carries out compressed encoding to its image data at image encoder 1453, and therefore, its image data is turned It is changed to the image data of coding.
Cellular phone 1400 is using above-mentioned picture coding device 61,351,451 and 551 as executing this processing Image encoder 1453.Therefore, in a manner of identical with picture coding device 51 and 351, image encoder 1453 is according to amount Change parameter and filter factor is arranged in prediction mode, also, before intra prediction, filtering processing is executed to neighborhood pixels.Alternatively, In a manner of identical with picture coding device 451 and 551, whether the control of image encoder 1453 is before intra prediction to neighbouring Pixel executes filtering processing.It is thus possible to improve code efficiency.
It note that at this point, simultaneously, while being shot with CCD camera 1416, cellular phone 1400 is in audio coding decoding Analog-digital Converter is carried out to the audio collected at microphone (microphone) 1421 at device 1459, and further it is compiled Code.
Cellular phone 1400 is using scheduled method to from image encoder 1453 at multiplexing/separative unit 1457 The coded image data of supply and from audio codec 1459 supply digital audio-frequency data multiplexed.Cellular phone 1400 carry out spectrum extension process to the multiplexed data obtained as its result at modulation/demodulation circuit unit 1458, And digital-to-analog conversion process and frequency conversion process are carried out to it at transmission/reception circuit unit 1463.Cellular phone 1400 send unshowned base station via antenna 1414 for the signal for being used to send obtained by its conversion process.It will send The signal (picture signal) for transmission to base station is supplied to communication parter via network etc..
It note that in the case where not sending image data, cellular phone 1400 can also be via LCD control unit 1455 Rather than image encoder 1453 is shown in the image data generated at CCD camera 1416 on liquid crystal display 1418.
In addition, for example, working as the data for the motion pictures files that reception and simple website etc. link in data communication mode In the case where, cellular phone 1400 receives the letter sent from base station via antenna 1414 at transmission/reception circuit unit 1463 Number, which is amplified, also, its further progress frequency conversion process and analog/digital conversion are handled.
Cellular phone 1400 carries out spectrum against at extension to the signal received at the modulation/demodulation circuit unit 1458 Reason, to restore original multiplexed data.Cellular phone 1400 is multiplexed at multiplexing/separative unit 1457 Data separating is the image data and audio data of coding.
Cellular phone 1400 is being schemed using codec format corresponding with the scheduled coded format of MPEG2, MPEG4 etc. As being decoded at decoder 1456 to the image data of coding, to generate the motion image data of playback, also, honeycomb electricity Words 1400 show the motion image data via LCD control unit 1455 on liquid crystal display 1418.Thus, for example, in liquid The motion image data for including in the motion pictures files with simple web site url is shown on crystal display 1418.
Cellular phone 1400 is using above-mentioned picture decoding apparatus 151,401,501 or 601 as executing this place The image decoder 1456 of reason.Therefore, in a manner of identical with picture decoding apparatus 151 and 401,1456 basis of image decoder Quantization parameter and prediction mode switch filter factor, also, before intra prediction, execute filtering processing to neighborhood pixels.Or Person, in a manner of identical with picture decoding apparatus 501 and 601, image decoder 1456 is based on opening/closing mark control in frame Whether filtering processing is executed to neighborhood pixels before interior prediction.It is thus possible to improve code efficiency.
At this point, simultaneously, digital audio-frequency data is converted to analog audio at audio codec 1459 by cellular phone 1400 Frequency signal, and it is exported from loudspeaker 1417.Thus, for example, playing in the motion pictures files with simple web site url The audio data for including.
Note that it is identical by with Email the case where in a manner of, cellular phone 1400 can be via record/playback unit 1462 by the data record (storage) received linked with simple website etc. in storage unit 1423.
In addition, two dimension of the cellular phone 1400 at main control unit 1450 to the imaging obtained by CCD camera 1416 Code is analyzed, it is hereby achieved that the information recorded in 2 d code.
In addition, infrared ray can be used at infrared communication unit 1481 and communication with external apparatus in cellular phone 1400.
For example, cellular phone 1400 is using picture coding device 51,351,451 or 551 as image encoder 1453, from And the volume of the coded data generated and encoding to the image data generated at CCD camera 1416 can be improved Code efficiency.As a result, cellular phone 1400 can provide the coded data (image with excellent code efficiency to another device Data).
In addition, cellular phone 1400 is using picture decoding apparatus 151,401,501 or 601 as image decoder 1456, So as to generate with high-precision forecast image.For example, as a result, cellular phone 1400 can from simple website The motion pictures files of link obtain the decoding image with higher precision, and show it.
It note that and such description has been carried out so far: where cellular phone 1400 utilizes CCD camera 1416, still, cellular phone 1400 can use the imaging sensor (CMOS using CMOS (complementary metal oxide semiconductor) Imaging sensor) replace the CCD camera 1416.In this case, with using CCD camera 1416 the case where, is identical Mode, cellular phone 1400 can also be imaged to subject and be generated the image data of the image of subject.
It is described in addition, being directed to cellular phone 1400 so far, still, with the feelings with cellular phone 1400 The identical mode of condition, picture coding device 51,351,451 and 55 and picture decoding apparatus 151,401,501 and 601 can be with Suitable for any kind of device, as long as it is that have imaging identical with the imaging function of cellular phone 1400 and image function The device of function and image function, for example, PDA (personal digital assistant), smart phone, UMPC (super mobile personal computer), Network book (net book), notebook personal computer etc..
Figure 48 is that diagram uses the hdd recorder for applying picture coding device and picture decoding apparatus of the invention The block diagram of main configuration example.
Hdd recorder shown in Figure 48 (HDD logger) 1500 is devices which that the device is in built-in hard disk The broadcast that middle storage is received by tuner and includes from the broadcast wave signal (TV signal) of the transmissions such as satellite or ground-plane antenna The audio data and video data of program, also, the data of storage are provided a user according to the instruction of user at a certain timing.
For example, hdd recorder 1500 can be from broadcast wave signal extraction audio data and video data, when appropriate Time is decoded these audio datas and video data, and stores it in built-in hard disk.In addition, for example, hard disk Logger 1500 can also obtain audio data and video data from another device via network, in due course to these sounds Frequency evidence and video data are decoded, and are stored it in built-in hard disk.
In addition, for example, hdd recorder 1500 carries out the audio data and video data recorded in built-in hard disk Decoding, is supplied into monitor 1460, also, its image is shown on the screen of monitor 1560.In addition, hdd recorder 1500 can export its audio from the loudspeaker of monitor 1560.
For example, hdd recorder 1500 is to the audio data and video from the broadcast wave signal extraction obtained via tuner Data or the audio data and video data obtained via network from another device are decoded, and are supplied into monitor 1560, also, its image is shown on the screen of monitor 1560.In addition, hdd recorder 1500 can be from monitor 1560 Loudspeaker export its audio.
Much less, the operation in addition to these operations can be executed.
As shown in figure 48, hdd recorder 1500 includes receiving unit 1521, demodulating unit 1522, demultiplexer 1523, sound Frequency decoder 1524, Video Decoder 1525 and logger control unit 1526.Hdd recorder 1500 further includes that EPG data is deposited Reservoir 1527, program storage 1528, working storage 1529, display converter 1530, OSD (show) control unit on screen 1531, display control unit 1523, record/playback unit 1533, D/A converter 1534 and communication unit 1535.
In addition, display converter 1530 includes video encoder 1541.Record/playback unit 1533 includes encoder 1551 With decoder 1552.
Receiving unit 1521 receives infrared signal from remote controler (not shown), is converted into electric signal, and export and give Logger control unit 1526.Logger control unit 1526 is made of such as microprocessor etc., also, according to being stored in program Program in memory 1528 executes various processing.At this point, logger control unit 1526 uses various memories as needed 1529。
Communication unit 1535 connected to the network executes image procossing via network and another device.For example, communication unit 1535 are controlled by logger control unit 1526, to communicate with tuner (not shown) also, mainly export channel to tuner Selection control signal.
Demodulating unit 1522 demodulates the signal supplied from tuner, and exports to demultiplexer 1523.Demultiplexer 1523 by the data separating supplied from demodulating unit 1522 be audio data, video data and EPG data, and respectively output give Audio decoder 1524, Video Decoder 1525 and logger control unit 1526.
Audio decoder 1524 is for example decoded using audio data of the mpeg format to input, and is exported to note Record/playback unit 1533.Video Decoder 1525 is for example decoded using video data of the mpeg format to input, and defeated Out to display converter 1530.Logger control unit 1526 by the EPG data of input be supplied to EPG data storage 1527 with Just it stores.
Display converter 1530 will be from Video Decoder 1525 or logger control unit using video encoder 1541 The video data encoding of 1526 supplies is the video data for for example meeting NTSC (National Television Standards Committee) format, and defeated Out to record/playback unit 1533.In addition, display converter 1530 will be from Video Decoder 1525 or logger control unit The size conversion of the picture of the video data of 1526 supplies is the size of the size corresponding to monitor 1560.Display converter 1530 are further converted into the video data that converted screen size using video encoder 1541 view for meeting NTSC format Frequency evidence, and analog signal is converted to, and export to display control unit 1532.
Under the control of logger control unit 1526, display control unit 1523 will be controlled from OSD (showing on screen) The osd signal that unit 1531 exports is overlapped from the vision signal that display converter 1530 inputs, and is exported to monitor 1560 display is to show.
In addition, the audio data exported from audio decoder 1524 is converted into analog signal using D/A converter 1534, And it is supplied into monitor 1560.Monitor 1560 exports the audio signal from built-in loudspeaker.
Record/playback unit 1533 includes hard disk, there is the recording medium of video data, video data etc. as record.
For example, record/playback unit 1533 is supplied by encoder 1551 from audio decoder 1524 using mpeg format The audio data answered is encoded.In addition, record/playback unit 1533 using mpeg format by encoder 1551 to from display The video data that the video encoder 1541 of converter 1530 is supplied is encoded.Record/playback unit 1533 is multiple using multichannel The coded data of its audio data and the coded data of its video data are synthesized with device.Record/playback unit 1533 passes through channel The data of coding amplification synthesis, and its data is written in a hard disk via record head.
Record/playback unit 1533 plays the data recorded in a hard disk via reproducing head, amplifies the data, also, make Audio data and video data are separated into demultiplexer.Record/playback unit 1533 passes through decoder using mpeg format 1552 pairs of audio datas and video data are decoded.Record/playback unit 1533 carries out digital mould to decoded audio data Quasi- conversion, and export to the loudspeaker of monitor 1560.
In addition, record/playback unit 1533 carries out digital-to-analogue conversion to decoded video data, and export to monitoring The display of device 1560.
Logger control unit 1526 via the received infrared signal from remote controler of receiving unit 1521 based on by referring to The instruction of the user shown reads newest EPG data from EPG data storage 1527, and is supplied to OSD control unit 1531. OSD control unit 1531 generates the image data for corresponding to the EPG data of input, and exports to display control unit 1532. Display control unit 1532 by the video data inputted from OSD control unit 1531 export the display to monitor 1560 so as to Display.Therefore, EPG (electronic program guides) is shown on the display of monitor 1560.
In addition, hdd recorder 1500 can be obtained via the network of internet etc. from the various of another device provisioning Data, for example, video data, audio data, EPG data etc..
Communication unit 1535 is controlled by logger control unit 1526, with obtain send via network from another device it is all Such as coded data of video data, audio data, EPG data, and it is supplied into logger control unit 1526.Example Such as, the coded data of the video data of acquisition and audio data is supplied to record/playback unit by logger control unit 1526 1533, and store it in hard disk.At this point, logger control unit 1526 and record/playback unit 153 can be according to need Execute the processing recompiled etc..
In addition, the coded data of 1526 pairs of the logger control unit video datas obtained and audio data is decoded, And the video data of acquisition is supplied to display converter 1530.With with the video data phase supplied from Video Decoder 1525 Same mode, display converter 1530 handles the video data supplied from logger control unit 1526, via display Control unit 1532 is supplied to monitor 1560, to show its image.
Alternatively, such arrangement can be carried out: where shown according to the image, logger control unit 1526 will decode Audio data be supplied to monitor 1560 via D/A converter 1534, and export its audio from loudspeaker.
In addition, the coded data of the EPG data of 1526 pairs of logger control unit acquisitions is decoded, also, will decoding EPG data be supplied to EPG data storage 1527.
The hdd recorder 1500 being configured so that is decoded using picture decoding apparatus 151,401,501 or 601 as video Device 1525, decoder 1552 and the decoder being contained in logger control unit 1526.Therefore, with picture decoding apparatus 151 and 401 identical modes, Video Decoder 1525, decoder 1552 and the solution being contained in logger control unit 1526 Code device switches filter factor according to quantization parameter and prediction mode, also, before intra prediction, executes filtering to neighborhood pixels Processing.Alternatively, in a manner of identical with picture decoding apparatus 501 and 601, Video Decoder 1525, decoder 1552 and receiving Whether the decoder in logger control unit 1526 is based on opening/closing mark control before intra prediction to neighbouring picture Element executes filtering processing.It is thus possible to improve code efficiency.
Therefore, hdd recorder 1500 can produce with high-precision forecast image.As a result, hdd recorder 1500 can for example from via the received video data of tuner coded data, from the hard disk of record/playback unit 1533 read The coded data of video data out or the coded data of the video data obtained via network, which obtain, has higher precision Image is decoded, and is shown on monitor 1560.
In addition, hdd recorder 1500 is using picture coding device 51,351,451 or 551 as encoder 1551.Cause This, in a manner of identical with picture coding device 51 and 351, encoder 1551 is according to quantization parameter and prediction mode setting filtering Coefficient, also, before intra prediction, filtering processing is executed to neighborhood pixels.Alternatively, with picture coding device 451 and 551 Whether identical mode, the control of encoder 1551 execute filtering processing to neighborhood pixels before intra prediction.Therefore, Ke Yiti High coding efficiency.
Thus, for example, the code efficiency that record coded data in a hard disk can be improved in hdd recorder 1500.Make For as a result, hdd recorder 1500 can use the storage region of hard disk in a more effective manner.
It note that so far to the hdd recorder for recording video data and audio data in a hard disk 1500 are described, and still, much less, can use any kind of recording medium.It is applied such as example, even using The logger of the recording medium in addition to hard disk of flash memory, CD, video tape etc., with above-mentioned hdd recorder 1500 the case where identical mode, 151,401,501 and of picture coding device 51,351,451 and 551 and picture decoding apparatus 601 are readily applicable to this.
Figure 49 is that diagram uses the main of the camera for applying picture coding device and picture decoding apparatus of the invention The block diagram of configuration example.
Subject is imaged in camera 1600 shown in Figure 49, and the image of subject is shown on LCD 1616, And as Imagery Data Recording in recording medium 1633.
Light (that is, video of subject) is input to CCD/CMOS 1612 by block of lense 1611.CCD/CMOS 1612 is benefit It is electric signal by the intensity-conversion of the light received with the imaging sensor of CCD or CMOS, and is supplied at camera signal Manage unit 1613.
The electric signal supplied from CCD/CMOS 1612 is converted to the face of Y, Cr and Cb by camera signal processing unit 1613 Colour difference signal, and it is supplied to image signal processing unit 1614.Under the control of controller 1621, image signal processing unit 1614 pairs of picture signals supplied from camera signal processing unit 1613 carry out scheduled image procossing, alternatively, using for example Mpeg format encodes its picture signal by encoder 1641.Image signal processing unit 1614 will be by believing image Number coded data for being encoded and being generated is supplied to decoder 1615.In addition, the acquisition of image signal processing unit 1614 is being shielded The data for display generated at display (OSD) 1620 on curtain, and it is supplied into decoder 1615.
For above-mentioned processing, camera signal processing unit 1613 utilizes as suitably desired to be connected via bus 1617 The DRAM (dynamic random access memory) 1617 connect is kept by image data, from coded data of its coded image data etc. In its DRAM 1618.
Decoder 1615 is decoded the coded data supplied from image signal processing unit 1614, also, will obtain Image data (decoding image data) be supplied to LCD 1616.In addition, decoder 1615 will be from image signal processing unit The data for display of 1614 supplies are supplied to LCD 1616.LCD 1616 is synthesized in due course to be supplied from decoder 1615 The image for the data for display answered and the image of decoding image data, and show its composograph.
Under the control of controller 1621, display 1620 will be by such as dish of symbol, character or figure constitution on screen The data for display of single-image or icon etc. are output to image signal processing unit 1614 via bus 1617.
Signal based on the content that instruction is ordered by user using operating unit 1622, control unit 1621 execute various places Reason, also, also control via bus 1617 image signal processing unit 1614, DRAM 1618, external interface 1619, on screen Display 1620, media drive 1623 etc..Procedure, data needed for executing various processing for controller 1621 etc. is stored in In flash ROM 1624.
For example, controller 1621 can replace image signal processing unit 1614 and decoder 1615 to being stored in DRAM Image data in 1618 is encoded, or is decoded to the coded data being stored in DRAM 1618.At this point, control Device 1621 can be used format identical with the coding and decoding format of image signal processing unit 1614 and decoder 1615 and hold Row coding and decoding processing, alternatively, the lattice that image signal processing unit 1614 and decoder 1615 cannot all be handled can be used Formula is handled to execute coding and decoding.
In addition, for example, in the case where image prints since indicating operating unit 1622, controller 1621 is from DRAM 1618 read image datas, and be supplied into via bus 1617 printer 1634 that is connect with external interface 1619 so as to Printing.
In addition, for example, in the case where indicating image recording from operating unit 1622, controller 1621 is from DRAM 1618 read coded data, and the recording medium being mounted on media drive 1623 is supplied into via bus 1617 1633 to store.
Recording medium 1633 is optional read-write removable medium, for example, disk, magneto-optic disk, CD, semiconductor are deposited Reservoir etc..Much less, about the type of removable medium, recording medium 1633 be also it is optional, as such, it can be that with dress It sets, perhaps can be disk or can be memory card.Much less, recording medium 1633 can be non-contact IC card etc..
Alternatively, media drive 1623 and recording medium 1633, which can be configured as, is integrated in non-portabillity recording medium In, the non-portabillity recording medium is, for example, built-in hard disk drive, SSD (solid state drive) etc..
External interface 1619 is made of such as USB input/output terminal etc., and execute image printing in the case where with Printer 1634 connects.In addition, driver 1631 is connect with external interface 1619 as needed, such as disk, CD or magneto-optic The removable medium 1632 of disk is mounted in due course, also, the computer program being read out from is as needed It is installed in flash ROM 1624.
In addition, external interface 1619 includes the network interface to connect with the predetermined network of LAN, internet etc..Example Such as, according to the instruction from operating unit 122, controller 1621 can read coded data from DRAM 1618, and by its from External interface 1619 is supplied to another device connected via a network.In addition, controller 1621 can be via external interface 1619 It obtains via network from the coded data or image data of another device provisioning, and holds it in DRAM 1618, or Person is supplied into image signal processing unit 1614.
The camera 1600 being configured so that is using picture decoding apparatus 151,401,501 or 601 as decoder 1615.
Therefore, in a manner of identical with picture decoding apparatus 151 and 401, decoder 1615 is according to quantization parameter and prediction Pattern switching filter factor, also, before intra prediction, filtering processing is executed to neighborhood pixels.Alternatively, with image decoding Whether the identical mode of device 501 and 601, decoder 1615 are based on opening/closing mark control before intra prediction to neighbour Nearly pixel executes filtering processing.It is thus possible to improve code efficiency.
Therefore, camera 1600 can produce with high-precision forecast image.As a result, for example, camera 1600 can be from the image data generated at CCD/CMOS 1612, the video read from DRAM 1618 or recording medium 1633 The coded data of data or the coded data of the video data obtained via network obtain the decoding figure with higher precision Picture, and shown on LCD 1616.
In addition, camera 1600 is using picture coding device 51,351,451 or 551 as encoder 1641.
Therefore, in a manner of identical with picture coding device 51 and 351, encoder 1641 is according to quantization parameter and prediction mould Filter factor is arranged in formula, also, before intra prediction, executes filtering processing to neighborhood pixels.Alternatively, being filled with being encoded with image 451 and 551 identical modes are set, whether the control of encoder 1641 executes filtering processing to neighborhood pixels before intra prediction. It is thus possible to improve code efficiency.
Thus, for example, the code efficiency that record coded data in a hard disk can be improved in camera 1600.As it As a result, camera 1600 can be in a more effective manner using DRAM 1618 or the storage region of recording medium 1633.
It note that the coding/decoding method of picture decoding apparatus 151,401,501 and 601 can be adapted for the execution of controller 1621 Decoding process.Similarly, the coding method of picture coding device 51,351,451 and 551 can be adapted for controller 1621 and hold Capable coded treatment.
In addition, the image data that camera 1600 is imaged can be moving image, or it can be static image.
Much less, picture coding device 51,351,451 and 551 and picture decoding apparatus 151,401,501 and 601 can With the device or system being suitable in addition to above-mentioned apparatus.
Reference signs list
51 picture coding devices
66 lossless coding units
74 intraprediction units
75 neighborhood pixels interpolation filtering switch units
81 neighborhood pixels setting units
82 forecast images generate unit
83 optimum prediction mode determination units
91 prediction mode buffers
92 quantization parameter buffers
93 low-pass filtering setting units
94 frame coefficient memories
111 neighborhood pixels setting units
112 forecast images generate unit
113 optimum prediction mode determination units
121 prediction mode buffers
122 optimum filtering computing units
151 picture decoding apparatus
162 lossless decoding units
171 intraprediction units
172 neighborhood pixels interpolation filtering switch units
181 forecast images generate unit
182 neighborhood pixels setting units
191 prediction mode buffers
192 quantization parameter buffers
193 low-pass filtering setting units
194 filter factor memories
202 low-pass filtering setting units
251 learning devices
261 adjacent to interpolation filtering computing unit
271 filter factor storage units
351 picture coding devices
361 re prediction units
362 neighborhood pixels interpolation filtering switch units
401 picture decoding apparatus
411 re prediction units
412 neighborhood pixels interpolation filtering switch units
451 picture coding devices
461 neighborhood pixels interpolation filtering control units
501 picture decoding apparatus
511 neighborhood pixels interpolation filtering control units
551 picture coding devices
561 neighborhood pixels interpolation filtering control units
601 picture decoding apparatus
611 neighborhood pixels interpolation filtering control units

Claims (10)

1. a kind of image processing apparatus, comprising:
Filter part is used using carrying out image obtained by decoding process as object to bit stream according to as decoding process The block size when pixel of the object block of object carries out intra prediction and the filter factor set, to the picture for being directed to the object block The neighborhood pixels that element carries out the reference of intra prediction when institute are filtered;
Control unit is carried out prediction mode when intra prediction to the pixel for the object block, is controlled using control information The filtering processing is made to select to be filtered the neighborhood pixels or not be filtered to the neighborhood pixels Processing, the control information control is to be filtered or be not filtered to the neighborhood pixels to the neighborhood pixels Processing;And
Intra prediction component carries out intra prediction using the neighborhood pixels come the pixel to the object block, generates prognostic chart Picture.
2. image processing apparatus according to claim 1, wherein
The filter part utilizes the filter for using 3 taps in the case where having selected to be filtered the neighborhood pixels The filtering operation of wave system number is filtered the neighborhood pixels.
3. image processing apparatus according to claim 1, further includes:
Decoding device is decoded place to the bit stream using by the intra prediction component forecast image generated Reason.
4. image processing apparatus according to claim 3, wherein
The bit stream be generated by being encoded according to the unit with layer structure to image, and
The decoding device is decoded processing to the bit stream according to the unit with layer structure.
5. image processing apparatus according to claim 1, wherein
Block size is set as 4 × 4 pixels, 8 × 8 pixels, 16 × 16 pixels, 32 × 32 pixels to carry out by the intra prediction component Intra prediction.
6. a kind of image processing method, the following steps carried out including image processing apparatus:
To carry out image obtained by decoding process as object to bit stream, use according to as the object of the object of decoding process The block size when pixel of block carries out intra prediction and the filter factor set, carry out in frame the pixel for the object block The neighborhood pixels of institute's reference are filtered when prediction;
Prediction mode when intra prediction is carried out to the pixel for the object block, controls the filtering using control information Processing is to select to be filtered the neighborhood pixels or not be filtered to the neighborhood pixels, the control Information control processed is to be filtered or be not filtered to the neighborhood pixels to the neighborhood pixels;And
Intra prediction is carried out come the pixel to the object block using the neighborhood pixels, generates forecast image.
7. image processing method according to claim 6, wherein
In the case where having selected to be filtered the neighborhood pixels, the filtering for the filter factor for using 3 taps is utilized Operation is filtered the neighborhood pixels.
8. image processing method according to claim 6, further includes:
Processing is decoded to the bit stream using forecast image generated.
9. image processing method according to claim 8, wherein
The bit stream be generated by being encoded according to the unit with layer structure to image, and
Processing is decoded to the bit stream according to the unit with layer structure.
10. image processing method according to claim 6, wherein
Block size is set as 4 × 4 pixels, 8 × 8 pixels, 16 × 16 pixels, 32 × 32 pixels to carry out intra prediction.
CN201610811282.2A 2009-07-01 2010-06-23 Image processing apparatus and method Active CN106331710B (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP2009156563 2009-07-01
JP2009-156563 2009-07-01
JP2009-244753 2009-10-23
JP2009244753A JP5597968B2 (en) 2009-07-01 2009-10-23 Image processing apparatus and method, program, and recording medium
CN201080028028.6A CN102474617B (en) 2009-07-01 2010-06-23 Image processing apparatus and method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN201080028028.6A Division CN102474617B (en) 2009-07-01 2010-06-23 Image processing apparatus and method

Publications (2)

Publication Number Publication Date
CN106331710A CN106331710A (en) 2017-01-11
CN106331710B true CN106331710B (en) 2019-07-16

Family

ID=

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1596546A (en) * 2001-11-30 2005-03-16 罗伯特-博希股份公司 Method for conducting a directed prediction of an image block
CN1851806A (en) * 2006-05-30 2006-10-25 北京中星微电子有限公司 Adaptive microphone array system and its voice signal processing method
CN101115201A (en) * 2007-08-30 2008-01-30 上海交通大学 Video decoding method and device
WO2008085109A1 (en) * 2007-01-09 2008-07-17 Telefonaktiebolaget Lm Ericsson (Publ) Adaptive filter representation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1596546A (en) * 2001-11-30 2005-03-16 罗伯特-博希股份公司 Method for conducting a directed prediction of an image block
CN1851806A (en) * 2006-05-30 2006-10-25 北京中星微电子有限公司 Adaptive microphone array system and its voice signal processing method
WO2008085109A1 (en) * 2007-01-09 2008-07-17 Telefonaktiebolaget Lm Ericsson (Publ) Adaptive filter representation
CN101115201A (en) * 2007-08-30 2008-01-30 上海交通大学 Video decoding method and device

Similar Documents

Publication Publication Date Title
CN106254876B (en) Image processing apparatus and method
CN106331710B (en) Image processing apparatus and method
JP6048774B2 (en) Image processing apparatus and method
AU2017201515B2 (en) Image processing device and method
AU2015203503B2 (en) Image processing device and method
AU2010267185B8 (en) Image processing device and method

Legal Events

Date Code Title Description
PB01 Publication
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20190121

Address after: Dublin, Ireland

Applicant after: Qualcomm Inc.

Address before: Tokyo, Japan

Applicant before: Sony Corp.

GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20221213

Address after: Tokyo

Patentee after: Sony Group

Address before: Irish Dublin

Patentee before: Qualcomm Inc.