CN101534432A - Method for controlling code rate based on human eye sensing model - Google Patents

Method for controlling code rate based on human eye sensing model Download PDF

Info

Publication number
CN101534432A
CN101534432A CN 200910049042 CN200910049042A CN101534432A CN 101534432 A CN101534432 A CN 101534432A CN 200910049042 CN200910049042 CN 200910049042 CN 200910049042 A CN200910049042 A CN 200910049042A CN 101534432 A CN101534432 A CN 101534432A
Authority
CN
China
Prior art keywords
pixel
model
rate control
sigma
macro block
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN 200910049042
Other languages
Chinese (zh)
Inventor
郭凤
潘琤雯
滕国伟
郁志明
石旭利
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SVA Group Co Ltd
Central Academy of SVA Group Co Ltd
Original Assignee
Central Academy of SVA Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central Academy of SVA Group Co Ltd filed Critical Central Academy of SVA Group Co Ltd
Priority to CN 200910049042 priority Critical patent/CN101534432A/en
Publication of CN101534432A publication Critical patent/CN101534432A/en
Pending legal-status Critical Current

Links

Images

Abstract

The invention provides a method for controlling code rate based on a human eye sensing model. The method comprises the following steps: through extracting static characters and dynamic characters of a video image, establishing a static character model and a dynamic character model respectively; further establishing a static and dynamic comprehensive character model; controlling code rate of frame level by the dynamic character model; and controlling code rate of macro block level by the static and dynamic comprehensive character model. The method for controlling the code rate not only can ensure that the decoded video has better visual effect, but also can realize distribution of target bit more effectively under a condition of limited number of bit.

Description

Bit rate control method based on human eye sensing model
Technical field
The present invention relates to the digital video coding technology, particularly relate to a kind of bit rate control method based on human eye sensing model.
Background technology
Rate Control is one of key technology of video coding, is that encoder is based on the process that will send to the video bits flow rate on the channel to the estimation decision of network availability bandwidth.The key of control is to find a balance between code check size and video compression quality.Therefore the quality of rate control algorithm has directly influenced the performance and the efficient of encoder.H.264/AVC and AVS in existing several compression standards, as, its bit rate control method only carries out Data Rate Distribution according to the complexity of video content, that is, distribute the more bits number for the part of texture complexity.But the final recipient of video is human, and the people can be decided to the apperceive characteristic of video with to the understanding of video content by human brain for the concern situation of video.Under many scientific workers' effort, we have realized that human eye can drop into the more concern degree for the edge contour of moving object in the video and object.Therefore, if think to be different frame, different macroblock allocation target bit more effectively, just must consider visual characteristics of human eyes in the Rate Control and go.
Summary of the invention
The object of the present invention is to provide a kind of bit rate control method based on human eye sensing model, described method is according to the behavioral characteristics and the static nature of video image, set up the static attention model and the dynamic attention model of human vision, and extract the zone that the human eye in the video image is paid close attention to by these two kinds of models, to distinguish the non-zone of paying close attention to, realization is carried out the classification quantitative coding at the different zone of significance level in the image, and then realizes the bit rate control method based on the human eye perception.
The object of the present invention is achieved like this: a kind of bit rate control method based on human eye sensing model, described bit rate control method carries out frame level and macro-block level Rate Control respectively, to determine the distribution of bit number, wherein, the size of macro block is h*h, h is a natural number, and described bit rate control method is realized by following steps:
The static nature of step 1, extraction video image, and set up the static nature model;
The behavioral characteristics of step 2, extraction video image, and set up the behavioral characteristics model;
Step 3, get the common factor of static nature model and behavioral characteristics model, obtain the moving image portion and the profile thereof of this video, i.e. sound attitude comprehensive characteristics model;
Step 4, in the frame level bit-rate control algolithm, utilize the behavioral characteristics model that extracts, calculate the ratio of the shared whole macroblock number of motion object piece in each frame, and with former linear prediction model MAD (the n)=k of mean absolute difference MAD of corresponding macro block pixels point in each macro block in this ratio correction current encoded frame and the corresponding reference frame of coding 1* MAD (n-1)+k 2, when the preassignment bit number,, reflect the motion complexity of present frame according to what of motion object piece, thus decision allocation bit number how much realize Rate Control, wherein k 1Be the linear scale factor of this model, k 2Be the constant coefficient of this model, n is the picture frame sequence number;
Step 5, in the macro-block level Rate Control, at first according to sound attitude comprehensive characteristics model, calculate the ratio of total pixel in the shared macro block of pixel in the sound attitude comprehensive characteristics model, and by an adjusting of this ratio acquisition parameter, in Rate Control, use and regulate the original target bit mean allocation of parameter modification formula Bit (n, I, J)=f Rb(n, I, J)/N Ub, realize the best Rate Control effect under the human eye perception, wherein, n is the picture frame sequence number, ((n, I J) are target bit, f to Bit for I, the J) coordinate of expression macro block Rb(n, I J) are total bit number to be allocated, N UbIt is number of macroblocks to be allocated.
The foundation of the extraction of the static nature in the described step 1 and static nature model further realizes by following steps:
Step 1.1, convert each pixel of current video to rgb format from yuv format; Extract red R, green G, blue B and yellow Y four primary colours, and the chromaticity figure that obtains four primary colours be respectively R (r, g, b), G (r, g, b), B (r, g, b) and Y (r, g, b); Extract red/green, indigo plant/yellow two groups of antagonistic pairs, and obtain red/green characteristic pattern RG (i, j)=| (R (and i, j)-G (i, j)) | and indigo plant/yellow characteristic pattern BY (i, j)=| (B (and i, j)-Y (i, j)) |, wherein (i, j) coordinate of remarked pixel point;
Step 1.2, calculate each pixel and the mean value of red/green, the indigo plant/yellow characteristic value of eight points on every side thereof Diff RG ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | RG ( i , j ) - RG ( n , m ) | With Diff BY ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | BY ( i , j ) - BY ( n , m ) | ;
Step 1.3, calculate the static nature figure of each pixel StaticMap ( i , j ) = Diff BY ( i , j ) 2 + Diff RG ( i , j ) 2 , The maximum of static nature figure is designated as StaticMap Max, minimum value is designated as StaticMap Min, establish and judge that the static nature figure threshold value whether pixel meets the static nature model is T, then T=(StaticMap Max+ StaticMap Min)/2, (i, during j) greater than T, this pixel promptly is included in the static nature model, otherwise is not included in the static nature model as the static nature figure of pixel StaticMap.
The foundation of the extraction of the behavioral characteristics in the described step 2 and behavioral characteristics model further realizes by following steps:
Step 2.1, establish n frame (I, J) motion vector of macro block be designated as PV (I, J)=(x N, I, J, y N, I, J), the motion vector of this macro block is defaulted as the motion vector of this each pixel of macro block, the direction of motion of this vector is expressed as θ N, i, i=arctan (y N, i, j/ x N, i, j);
Step 2.2, calculate the current pixel point and the probability histogram distribution function of the motion vector value of eight points on every side thereof P s ( n ) = SH i , j w ( n ) Σ l = 1 m SH i , j w ( l ) ; Wherein SH () is by the current pixel point and the direction value θ of the motion vector of eight points on every side thereof N, i, iThe histogram of being formed, m are histogram space size, and w represents the search window size of N*N; Calculate the spatial coherence entropy of the motion vector value of each pixel according to the probability distribution situation of gained Cs ( i , j ) = - Σ n = 1 m P s ( n ) Log ( P s ( n ) ) ; Wherein, the spatial information entropy of Cs () expression motion vector, P sIt is the corresponding probability-distribution function of histogram SH ();
The probability histogram distribution function of the motion vector value of the pixel on the same position of step 2.3, the motion vector value that calculates current pixel point and front and back three frames thereof P t ( n ) = TH i , j L ( n ) Σ l = 1 m TH i , j L ( l ) , Wherein TH () is the motion vector direction indication value θ by current pixel point and front and back three frame relevant position pixels thereof N, i, iThe histogram of being formed, P tBe the corresponding probability-distribution function of histogram TH (), m is a histogram space size, the relevant frame number on the L express time axle; Calculate the temporal correlation entropy of the motion vector value of each pixel thus Ct ( i , j ) = - Σ n = 1 m P t ( n ) Log ( P t ( n ) ) ; The temporal information entropy of Ct () expression motion vector;
Step 2.4, generalized time and spatial information, obtain final space time information entropy C (i, j)=a 1 *Ct (i, j)+a 2 *Cs (i, j), a wherein 1+ a 2=1;
Step 2.5, in a two field picture, the minimum space time information entropy of order be Min[C (i, j)], represent with message level 0, make maximum space time information entropy be Max[C (i, j)], l-1 represents with message level, R={0,1..., l-1} represent the set of message level; Definition N p(p ∈ R) pixel quantity when being p for message level promptly has the pixel number of identical information entropy; For threshold value t ∈ R, need find out wherein the pairing space time information entropy of certain one-level in the l-1 grade as threshold value t in 0 grade, and carry out self adaptation according to threshold value t and divide, the comentropy that promptly is lower than threshold value is E A = - Σ j = i + 1 l - 1 N j Σ n = t + 1 l - 1 N n Log ( N j Σ n = t + 1 l - 1 N n ) ; The comentropy that is higher than threshold value is E B = - Σ i = 0 t N i Σ m = 0 t N m Log ( N i Σ m = 0 t N m ) ; Threshold value wherein t = arg max R ( E A + E B ) ; Promptly when the space time information entropy of pixel during greater than threshold value t, promptly this pixel is in the moving region, otherwise is in non-moving region.
The ratio of the shared whole macroblock number of motion object piece is R in each frame in the described step 4 Mb(n)=N Motion(n)/N All, the linear prediction model of revised mean absolute difference MAD is MAD (n)=k 1* R Mb(n) * MAD (n-1)+k 2, N wherein Motion(n) be the number of the shared macro block of motion object in the current n frame, N AllBe the macro block sum that each frame comprised.
The ratio of total pixel is R in the shared macro block of pixel in the sound attitude comprehensive characteristics model in the described step 5 Pixel(n, I, J)=N (n, I, J)/N All, wherein (n, I J) are (I, the number of the pixel of the static nature that J) is extracted on the piece in the current encoded frame n frame to N; N AllBe the total pixel number of each macro block.
Adjusting parameter in the described step 5 be α (n, I, J)=R Pixel(n, I, J)+and b, the target bit mean allocation formula after regulating is Bit ( n , I , J ) ′ = α ( n , I , J ) f rb ( n , I , J ) / Σ l = I , k = J l = c , k = d α ( n , l , k ) , Wherein b prevents that bit number is assigned as 0 radix that is provided with;
Figure A200910049042D00101
For in the present frame from the (I, J) macro block begin all not coded macroblocks the adjusting parameter and, c represents the macro block number that every row comprised in each two field picture, d represents the macro block number that every row comprised in each two field picture.
In the step 1.1, think that every adjacent h pixel has identical YUV value in the described macro block.Described pixel is as follows from the transformational relation that yuv format converts rgb format to: R=Y+1.402 (V-128); G=Y-0.34414 (U-128)-0.71414 (V-128); B=Y+1.772 (U-128).The relation of described R, G, B and Y and three color component r, g and b is as follows: R=r-(g+b)/2; G=g-(r+b)/2; B=b-(r+g)/2; Y=r+g-2 (| r-g|+b).
Comprise also in the step 2.1 that motion vector is made further average value filtering to be handled.
The present invention compared with prior art, not only makes decoded video have better visual effect owing to adopted above-mentioned technical scheme; And under the situation of limit bit number, can realize the distribution of target bits more efficiently.
Description of drawings
Bit rate control method based on human eye sensing model of the present invention is provided in detail by following embodiment and accompanying drawing.
Fig. 1 is the implementation procedure schematic diagram of the embodiment of the invention based on the bit rate control method of human eye sensing model;
Fig. 2 is an a certain two field picture in the video of the embodiment of the invention;
The video static nature hum pattern of Fig. 3 for extracting according to the color contrast principle in the embodiment of the invention;
Fig. 4 is the video behavioral characteristics figure before adaptive threshold is divided in the embodiment of the invention;
Fig. 5 is the video behavioral characteristics figure after adaptive threshold is divided in the embodiment of the invention;
Fig. 6 is video sound attitude characteristic synthetic figure in the embodiment of the invention.
Embodiment
Below will be described in further detail the bit rate control method based on human eye sensing model of the present invention.
Accompanying drawing 1 has simply presented the implementation procedure of present embodiment based on the bit rate control method of human eye sensing model, when video image is as shown in Figure 2 encoded, described bit rate control method carries out frame level and macro-block level Rate Control respectively, the size of macro block is 4*4, and described bit rate control method is realized by following steps:
The static nature of step 1, extraction video image, and set up the static nature model, as shown in Figure 3, specifically comprise the following steps:
Step 1.1, convert each pixel of current video image to rgb format from yuv format, i.e. R=Y+1.402 (V-128); G=Y-0.34414 (U-128)-0.71414 (V-128); B=Y+1.772 (U-128); In order to handle conveniently, in the present embodiment, every adjacent 4 pixels can think to have identical YUV value in the macro block, promptly the YUV value of first pixel are composed the YUV value of giving other 3 pixels; Then extract red R, green G, blue B and yellow Y four primary colours, and the chromaticity figure that obtains four primary colours is respectively R (r, g, b), G (r, g, b), B (r, g, b) and Y (r, g, b), be expressed as follows with color three-component r, g and b: R=r-(g+b)/2, G=g-(r+b)/2, B=b-(r+g)/2, Y=r+g-2 (| r-g|+b); Extract red/green, indigo plant/yellow two groups of antagonistic pairs, and obtain red/green characteristic pattern RG (i, j)=| (R (and i, j)-G (i, j)) | and indigo plant/yellow characteristic pattern BY (i, j)=| (B (and i, j)-Y (i, j)) |, wherein (i, j) coordinate of remarked pixel point;
Step 1.2, calculate each pixel and the mean value of red/green, the indigo plant/yellow characteristic value of eight points on every side thereof Diff RG ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | RG ( i , j ) - RG ( n , m ) | With Diff BY ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | BY ( i , j ) - BY ( n , m ) | ;
Step 1.3, calculate each pixel static nature figure StaticMap (i, j), wherein StaticMap ( i , j ) = Diff BY ( i , j ) 2 + Diff RG ( i , j ) 2 , The maximum of static nature figure is designated as StaticMap Max, minimum value is designated as StaticMap Min, establish and judge that the static nature figure threshold value whether pixel meets the static nature model is T, then T=(StaticMap Max+ StaticMap Min)/2; That is to say that (i, during j) greater than T, this pixel promptly is included in the static nature model, otherwise is not included in the static nature model, the video information of paying close attention to for non-human eye as the static nature figure of pixel StaticMap.
The behavioral characteristics of step 2, extraction video image, and set up the behavioral characteristics model, as shown in Figure 4, specifically comprise the following steps:
Step 2.1, establish n frame (I, J) motion vector of macro block be designated as PV (I, J)=(x N, I, J, y N, I, J), the motion vector of this macro block is defaulted as the motion vector of this each pixel of macro block, the direction of motion of this vector can be expressed as θ N, i, i=arctan (y N, i, j/ x N, i, j), in order to obtain better effects, further motion vector is made average value filtering and handle;
Step 2.2, calculate the current pixel point and the probability histogram distribution function of the motion vector value of eight points on every side thereof P s ( n ) = SH i , j w ( n ) Σ l = 1 m SH i , j w ( l ) ; Wherein SH () is by the current pixel point and the direction value θ of the motion vector of eight points on every side thereof N, i, iThe histogram of being formed; Calculate the spatial coherence entropy of the motion vector value of each pixel according to the probability distribution situation of gained Cs ( i , j ) = - Σ n = 1 m P s ( n ) Log ( P s ( n ) ) ; Wherein, the spatial information entropy of Cs () expression motion vector, P sBe the corresponding probability-distribution function of histogram SH (), m is a histogram space size, is the amount of calculation in the probability histogram function, is obtained by the statistical computation value; W represents the search window size of N*N, and present embodiment adopts the 3*3 window, and promptly w is 9.
The probability histogram distribution function of the motion vector value of the pixel on the same position of the motion vector value of step 2.3, current pixel point and front and back three frames thereof is P t ( n ) = TH i , j L ( n ) Σ l = 1 m TH i , j L ( l ) ; Calculate the temporal correlation entropy of the motion vector value of each pixel thus Ct ( i , j ) = - Σ n = 1 m P t ( n ) Log ( P t ( n ) ) ; Wherein TH () is the motion vector direction indication value θ by current pixel point and front and back three frame relevant position pixels thereof N, i, iThe histogram of being formed; The temporal information entropy of Ct () expression motion vector, P tBe the corresponding probability-distribution function of histogram TH (), m is a histogram space size, and relevant frame number gets 7 in the present embodiment on the L express time axle;
Step 2.4, generalized time and spatial information, obtain final space time information entropy C (i, j)=a 1* Ct (i, j)+a 2* Cs (i, j), a wherein 1+ a 2=1, a in the present embodiment 1=0.7; a 2=0.3;
Step 2.5, in a two field picture, the minimum space time information entropy of order be Min[C (i, j)], represent with message level 0, make maximum space time information entropy be Max[C (i, j)], l-1 represents with message level.R={0,1..., l-1} represent the set of message level; Definition N p(p ∈ R) pixel quantity when being p for message level promptly has the pixel number of identical information entropy; For threshold value t ∈ R, need find out wherein the pairing space time information entropy of certain one-level in the 1-1 grade as threshold value t in 0 grade, and carry out self adaptation according to threshold value t and divide, the comentropy that promptly is lower than threshold value is E A = - Σ j = t + 1 l - 1 N j Σ n = t + 1 l - 1 N n Log ( N j Σ n = t + 1 l - 1 N n ) ; The comentropy that is higher than threshold value is E B = - Σ i = 0 t N i Σ m = 0 t N m Log ( N i Σ m = 0 t N m ) ; Threshold value wherein t = arg max R ( E A + E B ) ; Promptly when the space time information entropy of pixel during greater than threshold value t, promptly this pixel is in the moving region, otherwise is in non-moving region, the video behavioral characteristics figure before dividing as shown in Figure 4, the video behavioral characteristics figure after the division is as shown in Figure 5;
Step 3, get the common factor of static nature model and behavioral characteristics model, obtain the moving image portion and the profile thereof of this video, i.e. sound attitude comprehensive characteristics model, as shown in Figure 6;
Step 4, in the frame level bit-rate control algolithm, the behavioral characteristics model that utilize to extract calculates motion object piece in each frame, i.e. the macroblock number that comprises in the motion model, the ratio R of shared whole macroblock number Mb(n)=N Motion(n)/N All, and with former linear prediction model MAD (n)=k of this ratio correction mean absolute difference MAD 1* MAD (n-1)+k 2, revised MAD model be MAD ' (n)=k 1* R Mb(n) * MAD ' (n-1)+k 2, when the preassignment bit number, according to the shared macroblock number of motion object what, reflect the motion complexity of present frame, thus decision allocation bit number how much realize Rate Control, wherein k 1Be the linear scale factor of this model, k 2Be the constant coefficient of this model, N Motion(n) be the number of the shared macro block of motion object in the current n frame, N AllBe the macro block sum that each frame comprised;
Step 5, in the macro-block level Rate Control, at first according to sound attitude comprehensive characteristics model, calculate the pixel in the sound attitude comprehensive characteristics model, the pixel paid close attention to of human eye just, the ratio R of total pixel in the shared macro block Pixel(n, I, J)=N (n, I, J)/N All, and by this ratio obtain one regulate parameter alpha (n, I, J)=R Pixel(n, I, J)+b, the original target bit mean allocation of usefulness adjusting parameter modification formula Bit in Rate Control (n, I, J)=f Rb(n, I, J)/N Ub, the target bit mean allocation formula after regulating is Bit ( n , I , J ) ′ = α ( n , I , J ) f rb ( n , I , J ) / Σ l = I , k = J l = c , k = d α ( n , l , k ) , Wherein b prevents that bit number is assigned as 0 radix that is provided with, and gets 1 herein;
Figure A200910049042D00141
For in the present frame from (I, J) macro block begin all not coded macroblocks the adjusting parameter and, c represents the macro block number that every row comprised in each two field picture, and d represents the macro block number that every row comprised in each two field picture, thereby realizes the best Rate Control effect under the human eye perception.
By the bit rate control method based on human eye sensing model of the present invention, not only make decoded video have better visual effect; And under the situation of limit bit number, can realize the distribution of target bits more efficiently.

Claims (10)

1, a kind of bit rate control method based on human eye sensing model, described bit rate control method carries out frame level and macro-block level Rate Control respectively, to determine the distribution of target bit, wherein, the size of macro block is h*h, h is a natural number, it is characterized in that, described bit rate control method is realized by following steps:
The static nature of step 1, extraction video image, and set up the static nature model;
The behavioral characteristics of step 2, extraction video image, and set up the behavioral characteristics model;
Step 3, get the common factor of static nature model and behavioral characteristics model, obtain the moving image portion and the profile thereof of this video, i.e. sound attitude comprehensive characteristics model;
Step 4, in the frame level bit-rate control algolithm, utilize the behavioral characteristics model that extracts, calculate the ratio of the shared whole macroblock number of motion object piece in each frame, and with former linear prediction model MAD (the n)=k of mean absolute difference MAD of corresponding macro block pixels point in each macro block in this ratio correction current encoded frame and the corresponding reference frame of coding 1* MAD (n-1)+k 2, when the preassignment bit number,, reflect the motion complexity of present frame according to what of motion object piece, thus decision allocation bit number how much realize Rate Control, wherein k 1Be the linear scale factor of this model, k 2Be the constant coefficient of this model, n is the picture frame sequence number;
Step 5, in the macro-block level Rate Control, at first according to sound attitude comprehensive characteristics model, calculate the ratio of total pixel in the shared macro block of pixel in the sound attitude comprehensive characteristics model, and by an adjusting of this ratio acquisition parameter, in Rate Control, use and regulate the original target bit mean allocation of parameter modification formula Bit (n, I, J)=f Rb(n, I, J)/N Ub, realize the best Rate Control effect under the human eye perception, wherein, n is the picture frame sequence number, ((n, I J) are target bit, f to Bit for I, the J) coordinate of expression macro block Rb(n, I J) are total bit number to be allocated, N UbIt is number of macroblocks to be allocated.
2, the bit rate control method based on human eye sensing model as claimed in claim 1 is characterized in that, the foundation of the extraction of the static nature in the described step 1 and static nature model further realizes by following steps:
Step 1.1, convert each pixel of current video to rgb format from yuv format; Extract red R, green G, blue B and yellow Y four primary colours, and the chromaticity figure that obtains four primary colours be respectively R (r, g, b), G (r, g, b), B (r, g, b) and Y (r, g, b); Extract red/green, indigo plant/yellow two groups of antagonistic pairs, and obtain red/green characteristic pattern RG (i, j)=| (R (and i, j)-G (i, j)) | and indigo plant/yellow characteristic pattern BY (i, j)=| (B (and i, j)-Y (i, j)) |, wherein (i, j) coordinate of remarked pixel point;
Step 1.2, calculate each pixel and the mean value of red/green, the indigo plant/yellow characteristic value of eight points on every side thereof Diff RG ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | RG ( i , j ) - RG ( n , m ) | With Diff BY ( i , j ) = 1 9 Σ n = i - 1 i + 1 Σ m = j - 1 j + 1 | BY ( i , j ) - BY ( n , m ) | ;
Step 1.3, calculate the static nature figure of each pixel Stati cMap ( i , j ) = Diff BY ( i , j ) 2 + Diff RG ( i , j ) 2 , The maximum of static nature figure is designated as StaticMap Max, minimum value is designated as StaticMap Min, establish and judge that the static nature figure threshold value whether pixel meets the static nature model is T, then T=(StaticMap Max+ StaticMap Min)/2, (i, during j) greater than T, this pixel promptly is included in the static nature model, otherwise is not included in the static nature model as the static nature figure of pixel StaticMap.
3, the bit rate control method based on human eye sensing model as claimed in claim 1 is characterized in that, the foundation of the extraction of the behavioral characteristics in the described step 2 and behavioral characteristics model further realizes by following steps:
Step 2.1, establish n frame (I, J) motion vector of macro block be designated as PV (I, J)=(x N, I, J, y N, I, J), the motion vector of this macro block is defaulted as the motion vector of this each pixel of macro block, the direction of motion of this vector is expressed as θ N, i, j=arctan (y N, i, j/ x N, i, j);
Step 2.2, calculate the current pixel point and the probability histogram distribution function of the motion vector value of eight points on every side thereof P s ( n ) = SH i , j w ( n ) Σ l = 1 m SH i , j w ( l ) ; Wherein SH () is by the current pixel point and the direction value θ of the motion vector of eight points on every side thereof N, i, iThe histogram of being formed, m are histogram space size, and w represents the search window size of N*N; Calculate the spatial coherence entropy of the motion vector value of each pixel according to the probability distribution situation of gained Cs ( i , j ) = - Σ n = 1 m P s ( n ) Log ( P S ( n ) ) ; Wherein, the spatial information entropy of Cs () expression motion vector, P sIt is the corresponding probability-distribution function of histogram SH ();
The probability histogram distribution function of the motion vector value of the pixel on the same position of step 2.3, the motion vector value that calculates current pixel point and front and back three frames thereof P t ( n ) = TH i , j L ( n ) Σ l = 1 m TH i , j L ( l ) , Wherein TH () is the motion vector direction indication value θ by current pixel point and front and back three frame relevant position pixels thereof N, i, iThe histogram of being formed, P tBe the corresponding probability-distribution function of histogram TH (), m is a histogram space size, the relevant frame number on the L express time axle; Calculate the temporal correlation entropy of the motion vector value of each pixel thus Ct ( i , j ) = - Σ n = 1 m P t ( n ) Log ( P t ( n ) ) ; The temporal information entropy of Ct () expression motion vector;
Step 2.4, generalized time and spatial information, obtain final space time information entropy C (i, j)=a 1* Ct (i, j)+a 2* Cs (i, j), a wherein 1+ a 2=1;
Step 2.5, in a two field picture, the minimum space time information entropy of order be Min[C (i, j)], represent with message level 0, make maximum space time information entropy be Max[C (i, j)], l-1 represents with message level, R={0,1..., l-1} represent the set of message level; Definition N p(p ∈ R) pixel quantity when being p for message level promptly has the pixel number of identical information entropy; For threshold value t ∈ R, need find out wherein the pairing space time information entropy of certain one-level in the l-1 grade as threshold value t in 0 grade, and carry out self adaptation according to threshold value t and divide, the comentropy that promptly is lower than threshold value is E A = - Σ j = t + 1 l - 1 N j Σ n = t + 1 l - 1 N n Log ( N j Σ n = t + 1 l - 1 N n ) ; The comentropy that is higher than threshold value is E B = - Σ i = 0 t N i Σ m = 0 t N m Log ( N i Σ m = 0 t N m ) ; Threshold value t=arg max (E wherein A+ E B); Promptly when the space time information entropy of pixel during greater than threshold value t, promptly this pixel is in the moving region, otherwise is in non-moving region.
4, the bit rate control method based on human eye sensing model as claimed in claim 1 is characterized in that, the ratio of the shared whole macroblock number of motion object piece is R in each frame in the described step 4 Mb(n)=N Motion(n)/N All, the linear prediction model of revised mean absolute difference MAD be MAD ' (n)=k 1* R Mb(n) * MAD ' (n-1)+k 2, N wherein Motion(n) be the number of the shared macro block of motion object in the current n frame, N AllBe the macro block sum that each frame comprised.
5, the bit rate control method based on human eye sensing model as claimed in claim 1 is characterized in that, the ratio of total pixel is R in the shared macro block of pixel in the sound attitude comprehensive characteristics model in the described step 5 Pixel(n, I, J)=N (n, I, J)/N All, wherein (n, I J) are (I, the number of the pixel of the static nature that J) is extracted on the piece in the current encoded frame n frame to N; N AllBe the total pixel number of each macro block.
6, the bit rate control method based on human eye sensing model as claimed in claim 5 is characterized in that, the adjusting parameter in the described step 5 be α (n, I, J)=R Pixel(n, I, J)+and b, the target bit mean allocation formula after regulating is Bit ( n , I , J ) ′ = α ( n , I , J ) f rb ( n , I , J ) / Σ l = I , k = J l = c , k = d α ( n , l , k ) , Wherein b prevents that bit number is assigned as 0 radix that is provided with;
Figure A200910049042C00052
For in the present frame from the (I, J) macro block begin all not coded macroblocks the adjusting parameter and, c represents the macro block number that every row comprised in each two field picture, d represents the macro block number that every row comprised in each two field picture.
7, the bit rate control method based on human eye sensing model as claimed in claim 2 is characterized in that, in the step 1.1, every adjacent h pixel is considered to have identical YUV value in the described macro block.
8, the bit rate control method based on human eye sensing model as claimed in claim 2 is characterized in that, in the step 1.1, described pixel is as follows from the transformational relation that yuv format converts rgb format to: R=Y+1.402 (V-128); G=Y-0.34414 (U-128)-0.71414 (V-128); B=Y+1.772 (U-128).
9, the bit rate control method based on human eye sensing model as claimed in claim 2 is characterized in that, in the step 1.1, the relation of described R, G, B and Y and three color component r, g and b is as follows: R=r-(g+b)/2; G=g-(r+b)/2; B=b-(r+g)/2; Y=r+g-2 (| r-g|+b).
10, the bit rate control method based on human eye sensing model as claimed in claim 3 is characterized in that: comprise also in the step 2.1 that motion vector is made further average value filtering to be handled.
CN 200910049042 2009-04-09 2009-04-09 Method for controlling code rate based on human eye sensing model Pending CN101534432A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 200910049042 CN101534432A (en) 2009-04-09 2009-04-09 Method for controlling code rate based on human eye sensing model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 200910049042 CN101534432A (en) 2009-04-09 2009-04-09 Method for controlling code rate based on human eye sensing model

Publications (1)

Publication Number Publication Date
CN101534432A true CN101534432A (en) 2009-09-16

Family

ID=41104782

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 200910049042 Pending CN101534432A (en) 2009-04-09 2009-04-09 Method for controlling code rate based on human eye sensing model

Country Status (1)

Country Link
CN (1) CN101534432A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101827267A (en) * 2010-04-20 2010-09-08 上海大学 Code rate control method based on video image segmentation technology
CN102036073A (en) * 2010-12-21 2011-04-27 西安交通大学 Method for encoding and decoding JPEG2000 image based on vision potential attention target area
CN103024387A (en) * 2012-12-17 2013-04-03 宁波大学 Multi-view video bit rate control method based on sensing
CN103258334A (en) * 2013-05-08 2013-08-21 电子科技大学 Method of estimating scene light source colors of color image
CN110062236A (en) * 2019-05-10 2019-07-26 上海大学 Based on Space-time domain just can perceptual distortion code rate allocation method, system and medium
CN110365983A (en) * 2019-09-02 2019-10-22 珠海亿智电子科技有限公司 A kind of macro-block level bit rate control method and device based on human visual system
CN110708570A (en) * 2019-10-21 2020-01-17 腾讯科技(深圳)有限公司 Video coding rate determining method, device, equipment and storage medium
CN110769254A (en) * 2019-10-10 2020-02-07 网宿科技股份有限公司 Code rate configuration method, system and equipment for video frame
CN112738518A (en) * 2019-10-28 2021-04-30 北京博雅慧视智能技术研究院有限公司 Code rate control method for CTU (China train unit) -level video coding based on perception
WO2022156688A1 (en) * 2021-01-19 2022-07-28 华为技术有限公司 Layered encoding and decoding methods and apparatuses

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101827267A (en) * 2010-04-20 2010-09-08 上海大学 Code rate control method based on video image segmentation technology
CN101827267B (en) * 2010-04-20 2012-07-04 上海大学 Code rate control method based on video image segmentation technology
CN102036073A (en) * 2010-12-21 2011-04-27 西安交通大学 Method for encoding and decoding JPEG2000 image based on vision potential attention target area
CN102036073B (en) * 2010-12-21 2012-11-28 西安交通大学 Method for encoding and decoding JPEG2000 image based on vision potential attention target area
CN103024387A (en) * 2012-12-17 2013-04-03 宁波大学 Multi-view video bit rate control method based on sensing
CN103024387B (en) * 2012-12-17 2015-12-09 宁波大学 A kind of multi-view video rate control based on perception
CN103258334A (en) * 2013-05-08 2013-08-21 电子科技大学 Method of estimating scene light source colors of color image
CN103258334B (en) * 2013-05-08 2015-11-18 电子科技大学 The scene light source colour method of estimation of coloured image
CN110062236A (en) * 2019-05-10 2019-07-26 上海大学 Based on Space-time domain just can perceptual distortion code rate allocation method, system and medium
CN110062236B (en) * 2019-05-10 2021-04-23 上海大学 Code rate allocation method, system and medium based on just-perceivable distortion of space-time domain
CN110365983A (en) * 2019-09-02 2019-10-22 珠海亿智电子科技有限公司 A kind of macro-block level bit rate control method and device based on human visual system
CN110769254A (en) * 2019-10-10 2020-02-07 网宿科技股份有限公司 Code rate configuration method, system and equipment for video frame
CN110769254B (en) * 2019-10-10 2022-04-22 网宿科技股份有限公司 Code rate configuration method, system and equipment for video frame
CN110708570A (en) * 2019-10-21 2020-01-17 腾讯科技(深圳)有限公司 Video coding rate determining method, device, equipment and storage medium
CN112738518A (en) * 2019-10-28 2021-04-30 北京博雅慧视智能技术研究院有限公司 Code rate control method for CTU (China train unit) -level video coding based on perception
CN112738518B (en) * 2019-10-28 2022-08-19 北京博雅慧视智能技术研究院有限公司 Code rate control method for CTU (China train unit) level video coding based on perception
WO2022156688A1 (en) * 2021-01-19 2022-07-28 华为技术有限公司 Layered encoding and decoding methods and apparatuses

Similar Documents

Publication Publication Date Title
CN101534432A (en) Method for controlling code rate based on human eye sensing model
CN104539962B (en) It is a kind of merge visually-perceptible feature can scalable video coding method
KR100997060B1 (en) Video sensor-based automatic region-of-interest detection
CN110446041B (en) Video encoding and decoding method, device, system and storage medium
CN108293124A (en) The coding of picture in video
CN101341494B (en) Video frame motion-based automatic region-of-interest detection
CN100484244C (en) Image coding and decoding processing method based on picture element statistical characteristic and visual characteristic
CN107483934A (en) Decoding method, device and system
WO2019092463A1 (en) Video image processing
CN107222748B (en) The treating method and apparatus of image data code rate
CN117640942A (en) Coding method and device for video image
US20170094281A1 (en) Compressing high dynamic range images
CN106303521B (en) A kind of HEVC Rate-distortion optimization method based on sensitivity of awareness
CN108810530A (en) A kind of AVC bit rate control methods based on human visual system
CN106101703B (en) A kind of screen video compression method towards digital KVM switcher
CN103561270A (en) Coding control method and device for HEVC
CN104811730B (en) A kind of texture analysis of video image intraframe coding unit and coding unit system of selection
CN107197266B (en) HDR video coding method
US9392294B2 (en) Color image data compression
CN101710985A (en) Image brightness compensation method for image coding
CN102663682A (en) Adaptive image enhancement method based on interesting area
CN103313055A (en) Intra-frame prediction method based on segmented chrominance and video coding and decoding method
US8369423B2 (en) Method and device for coding
CN106295587B (en) A kind of video interested region quick calibrating method
CN105791825B (en) A kind of screen picture coding method based on H.264 with hsv color quantization

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Open date: 20090916