CN106162167B - Efficient video coding method based on study - Google Patents
Efficient video coding method based on study Download PDFInfo
- Publication number
- CN106162167B CN106162167B CN201510137157.3A CN201510137157A CN106162167B CN 106162167 B CN106162167 B CN 106162167B CN 201510137157 A CN201510137157 A CN 201510137157A CN 106162167 B CN106162167 B CN 106162167B
- Authority
- CN
- China
- Prior art keywords
- coding
- coding unit
- unit
- mode
- depth
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Landscapes
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Efficient video coding method based on study uses efficient video coding device encoded video sequence, extracts the corresponding feature vector of each coding unit block;The learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, establishes learning model.Early-abort strategy structure is added in the selection course that coding unit size is carried out in efficient video coding device, and Direct Model and fusion mode current block is first carried out, and extracts the corresponding feature vector of corresponding present encoding.Feature vector is inputted to the learning machine model succeeded in school, exports predicted value, executes current coded unit size according to corresponding early-abort strategy structure.Until coding unit layer all in coding tree unit all encodes completion;It repeats until coding tree unit all encodes completion in all video frames.Therefore, optimal cataloged procedure can be exported according to rate distortion costs and computation complexity are corresponding, improves study and the classification performance of classifier, and then improve the code efficiency of Video coding.
Description
Technical field
The present invention relates to image-signal processing methods, more particularly to one kind efficiently based on the efficient video coding of study
Method.
Background technique
Due to being capable of providing better perceived quality and visual experience more true to nature, high definition (High Definition, HD)
It becomes more and more popular with ultra high-definition (Ultra High Definition) video and is liked by people.These high definitions and superelevation
Clear video has wide application market, including high definition television broadcast, MAX film, immersion video communication, network video-on-demand
And high-definition video monitoring etc..However, since high definition and ultra high-definition video have higher clarity and video frame rate, video counts
According to amount also great increase.For example, the video that 120 frame of 8K × 4K@high definition/ultra high-definition video per second has 11.5GB per second
It is effectively stored and transmitted and just needs very efficient video compress by initial data.For effective solution HD video
Compression problem, Video coding joint working group (Joint Collaborative Team on Video Coding, JCT-
VC efficient video coding (High Efficiency Video Coding, HEVC) standard) is proposed.Compared to H.264/AVC
The top grade (high profile) of standard reduces by 50% code rate under the premise of identical visual quality, i.e. compression ratio improves one
Times.Efficient video coding HEVC introduces a variety of advanced coding techniques, including flexible quaternary tree block partition mode, 35 kinds of frames
Inner estimation mode, discrete sine transform, and complicated interpolation and filtering technique etc..These correlative coding technologies effectively improve
Video compression efficiency, however encoder complexity is substantially increased, including computation complexity, CPU consumption, internal storage access disappear
Consumption, battery consumption etc. are unfavorable for the real-time application of high definition and ultra high-definition.
H.264/ coding tree unit structure (Coding Tree Unit, CTU) in efficient video coding HEVC is similar to
The concept of macro block in AVC.CTU includes a luminance block (Coding Tree Block, CTB) and several chrominance blocks, Yi Jiruo
Dry syntactic element.Each luminance block CTB includes a coding unit (Coding Unit, CU) according to video content or is divided
For multiple coding units.The size of each coding unit supports 8 × 8,16 × 16,32 × 32 and 64 in efficient video coding HEVC
Coding unit segmentation sample in × 64, luminance block CTB is as shown in Figure 1, Depth 0 to Depth 4 respectively indicates 64 × 64 to 8
× 8 coding unit size.In addition, each coding unit can be further divided into the predicting unit of different mode and size again
(Prediction Unit, PU), wherein including SKIP, MERGE mode, 8 kinds of inter-frame modes and 2 kinds of frame modes.Finally,
Each predicting unit PU will carry out transition coding using various sizes of converter unit (TransformUnit, TU) again.Efficiently view
Coding unit, predicting unit and converter unit are level recurrence relations in frequency coding HEVC, each coding unit, predicting unit and
Converter unit level has multiple modes, and each layer of optimal mode mainly passes through calculating rate distortion costs (Rate-
Distortion Cost) it obtains, that is, selecting the smallest mode of rate distortion costs is optimal mode.However, this just needs to calculate institute
The rate distortion costs of some levels and mode simultaneously compare to obtain optimal mode, and very time-consuming, computation complexity is very high.
Numerous researchers propose the low complex degree optimization method for efficient video coding HEVC as a result,.Such as pass through
Encoded (Slice) or coding unit are predicted to predict the depth of the coding unit of the coded slice where current coded unit
Range, for reducing computation complexity without test except estimation range.For another example people is believed using the differentiation of movement
Breath is used as main feature, by the movement differential of adjacent encoder cell block and current coded unit block come decision present encoding list
Whether member is divided.Have again and then utilizes rate distortion costs, the pattern dependency progress frame of different depth layer and adjacent encoder unit
The prediction of interior coding unit depth.Machine learning is the hot spot in artificial intelligence, pattern-recognition and signal processing, passes through study
Mode give effective optimal solution.The related algorithm of study is also applied to Video coding by researcher.It such as will be normalized
Rate distortion costs etc. are used as characteristic quantity, its effective mould is realized in input support vector machines (Support Vector Machine, SVM)
Formula classification problem.It is also considered simultaneously because the distortion of rate caused by misclassification increases.Therefore have and utilize back-propagating nerve net
Network (Back Propagation Neural Network, BPNN) solves the pattern classification problem in Video coding.In addition, certainly
Plan tree (Decision Tree) etc. is used for the pattern classification H.264/AVC and its in stereo scopic video coding, such as by SKIP mode
Progress premature termination mode selection processes are fetched for the block differentiation of optimal mode.These algorithms are mainly used for volume H.264/AVC
Code optimization, it is difficult to simply be transplanted in efficient video coding HEVC.
For the efficient video coding optimization based on machine learning, has and propose that the SVM based on weighting carries out coding unit point
Cut with ameristic premature termination algorithm, wherein Feature Selection is optimized, meanwhile, rate distortion costs participate in and conduct
The weighted information of SVM off-line learning.Therefore, there is the code stream information being utilized in MPEG-2 and H.264/AVC video flowing progress
The CU model prediction of HEVC.At the same time, statistical threshold is added, and the situation true to forecasting inaccuracy is further screened.
Generally, the existing coding unit mode prediction method based on machine learning is highly dependent on feature selecting and learning machine
Nicety of grading, once forecasting inaccuracy really will lead to huge compression efficiency decline.Meanwhile conventional method once it is determined that, it is difficult to it is logical
The conversion that code efficiency, computation complexity are realized in parameter adjustment is crossed, the coding requirement suitable for different video system is thus difficult to.
Summary of the invention
Based on this, it is necessary to provide a kind of efficiently efficient video coding method based on study.
A kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature of each coding unit block to
Amount;
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration are learned
Parameter and mode of learning are practised, learning model is established;
Early-abort strategy is added in step 130, the selection course that coding unit size is carried out in efficient video coding device
Structure, wherein in each coding unit depth layer i, Direct Model (SKIP mode) and fusion mode (MERGE mould is first carried out
Formula) current block, the corresponding feature vector of extraction present encoding corresponding with step 110;
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if
Predicted value is not divide, then executes and test current coded unit size, at the same skip the test of partition encoding unit size with
Coding;If predicted value be segmentation, skip test current coded unit size, directly execute segmentation coding size test and
Coding;If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested;
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit has all encoded
At;
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames.
Described eigenvector includes the feature of current coded unit block, motion information, up and down in one of the embodiments,
Literary information, quantization parameter etc. and forced coding unit size.
The feature of the current coded unit block includes encoding block marker x in one of the embodiments,CBF_Meg(i)、
Rate distortion costs value xRD_Meg(i), it is distorted xD_Meg(i) and number of coded bits xBit_Meg(i);Wherein, i is current coded unit
Depth;
The calculation formula of the motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicates fortune
Dynamic and vertical movement amplitude, i are the depth of current coded unit.
Using the time of video macroblock mode, the rate distortion costs in space and coding unit depth as current coded unit
The characteristic quantity of depth decision, is denoted as x respectivelyNB_RD(i) and xCU_depth(i);Wherein, rate distortion costs xNB_RDIt (i) is the adjacent left side
With the evaluation of estimate of the rate distortion costs value of top coding unit;Coding unit depth xCU_depthIt (i) is the flat of adjacent encoder unit
Equal depth;
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABVIt (i) is a left side
4 × 4 pieces of numbers in side and top coding unit.
The feature vector by extraction and forced coding unit size input the output of three values in one of the embodiments,
Learning machine, the step of configuring learning parameter and mode of learning, establishing learning model includes:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value
Oi,+1 or -1, wherein i indicates the label of learning machine, is 1 to m;
To m output OiIt is merged, obtaining final output is QALL;
Wherein, TAAnd TBFor 0 to m two threshold values.
It in one of the embodiments, further include that feature is inputted to classifier, the classifier will be according in current video
Hold characteristic and the model parameter learned is made prediction A (+1), B (- 1) or C (U is not known);
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn;
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1;
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1。
The incrementss of rate distortion costs caused by misprediction can indicate in one of the embodiments, are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALLIt (i) is to divide (B) and do not divide (A)
Misprediction rate, NBA,1(i) and NAB,2It (i) is respectively B misprediction is A in No.1 classifier number of samples, A in No. 2 classifiers
Misprediction is the number of samples of B, NALL(i) number of samples predicted for i-th layer of coding unit, wherein NALL(1) i.e.
For the coding unit number of whole image;
Divide and is calculated as with the computation complexity that do not divide after optimizing
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to pass through segmentation and ameristic prediction in coding unit depth layer i
Caused by computation complexity reduce percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i),
Wherein TS(i)、TnS(i) and TALL(i) it is respectively segmentation, does not divide and the computation complexity of origin operation;
The calculating after optimization is not divided according to the incrementss of rate distortion costs caused by the misprediction and the segmentation and
Complexity sets objective function;
The objective function is expressed as
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two models ginseng of i-th layer of three value output category device
Number, is expressed asWA(j, i) and WBIt is single that (j, i) respectively indicates i-th layer of coding
In member in j-th of classifier positive and negative samples weighted value.
It in one of the embodiments, further include to configuration learning parameter and mode of learning setting Optimal Learning parameter;
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、
ΔηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti,
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basis
Obtain WA(j,i)/WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、Δ
ηS→nS(i)、ΔηnS→S(i)、bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters.
The training method of the learning machine includes online mode and offline mode in one of the embodiments,.
The online mode includes to n frame video using original HM model based coding, output face in one of the embodiments,
To the characteristic vector X of the learning machine of every a kind of coding unit depth iiAnd the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
The coding unit depth prediction trained support vector machines learning machine being used in coding video frames.
The offline mode includes choose several particular sequences and each sequence several in one of the embodiments,
Frame is encoded using original HM model, exports the study towards every a kind of coding unit depth i in these encoded video frames
The characteristic vector X of machineiAnd the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
It is pre- that trained support vector machines learning machine is used for video sequence, the coding of video frame and coding unit depth
It surveys.
The above-mentioned efficient video coding method based on study establishes morning by setting the classifier of different niceties of grading
Final only policy construction.The early-abort strategy structure can be converted by adjusting the ratio of output becomes the several classics of tradition
Decision structure.Therefore, it can be switched over according to actual needs with the characteristic of classifier.It is preparatory for the feature vector of extraction
The learning machine model succeeded in school can correspond to output predicted value, and different coding modes is selected according to predicted value.It can basis
Rate distortion costs and computation complexity in efficient video coding, which correspond to, exports optimal cataloged procedure, to improve classifier
Study and classification performance, and then improve Video coding code efficiency.
Detailed description of the invention
Fig. 1 is the flow chart of the efficient video coding method based on study;
Fig. 2 is the Fractionation regimen figure of coding unit in coding tree unit;
Fig. 3 is coding tree unit cataloged procedure schematic diagram;
Fig. 4 is that coding unit decision divides schematic diagram;
Fig. 5 (a) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (b) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (c) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (d) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 6 is three output category device flow charts;
Fig. 7 is Δ ηT,iAverage computation complexity under configuring condition reduces and the increased curve graph of code rate;
Fig. 8 is online mode training schematic diagram;
Fig. 9 is offline mode training schematic diagram.
Specific embodiment
As shown in Figure 1, being the flow chart of the efficient video coding method based on study.
A kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature of each coding unit block to
Amount.
Described eigenvector include feature, motion information, contextual information, quantization parameter of current coded unit block etc. with
And forced coding unit size.
The feature of current coded unit block includes encoding block marker xCBF_Meg(i), rate distortion costs value xRD_Meg(i), it loses
True xD_Meg(i) and number of coded bits xBit_Meg(i);Wherein, i is the depth of current coded unit.
The calculation formula of motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicate movement and
Vertical movement amplitude, i are the depth of current coded unit.
Specifically, the information of current coded unit, main includes using SKIP mode and MERGE pattern-coding present encoding
The output information of unit process, including encoding block marker (Coded Block Flag, identification code residual error coefficient number), rate
Distortion cost value, distortion and number of coded bits, this four information are expressed as xCBF_Meg(i)、xRD_Meg(i)、xD_Meg(i)、
xBit_Meg(i), wherein i is the depth of current coded unit, is 0,1,2,3.In addition, after being coding there are one SKIP marker bit
Output information position, 0 or 1, be denoted as xSKIP(i)。
Motion information is mainly used to characterize the motion intense degree of current coded unit, bigger general of general motion intense
Rate uses the coding unit of small size, that is, the probability divided is higher.Using the movement arrow after merge pattern-coding in the present embodiment
Scale levies the movement of current coded unit, is specifically calculated as xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy are respectively indicated
Movement and vertical movement amplitude.
When usually having stronger due to the macroblock mode of video, spatial coherence, as a result, when, space rate be distorted generation
The characteristic quantity of valence and coding unit depth as current coded unit depth decision, is denoted as x respectivelyNB_RD(i) and xCU_depth
(i).In the present embodiment, xNB_RDIt (i) is the evaluation of estimate on the adjacent left side and the rate distortion costs value of top coding unit;xCU_depth
(i) it is the mean depth of adjacent encoder unit, is calculated by the following formula to obtain
Wherein, djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABV(i) it is
4 × 4 pieces of numbers in the left side and top coding unit.
The quantization parameter for encoding current coded unit, is denoted as xQP, usual quantization parameter is bigger, and current coded unit more has can
It can be encoded using biggish piece.
The features above amount enumerated in the present embodiment is main feature amount, the including but not limited to above content, furthermore
It can be to wherein deleting.
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration are learned
Parameter and mode of learning are practised, learning model is established.
Incorporated by reference to Fig. 6.Specifically, three value output category devices are made of the classifier that a two-value of m (m >=2) exports, structure
As shown in figure 5, these two-value output category devices can be well known support vector machines, neural network learning machine or Bayes's classification
Device etc..Finally, merging different output results by combining unit, form the output of three values, be positive respectively (+1), it is negative (-
1) and it is uncertain (U), A, B and C in Fig. 5 (d) are corresponded respectively to, amalgamation mode is embodied as.
Wherein TAAnd TBFor 0 to m two threshold values, work as TA=m, then the A in Fig. 5 (d) is 0, works as TB=0 three output point
The B of class device is 0, TA=m while TB=0, then exporting C is 100%.If TA<TB, then C output is 0.In practical operation, general TA>
TB。
The classifier of one three value output can be made of m different learning machines, can also be by the same learning machine but ginseng
Number composition different with mode of learning, or different parameters form different learning machines simultaneously.In the present embodiment, using multiple supports
Vector machine, different weighting coefficients, m=2, TA=2, TB=1.
Support vector machines learning machine is used in the present embodiment, training process can be divided into two major classes, offline mode and online
Mode.
As shown in figure 8, for online mode training schematic diagram.
In online mode, by using original HM model based coding to n frame video, export towards each class coding unit depth
The characteristic vector X of the learning machine of iiThe best macroblock mode Y of (including 9 above-mentioned characteristic quantities) and each coding unit, by Xi
With Y input support vector machines learning machine training, trained support vector machines learning machine is then used for subsequent video frame coding
In coding unit depth prediction;One sequence of every coding requires re -training in which.
As shown in figure 9, for offline mode training schematic diagram.
In offline mode, several frames of several particular sequences and each sequence are chosen, are encoded using original HM model,
Export the characteristic vector X of the learning machine towards every a kind of coding unit depth i in these encoded video framesi(include above-mentioned 9
A characteristic quantity) and each coding unit best macroblock mode Y, by XiWith Y input support vector machines learning machine training.Then
Trained support vector machines learning machine is used for other video sequences, the coding of video frame and coding unit depth prediction.It should
Mode can be trained the selection of video as required and training updates.
Early-abort strategy is added in step 130, the selection course that coding unit size is carried out in efficient video coding device
Structure, wherein in each coding unit depth layer i, SKIP mode and MERGE mode current block, extraction and step is first carried out
The corresponding feature vector of 110 corresponding present encodings.
In efficient video coding device HEVC Video coding, each image is made of the CTU of some column, coding therein
Final coding unit segmentation is obtained in tree unit and ameristic decision is not an individual two-value decision problem, Er Qieyou
Multiple decision problem compositions.In efficient video coding device HEVC coding, the coding tree unit luma unit of each 64x64 will be first
The coding unit size coding for first using 64x64, calculates rate distortion costs;Then, it is divided into the coding list of 4 32x32
Elemental size calculates separately the rate distortion costs of 4 units.In the process, the coding unit size of each 32x32 can be divided again
For the size coding unit of 4 16x16, successively recurrence, until the coding unit unit of 8x8.As shown in figure 3, wherein DnIt indicates
Coding current coded unit simultaneously calculates rate distortion costs, and n 0,1,2,3 respectively corresponds coding size 64x64 to 8x8 size, Pn
It (i) is then recursive coding unit size decision process, i is the call number of four sub-blocks.Finally, from the coding unit of small size
Recurrence upwards in turn, the rate distortion costs compared with upper one layer, cost it is small be set to better model selection;Successively recurrence, simultaneously
Compare to the end, thus obtains the optimum code unit size partitioning scheme of coding tree unit.
In coding unit size selection process, current coded unit size or smaller 4 coding unit rulers are selected
It is very little can be defined as one point with regardless of the problem of.The coding unit size decision problem of entire coding tree unit can be with as a result,
It is described as the two-value decision problem of 3 levels.As shown in figure 4, the selection of 64x64 or 4 32x32 can be defined as one
Point with regardless of two-value decision problem, need a classifier, be defined as L1 first layer;Then, for second, there will be 4
Point with regardless of the problem of, thus need 4 classifiers (L2);Similar, third layer needs 16 classifiers.These classifiers,
Same layer belongs to the classifier of the same attribute, can be used for multiple times with a classifier or multiple identical classifiers.
Due to the P in Fig. 3n(i) structure is a recursive structure, is repeated in different levels, thus for convenience's sake,
We use Pn(i) optimization of structure describes the optimization of entire coding unit size decision process.
For the P in Fig. 3n(i), in the Knowledge Verification Model of efficient video coding device HEVC (i.e. original video encoder)
It can be described as by a DnWith 4 Pn+1(i) the advantages of composition, process are that structure is executed such as the sequence of Fig. 5 (a), the structure
It is that selection optimal mode code efficiency is high, the disadvantage is that there is a large amount of non-essential calculating, computation complexity is high.During this,
Since optimal mode can select DnOr 4 Pn+1, it is the process of alternative.
Thus multiple pre- geodesic structures, Fig. 5 (b) are one of premature termination scheme, are finishing DnLater, pre- using one
Classifier is surveyed, predicts current best mode and judge whether to skip to execute subsequent Pn+1, thus reduce computation complexity.However
This method will execute D to all blocksnOperation, for selecting Pn+1It is obviously wasted for optimal mode unnecessary.
It is proposed shown in premature termination scheme such as Fig. 5 (c) as a result, i.e., as execution DnAnd Pn+1It is preceding to be predicted first by classifier
Current optimal mode, if it is Dn, then it is set as Y and only carries out Dn, otherwise it is set as N and executes Pn+1.The advantages of structure is only
Predict accurate, not additional calculating cost;The disadvantage is that it is highly dependent on the precision of prediction of classifier, once prediction is not quasi- enough
Really, it will biggish compression efficiency is caused to decline.
It is proposed to this end that as shown in Fig. 5 (d), classifier will make prediction A, B or C according to current video content characteristic, right
In A, i.e. prediction DnFor optimal mode, then D is only carried outn;It then predicts that Pn+1 is optimal mode for B, then only carries out Pn+1;For C
Then indicate uncertain DnOr Pn+1, then D is executednWith 4 Pn+1.The structure is very flexible, has multiple advantages: due to that can pass through
Tri- output items of ABC in sorting algorithm and parameter adjustment classifier, as a result, when it is 100% that A and B, which is 0, C, the structure
It can be exchanged into Fig. 5 (a), there is highest compression efficiency maximum complexity.When B is 0, then the structure becomes Fig. 5 (b);Finally,
When C is 0, then the structure is converted to structure shown in Fig. 5 (c).The structure can realize code efficiency and meter by adjusting A, B, C
The conversion of complexity is calculated, code efficiency and encoder complexity, can be according to reality between structure chart 5 (a) and Fig. 5 (c)
The adjustment of application system demand.
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if
Predicted value is not divide, then executes and test current coded unit size, at the same skip the test of partition encoding unit size with
Coding;If predicted value be segmentation, skip test current coded unit size, directly execute segmentation coding size test and
Coding;If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested.
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit has all encoded
At.
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames.
Efficient video coding method based on study further includes by the rate distortion costs of the time of video macroblock mode, space
And characteristic quantity of the coding unit depth as current coded unit depth decision, it is denoted as x respectivelyNB_RD(i) and xCU_depth(i);
Wherein, rate distortion costs xNB_RDIt (i) is the evaluation of estimate on the adjacent left side and the rate distortion costs value of top coding unit;Coding unit
Depth xCU_depthIt (i) is the mean depth of adjacent encoder unit.
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABVIt (i) is a left side
4 × 4 pieces of numbers in side and top coding unit.
The feature vector by extraction and forced coding unit size input the learning machine of three values output, configuration study ginseng
Several and mode of learning, the step of establishing learning model include:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value
Oi,+1 or -1, wherein i indicates the label of learning machine, is 1 to m.
To m output OiIt is merged, obtaining final output is QALL;
Wherein, TAAnd TBFor 0 to m two threshold values.
Learning machine is support vector machines, but weighting parameters are different.Preferably, m is set as 2, TA=2, TB=1.
Efficient video coding method based on study further includes that feature is inputted to classifier, and the classifier will be according to current
Video content characteristic and the model parameter learned make prediction A (+1), B (- 1) or C (U is uncertain).
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn。
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1。
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1。
Efficient video coding method based on study further includes to configuration learning parameter and mode of learning setting Optimal Learning
Parameter.
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、
ΔηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti。
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basis
Obtain WA(j,i)/WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、Δ
ηS→nS(i)、ΔηnS→S(i)、bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters.
Specifically, needing in the support vector machines learning machine training process of three values output for two support vector machines
Habit machine configures reasonable parameter, has reached optimal prediction effect.Mainly by adjusting support vector machines study in the present embodiment
Weighting coefficient W in machineAAnd WB, carry out adjusting training process.WAAnd WBThe importance of positive and negative samples in sample is respectively indicated, it is bigger
It indicates easier and is divided into such, false acceptance rate will increase, and false rejection rate will be reduced.The knot of these mistake classification or prediction
Fruit will lead to the coding of the code efficiency and computation complexity that finally encode.
Define Δ ηS→nS(i) do not divide those coding units for selecting Fractionation regimen as optimal mode of pattern-coding to use
The incrementss of caused rate distortion costs are represented by Δ ηS→nS(i)=(1-JnS(i)/JBest× 100%, (i)) wherein i
For coding unit decision-making level, i ∈ { 1,2,3 }, JnS(i) and JBest(i) it respectively indicates and is compiled using not Fractionation regimen and optimal mode
Rate distortion costs when code current coded unit.Equally, Δ η is definednS→S(i) for using Fractionation regimen encode those selection regardless of
The incrementss for cutting rate distortion costs caused by the coding unit that mode is optimal mode, are represented by Δ ηnS→S(i)=(1-JS
(i)/JBest(i)) × 100%, wherein JSIt (i) is the rate distortion costs that current coded unit is encoded using Fractionation regimen.As a result,
In the decision process of each layer of coding unit, the incrementss of the rate distortion costs as caused by misprediction can be indicated are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALLIt (i) is to divide (B) and do not divide (A)
Misprediction rate, NBA,1(i) and NAB,2It (i) is respectively B misprediction is A in No.1 classifier number of samples, A in No. 2 classifiers
Misprediction is the number of samples of B, NALL(i) number of samples predicted for i-th layer of coding unit, wherein NALL(1) i.e.
For the coding unit number of whole image.
Define qS,1(i)=NS,1(i)/NALLIt (i) is the percentage that segmentation is predicted as by the classifier that three values export, definition
qnS,2(i)=NnS,2(i)/NALL(i) to be predicted as ameristic percentage, wherein NS,1(i) and NnS,2It (i) is respectively that three values are defeated
The segmentation and ameristic number of samples that classifier 1 and 2 is predicted respectively in classifier out.As a result, by segmentation with do not divide it is excellent
Computation complexity after change may be calculated
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to pass through segmentation and ameristic prediction in coding unit depth layer i
Caused by computation complexity reduce percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i),
Wherein TS(i)、TnS(i) and TALLIt (i) is respectively that segmentation (skips DnOperation), do not divide and (omits 4 Pn+1Operation), and former behaviour
Make (DnWith 4 Pn+1Be carried out) computation complexity.
It is encoded and is counted for 20 frames of BQMall, FourPeople etc. 5 different cycle tests, by above-mentioned parameter
It is fitted to
Wherein bi、ai、ti、BiAnd AiFor fitting parameter;
Wherein ui、viAnd TiFor fitting parameter, these fitting parameters by with the cycle tests of selection, coding frame number etc. no
Together, it is varied.In addition, Δ TnS(i)、ΔTS(i)、ΔηS→nS(i) and Δ ηnS→SIt (i) can also be by partial test sequence
The coding of column and count to obtain.
In the present embodiment, above-mentioned fitting parameter is as follows:
Wherein R2Indicate fitting precision, it is better closer to 1.
Thus the performance optimized in order to obtain, function of setting objectives, i.e., under the conditions of rate distortion costs increase limited,
The computation complexity 1- Δ T (i) for minimizing encoder, is represented by
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two models ginseng of i-th layer of three value output category device
Number, is expressed asWA(j, i) and WBIt is single that (j, i) respectively indicates i-th layer of coding
In member in j-th of classifier positive and negative samples weighted value.Above problem conversion are as follows:
To above-mentioned formula respectively to xi、yiAnd λiLocal derviation is sought, and is set to 0, is obtained
Above-mentioned formula is solved, available:
In the present embodiment, only λiWith Δ ηT,iFor parameter, other coefficients are known constant, although can not be dominant
It indicates, but as long as giving a Δ ηT,iCorresponding λ can be obtained by the methods of known least square methodi, then by λiIt brings into
X can be calculated in formulaiAnd yi, finally obtain WA(j,i)/WBThe ratio of (j, i), the instruction as support vector machines learning machine
Practice parameter.
Based on above-mentioned all embodiments, it is based on using the reference software platform HM12.0 verifying of efficient video coding device is above-mentioned
The efficient video coding method of study.Configuration information includes low delay B frame class, and coded sequence first frame is I frame, remaining frame is
P frame, coding unit size support 64 × 64 to 8 × 8, and motion estimation range 64, other parameters are default parameters.Coding experiments
Implement on computers.
Encoding verification experiment is divided into two parts, encodes 5 cycle tests Basketballpass (416 × 240) first,
Partyscene (832 × 480), Johnny (1280 × 720), Kimono (1920 × 1080), Traffic (2560 ×
1600), using different user configuration parameter, Δ ηT,i, it can be expressed as { 0.1%, 0.1%, 0.1% } respectively, 0.3%,
0.3%, 0.3% }, { 0.5%, 0.5%, 0.5% }, { 0.7%, 0.7%, 0.7% } is denoted as Para_111, Para_333,
Para_555 and Para_777.In addition, { 0.3%, 0.2%, 0.1% }, { 0.6%, 0.4%, 0.2% }, 0.9%, 0.6%,
0.3% }, it is denoted as Para_321, Para_642, Para_963, and thus obtains training parameter WAAnd WBTraining learning machine, then
Learning machine is used for coding unit depth prediction in cataloged procedure.
It is illustrated in figure 7 the volume that the encoder encoded video before 5 videos and optimization is encoded using the encoder after optimization
Code efficiency and computation complexity compare, the saving degree of Y-axis position computation complexity, compared to the code rate before optimization after the what happened latter of X-axis position
Increase percentage, as seen from the figure, in different configuration parameter Δ ηT,iAdjusting under, average computation complexity can reduce by 42%
~56%.
In addition, using parameter Para_642 training learning machine, and the model that training obtains is used for the CU depth in coding
Prediction optimization complexity encodes all frames of 21 sequences in this experiment, while compared existing state-of-the-art three kinds of codings
Method ShenEVIP, ShenTMM and XiongTMM compare, and the present invention is soft compared to former efficient video coding device HEVC Knowledge Verification Model
Part platform HM can reduce computation complexity 28.82% to 70.93%, and average 51.45%, average BDPSNR and BDBR is respectively-
0.061dB and 1.98%, it is almost the same with the compression efficiency of former HM.Due to currently advanced from compression efficiency and computation complexity
Three schemes ShenEVIP, ShenTMM and XiongTMM.
It is that code efficiency and computation complexity compare such as following table:
It is above-mentioned it is all it is used in the examples by two Support vector machines form three values output learning machines,
Learning machine therein could alternatively be other kinds of learning machine, such as Bayes, neural network, decision tree, while learning machine
Quantity can be with more than two.Furthermore the learning machine of three values output both can export learning machine by multiple two-values and form, can also be by
Classification learning machine is directly realized by more than one.
The feature vector of input learning machine includes, but are not limited to four classes mentioned in above-described embodiment, may also include figure
As texture edge, brightness etc.;Meanwhile the characteristic quantity in the present invention can have many forms, such as current coded unit
Feature, the motion information of block are calculated as xMV_Meg(i)=| MVx |+| MVy |, contextual information, quantization parameter etc. and optimal encoding
Code unit size etc. can actually be replaced using other forms of expression.
Classification method based on study is used for the decision of coding unit depth, complicated to reduce the model selection in coding
Degree.Process with a variety of " multiselects one " selected similar to size/depth as coding in actual video cataloged procedure, is removed
Outside coding unit depth decision, there are also predicting unit model selection predicting unit, converter unit model selection converter unit join more
Frame selection is examined, the classification method in above-described embodiment based on study can be used in the processes such as estimation, solves related " multiselect
One " the problem of.
The above-mentioned efficient video coding method based on study establishes morning by setting the classifier of different niceties of grading
Final only policy construction.The early-abort strategy structure can be converted by adjusting the ratio of output becomes the several classics of tradition
Decision structure.Therefore, it can be switched over according to actual needs with the characteristic of classifier.It is preparatory for the feature vector of extraction
The learning machine model succeeded in school can correspond to output predicted value, and different coding modes is selected according to predicted value.It can basis
Rate distortion costs and computation complexity in efficient video coding, which correspond to, exports optimal cataloged procedure, to improve classifier
Study and classification performance, and then improve Video coding code efficiency.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality
It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited
In contradiction, all should be considered as described in this specification.
The embodiments described above only express several embodiments of the present invention, and the description thereof is more specific and detailed, but simultaneously
It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art
It says, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to protection of the invention
Range.Therefore, the scope of protection of the patent of the invention shall be subject to the appended claims.
Claims (9)
1. a kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature vector of each coding unit block;
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration study ginseng
Several and mode of learning, establishes learning model;
Early-abort strategy structure is added in step 130, the selection course that coding unit size is carried out in efficient video coding device,
Wherein, in each coding unit depth layer i, SKIP mode and MERGE mode is first carried out in current coded unit block, is extracted and is walked
The corresponding feature vector of rapid 110 corresponding current coded unit block;
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if prediction
Value is not divide, then executes and test current coded unit size, while skipping test and the coding of partition encoding unit size;
If predicted value is segmentation, skip test current coded unit size directly executes test and the coding of the coding size of segmentation;
If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested;
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit all encodes completion;
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames;
The feature vector of extraction and forced coding unit size, are inputted the learning machine of three values output by the step 120, and configuration is learned
The step of practising parameter and mode of learning, establishing learning model include:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value Oi,+1
Or -1, wherein i indicates the label of learning machine, is 1 to m;
To m output OiIt is merged, obtaining final output is QALL;
Wherein, TAAnd TBFor 0 to m two threshold values.
2. the efficient video coding method according to claim 1 based on study, which is characterized in that described eigenvector packet
Include feature, motion information, contextual information, quantization parameter of current coded unit block etc. and forced coding unit size.
3. the efficient video coding method according to claim 2 based on study, which is characterized in that the present encoding list
The feature of first block includes encoding block marker xCBF_Meg(i), rate distortion costs value xRD_Meg(i), it is distorted xD_Meg(i) it and encodes
Bit number xBit_Meg(i);Wherein, i is the depth of current coded unit;
The calculation formula of the motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicate movement and
Vertical movement amplitude, i are the depth of current coded unit;
Using the time of video macroblock mode, the rate distortion costs in space and coding unit depth as current coded unit depth
The characteristic quantity of decision, is denoted as x respectivelyNB_RD(i) and xCU_depth(i);Wherein, rate distortion costs xNB_RDIt (i) is the adjacent left side and upper
The evaluation of estimate of the rate distortion costs value of side coding unit;Coding unit depth xCU_depthIt (i) is the average depth of adjacent encoder unit
Degree;
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABV(i) for the left side and
4 × 4 pieces of numbers in the coding unit of top.
4. the efficient video coding method according to claim 1 based on study, which is characterized in that further include that feature is defeated
Enter classifier, the classifier will make prediction A, B or C according to current video content characteristic and the model parameter learned;Wherein A
It indicates just, B indicates negative, and C indicates uncertain;
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn;
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1;
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1。
5. the efficient video coding method according to claim 4 based on study, which is characterized in that rate caused by misprediction
The incrementss of distortion cost can indicate are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein Δ ηnS→S(i) for use Fractionation regimen coding select not Fractionation regimen for caused by the coding unit of optimal mode
The incrementss of rate distortion costs, Δ ηS→nS(i) do not divide the coding that pattern-coding selects Fractionation regimen as optimal mode to use
The incrementss of rate distortion costs caused by unit, pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALL(i) it is
Segmentation B and the misprediction rate for not dividing A, NBA,1(i) and NAB,2(i) be respectively in No.1 classifier B misprediction be A sample
Number, A misprediction is the number of samples of B, N in No. 2 classifiersALL(i) sample predicted for i-th layer of coding unit
Number, wherein NALL(1) be whole image coding unit number;
Divide and is calculated as with the computation complexity that do not divide after optimizing
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to be led in coding unit depth layer i by segmentation and ameristic prediction
The computation complexity of cause reduces percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i), wherein
TS(i)、TnS(i) and TALL(i) it is respectively segmentation, does not divide and the computation complexity of origin operation;
The calculating complexity after optimization is not divided according to the incrementss of rate distortion costs caused by the misprediction and the segmentation and
Degree setting objective function;
The objective function is expressed as
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two model parameters of i-th layer of three value output category device,
It is expressed asWA(j, i) and WB(j, i) is respectively indicated in i-th layer of coding unit
The weighted value of positive and negative samples in j-th of classifier.
6. the efficient video coding method according to claim 4 based on study, which is characterized in that further include learning configuration
It practises parameter and Optimal Learning parameter is arranged in mode of learning;
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、Δ
ηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti,
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basisObtain WA(j,i)/
WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、ΔηS→nS(i)、ΔηnS→S(i)、
bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters, h1(i), h2(i), k1(i), k2It (i) is known constant.
7. the efficient video coding method according to claim 1 based on study, which is characterized in that the instruction of the learning machine
The mode of white silk includes online mode and offline mode.
8. the efficient video coding method according to claim 7 based on study, which is characterized in that the online mode packet
It includes to n frame video using original HM model based coding, exports the characteristic vector X of the learning machine towards each class coding unit depth ii
And the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
The coding unit depth prediction trained support vector machines learning machine being used in coding video frames.
9. the efficient video coding method according to claim 7 based on study, which is characterized in that the offline mode packet
Several frames for choosing several particular sequences and each sequence are included, is encoded using original HM model, exports these encoded views
The characteristic vector X of learning machine towards every a kind of coding unit depth i in frequency frameiAnd the best piecemeal mould of each coding unit
Formula Y;
By XiWith Y input support vector machines learning machine training;
Trained support vector machines learning machine is used for video sequence, the coding of video frame and coding unit depth prediction.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510137157.3A CN106162167B (en) | 2015-03-26 | 2015-03-26 | Efficient video coding method based on study |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510137157.3A CN106162167B (en) | 2015-03-26 | 2015-03-26 | Efficient video coding method based on study |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106162167A CN106162167A (en) | 2016-11-23 |
CN106162167B true CN106162167B (en) | 2019-05-17 |
Family
ID=57340278
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510137157.3A Active CN106162167B (en) | 2015-03-26 | 2015-03-26 | Efficient video coding method based on study |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106162167B (en) |
Families Citing this family (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10382770B2 (en) * | 2017-02-06 | 2019-08-13 | Google Llc | Multi-level machine learning-based early termination in partition search for video encoding |
CN106713929B (en) * | 2017-02-16 | 2019-06-28 | 清华大学深圳研究生院 | A kind of video inter-prediction Enhancement Method based on deep neural network |
CN108737841B (en) * | 2017-04-21 | 2020-11-24 | 腾讯科技(深圳)有限公司 | Coding unit depth determination method and device |
CN107690069B (en) * | 2017-08-28 | 2021-01-01 | 中国科学院深圳先进技术研究院 | Data-driven cascade video coding method |
US11436471B2 (en) * | 2017-10-13 | 2022-09-06 | Panasonic Intellectual Property Corporation Of America | Prediction model sharing method and prediction model sharing system |
KR102535361B1 (en) | 2017-10-19 | 2023-05-24 | 삼성전자주식회사 | Image encoder using machine learning and data processing method thereof |
US20200344474A1 (en) * | 2017-12-14 | 2020-10-29 | Interdigital Vc Holdings, Inc. | Deep learning based image partitioning for video compression |
CN108012150B (en) * | 2017-12-14 | 2020-05-05 | 湖南兴天电子科技有限公司 | Video interframe coding method and device |
CN108200442B (en) * | 2018-01-23 | 2021-11-12 | 北京易智能科技有限公司 | HEVC intra-frame coding unit dividing method based on neural network |
CN108174208B (en) * | 2018-02-12 | 2020-05-12 | 杭州电子科技大学 | Efficient video coding method based on feature classification |
CN108495129B (en) * | 2018-03-22 | 2019-03-08 | 北京航空航天大学 | The complexity optimized method and device of block partition encoding based on deep learning method |
CN108924558B (en) * | 2018-06-22 | 2021-10-22 | 电子科技大学 | Video predictive coding method based on neural network |
EP3743855A1 (en) * | 2018-09-18 | 2020-12-02 | Google LLC | Receptive-field-conforming convolution models for video coding |
CN109089114B (en) * | 2018-09-28 | 2022-01-28 | 河海大学 | Complexity-adaptive screen content coding method for virtual reality equipment |
CN109769119B (en) * | 2018-12-18 | 2021-01-19 | 中国科学院深圳先进技术研究院 | Low-complexity video signal coding processing method |
CN110072119B (en) * | 2019-04-11 | 2020-04-10 | 西安交通大学 | Content-aware video self-adaptive transmission method based on deep learning network |
CN110401834B (en) * | 2019-08-06 | 2021-07-27 | 杭州微帧信息科技有限公司 | Self-adaptive video coding method based on deep learning |
CN111083480B (en) * | 2019-12-10 | 2022-11-04 | 四川新视创伟超高清科技有限公司 | High-speed coding method for 8K ultra-high definition video based on CPU |
CN113347415A (en) * | 2020-03-02 | 2021-09-03 | 阿里巴巴集团控股有限公司 | Coding mode determining method and device |
CN111988628B (en) * | 2020-09-08 | 2023-10-17 | 福州大学 | VVC rapid intra-frame coding method based on reinforcement learning |
CN114584771B (en) * | 2022-05-06 | 2022-09-06 | 宁波康达凯能医疗科技有限公司 | Method and system for dividing intra-frame image coding unit based on content self-adaption |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101217663A (en) * | 2008-01-09 | 2008-07-09 | 上海华平信息技术股份有限公司 | A quick selecting method of the encode mode of image pixel block for the encoder |
CN102387356A (en) * | 2010-09-06 | 2012-03-21 | 索尼公司 | Image processing apparatus and method |
CN103517069A (en) * | 2013-09-25 | 2014-01-15 | 北京航空航天大学 | HEVC intra-frame prediction quick mode selection method based on texture analysis |
WO2015034061A1 (en) * | 2013-09-06 | 2015-03-12 | 三菱電機株式会社 | Video encoding device, video transcoding device, video encoding method, video transcoding method and video stream transmission system |
-
2015
- 2015-03-26 CN CN201510137157.3A patent/CN106162167B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101217663A (en) * | 2008-01-09 | 2008-07-09 | 上海华平信息技术股份有限公司 | A quick selecting method of the encode mode of image pixel block for the encoder |
CN102387356A (en) * | 2010-09-06 | 2012-03-21 | 索尼公司 | Image processing apparatus and method |
WO2015034061A1 (en) * | 2013-09-06 | 2015-03-12 | 三菱電機株式会社 | Video encoding device, video transcoding device, video encoding method, video transcoding method and video stream transmission system |
CN103517069A (en) * | 2013-09-25 | 2014-01-15 | 北京航空航天大学 | HEVC intra-frame prediction quick mode selection method based on texture analysis |
Non-Patent Citations (1)
Title |
---|
Fast transrating for high efficiency video coding based on machine learning;Yun Zhang,Sam Kwong,Xu Wang,Hui Yuan,Zhaoqing Pan;《2013 IEEE International Conference on Image Processing》;20140213;第26卷(第3期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN106162167A (en) | 2016-11-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106162167B (en) | Efficient video coding method based on study | |
CN105306947B (en) | video transcoding method based on machine learning | |
CN103873861B (en) | Coding mode selection method for HEVC (high efficiency video coding) | |
Jin et al. | CNN oriented fast QTBT partition algorithm for JVET intra coding | |
CN106454342B (en) | A kind of the inter-frame mode fast selecting method and system of video compression coding | |
CN106937118B (en) | A kind of bit rate control method combined based on subjective area-of-interest and time-space domain | |
CN106131546B (en) | A method of determining that HEVC merges and skip coding mode in advance | |
CN111654698B (en) | Fast CU partition decision method for H.266/VVC | |
CN108737841A (en) | Coding unit depth determination method and device | |
CN107371022A (en) | The quick division methods of interframe encode unit applied to HEVC medical image lossless codings | |
CN111355956A (en) | Rate distortion optimization fast decision making system and method based on deep learning in HEVC intra-frame coding | |
CN105959611A (en) | Adaptive H264-to-HEVC (High Efficiency Video Coding) inter-frame fast transcoding method and apparatus | |
CN109040764A (en) | Fast coding algorithm in a kind of HEVC screen content frame based on decision tree | |
CN110049338B (en) | HEVC (high efficiency video coding) rapid inter-frame coding method based on multi-level classification | |
CN107852492A (en) | Method and apparatus for being coded and decoded to image | |
Chen et al. | A novel fast intra mode decision for versatile video coding | |
CN108989799A (en) | A kind of selection method, device and the electronic equipment of coding unit reference frame | |
CN103313058B (en) | The HEVC Video coding multimode optimization method realized for chip and system | |
CN116489386A (en) | VVC inter-frame rapid coding method based on reference block | |
CN106937116A (en) | Low-complexity video coding method based on random training set adaptive learning | |
CN107690069B (en) | Data-driven cascade video coding method | |
CN108200431A (en) | A kind of video frequency coding rate controls frame-layer Bit distribution method | |
CN103888770B (en) | A kind of video code conversion system efficiently and adaptively based on data mining | |
CN109361920A (en) | A kind of interframe quick predict algorithm of the adaptive decision-making tree selection towards more scenes | |
Huang et al. | Modeling acceleration properties for flexible INTRA HEVC complexity control |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |