CN106162167B - Efficient video coding method based on study - Google Patents

Efficient video coding method based on study Download PDF

Info

Publication number
CN106162167B
CN106162167B CN201510137157.3A CN201510137157A CN106162167B CN 106162167 B CN106162167 B CN 106162167B CN 201510137157 A CN201510137157 A CN 201510137157A CN 106162167 B CN106162167 B CN 106162167B
Authority
CN
China
Prior art keywords
coding
coding unit
unit
mode
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510137157.3A
Other languages
Chinese (zh)
Other versions
CN106162167A (en
Inventor
张云
朱林卫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Institute of Advanced Technology of CAS
Original Assignee
Shenzhen Institute of Advanced Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Institute of Advanced Technology of CAS filed Critical Shenzhen Institute of Advanced Technology of CAS
Priority to CN201510137157.3A priority Critical patent/CN106162167B/en
Publication of CN106162167A publication Critical patent/CN106162167A/en
Application granted granted Critical
Publication of CN106162167B publication Critical patent/CN106162167B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Efficient video coding method based on study uses efficient video coding device encoded video sequence, extracts the corresponding feature vector of each coding unit block;The learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, establishes learning model.Early-abort strategy structure is added in the selection course that coding unit size is carried out in efficient video coding device, and Direct Model and fusion mode current block is first carried out, and extracts the corresponding feature vector of corresponding present encoding.Feature vector is inputted to the learning machine model succeeded in school, exports predicted value, executes current coded unit size according to corresponding early-abort strategy structure.Until coding unit layer all in coding tree unit all encodes completion;It repeats until coding tree unit all encodes completion in all video frames.Therefore, optimal cataloged procedure can be exported according to rate distortion costs and computation complexity are corresponding, improves study and the classification performance of classifier, and then improve the code efficiency of Video coding.

Description

Efficient video coding method based on study
Technical field
The present invention relates to image-signal processing methods, more particularly to one kind efficiently based on the efficient video coding of study Method.
Background technique
Due to being capable of providing better perceived quality and visual experience more true to nature, high definition (High Definition, HD) It becomes more and more popular with ultra high-definition (Ultra High Definition) video and is liked by people.These high definitions and superelevation Clear video has wide application market, including high definition television broadcast, MAX film, immersion video communication, network video-on-demand And high-definition video monitoring etc..However, since high definition and ultra high-definition video have higher clarity and video frame rate, video counts According to amount also great increase.For example, the video that 120 frame of 8K × 4K@high definition/ultra high-definition video per second has 11.5GB per second It is effectively stored and transmitted and just needs very efficient video compress by initial data.For effective solution HD video Compression problem, Video coding joint working group (Joint Collaborative Team on Video Coding, JCT- VC efficient video coding (High Efficiency Video Coding, HEVC) standard) is proposed.Compared to H.264/AVC The top grade (high profile) of standard reduces by 50% code rate under the premise of identical visual quality, i.e. compression ratio improves one Times.Efficient video coding HEVC introduces a variety of advanced coding techniques, including flexible quaternary tree block partition mode, 35 kinds of frames Inner estimation mode, discrete sine transform, and complicated interpolation and filtering technique etc..These correlative coding technologies effectively improve Video compression efficiency, however encoder complexity is substantially increased, including computation complexity, CPU consumption, internal storage access disappear Consumption, battery consumption etc. are unfavorable for the real-time application of high definition and ultra high-definition.
H.264/ coding tree unit structure (Coding Tree Unit, CTU) in efficient video coding HEVC is similar to The concept of macro block in AVC.CTU includes a luminance block (Coding Tree Block, CTB) and several chrominance blocks, Yi Jiruo Dry syntactic element.Each luminance block CTB includes a coding unit (Coding Unit, CU) according to video content or is divided For multiple coding units.The size of each coding unit supports 8 × 8,16 × 16,32 × 32 and 64 in efficient video coding HEVC Coding unit segmentation sample in × 64, luminance block CTB is as shown in Figure 1, Depth 0 to Depth 4 respectively indicates 64 × 64 to 8 × 8 coding unit size.In addition, each coding unit can be further divided into the predicting unit of different mode and size again (Prediction Unit, PU), wherein including SKIP, MERGE mode, 8 kinds of inter-frame modes and 2 kinds of frame modes.Finally, Each predicting unit PU will carry out transition coding using various sizes of converter unit (TransformUnit, TU) again.Efficiently view Coding unit, predicting unit and converter unit are level recurrence relations in frequency coding HEVC, each coding unit, predicting unit and Converter unit level has multiple modes, and each layer of optimal mode mainly passes through calculating rate distortion costs (Rate- Distortion Cost) it obtains, that is, selecting the smallest mode of rate distortion costs is optimal mode.However, this just needs to calculate institute The rate distortion costs of some levels and mode simultaneously compare to obtain optimal mode, and very time-consuming, computation complexity is very high.
Numerous researchers propose the low complex degree optimization method for efficient video coding HEVC as a result,.Such as pass through Encoded (Slice) or coding unit are predicted to predict the depth of the coding unit of the coded slice where current coded unit Range, for reducing computation complexity without test except estimation range.For another example people is believed using the differentiation of movement Breath is used as main feature, by the movement differential of adjacent encoder cell block and current coded unit block come decision present encoding list Whether member is divided.Have again and then utilizes rate distortion costs, the pattern dependency progress frame of different depth layer and adjacent encoder unit The prediction of interior coding unit depth.Machine learning is the hot spot in artificial intelligence, pattern-recognition and signal processing, passes through study Mode give effective optimal solution.The related algorithm of study is also applied to Video coding by researcher.It such as will be normalized Rate distortion costs etc. are used as characteristic quantity, its effective mould is realized in input support vector machines (Support Vector Machine, SVM) Formula classification problem.It is also considered simultaneously because the distortion of rate caused by misclassification increases.Therefore have and utilize back-propagating nerve net Network (Back Propagation Neural Network, BPNN) solves the pattern classification problem in Video coding.In addition, certainly Plan tree (Decision Tree) etc. is used for the pattern classification H.264/AVC and its in stereo scopic video coding, such as by SKIP mode Progress premature termination mode selection processes are fetched for the block differentiation of optimal mode.These algorithms are mainly used for volume H.264/AVC Code optimization, it is difficult to simply be transplanted in efficient video coding HEVC.
For the efficient video coding optimization based on machine learning, has and propose that the SVM based on weighting carries out coding unit point Cut with ameristic premature termination algorithm, wherein Feature Selection is optimized, meanwhile, rate distortion costs participate in and conduct The weighted information of SVM off-line learning.Therefore, there is the code stream information being utilized in MPEG-2 and H.264/AVC video flowing progress The CU model prediction of HEVC.At the same time, statistical threshold is added, and the situation true to forecasting inaccuracy is further screened. Generally, the existing coding unit mode prediction method based on machine learning is highly dependent on feature selecting and learning machine Nicety of grading, once forecasting inaccuracy really will lead to huge compression efficiency decline.Meanwhile conventional method once it is determined that, it is difficult to it is logical The conversion that code efficiency, computation complexity are realized in parameter adjustment is crossed, the coding requirement suitable for different video system is thus difficult to.
Summary of the invention
Based on this, it is necessary to provide a kind of efficiently efficient video coding method based on study.
A kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature of each coding unit block to Amount;
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration are learned Parameter and mode of learning are practised, learning model is established;
Early-abort strategy is added in step 130, the selection course that coding unit size is carried out in efficient video coding device Structure, wherein in each coding unit depth layer i, Direct Model (SKIP mode) and fusion mode (MERGE mould is first carried out Formula) current block, the corresponding feature vector of extraction present encoding corresponding with step 110;
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if Predicted value is not divide, then executes and test current coded unit size, at the same skip the test of partition encoding unit size with Coding;If predicted value be segmentation, skip test current coded unit size, directly execute segmentation coding size test and Coding;If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested;
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit has all encoded At;
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames.
Described eigenvector includes the feature of current coded unit block, motion information, up and down in one of the embodiments, Literary information, quantization parameter etc. and forced coding unit size.
The feature of the current coded unit block includes encoding block marker x in one of the embodiments,CBF_Meg(i)、 Rate distortion costs value xRD_Meg(i), it is distorted xD_Meg(i) and number of coded bits xBit_Meg(i);Wherein, i is current coded unit Depth;
The calculation formula of the motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicates fortune Dynamic and vertical movement amplitude, i are the depth of current coded unit.
Using the time of video macroblock mode, the rate distortion costs in space and coding unit depth as current coded unit The characteristic quantity of depth decision, is denoted as x respectivelyNB_RD(i) and xCU_depth(i);Wherein, rate distortion costs xNB_RDIt (i) is the adjacent left side With the evaluation of estimate of the rate distortion costs value of top coding unit;Coding unit depth xCU_depthIt (i) is the flat of adjacent encoder unit Equal depth;
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABVIt (i) is a left side 4 × 4 pieces of numbers in side and top coding unit.
The feature vector by extraction and forced coding unit size input the output of three values in one of the embodiments, Learning machine, the step of configuring learning parameter and mode of learning, establishing learning model includes:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value Oi,+1 or -1, wherein i indicates the label of learning machine, is 1 to m;
To m output OiIt is merged, obtaining final output is QALL
Wherein, TAAnd TBFor 0 to m two threshold values.
It in one of the embodiments, further include that feature is inputted to classifier, the classifier will be according in current video Hold characteristic and the model parameter learned is made prediction A (+1), B (- 1) or C (U is not known);
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1
The incrementss of rate distortion costs caused by misprediction can indicate in one of the embodiments, are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALLIt (i) is to divide (B) and do not divide (A) Misprediction rate, NBA,1(i) and NAB,2It (i) is respectively B misprediction is A in No.1 classifier number of samples, A in No. 2 classifiers Misprediction is the number of samples of B, NALL(i) number of samples predicted for i-th layer of coding unit, wherein NALL(1) i.e. For the coding unit number of whole image;
Divide and is calculated as with the computation complexity that do not divide after optimizing
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to pass through segmentation and ameristic prediction in coding unit depth layer i Caused by computation complexity reduce percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i), Wherein TS(i)、TnS(i) and TALL(i) it is respectively segmentation, does not divide and the computation complexity of origin operation;
The calculating after optimization is not divided according to the incrementss of rate distortion costs caused by the misprediction and the segmentation and Complexity sets objective function;
The objective function is expressed as
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two models ginseng of i-th layer of three value output category device Number, is expressed asWA(j, i) and WBIt is single that (j, i) respectively indicates i-th layer of coding In member in j-th of classifier positive and negative samples weighted value.
It in one of the embodiments, further include to configuration learning parameter and mode of learning setting Optimal Learning parameter;
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、 ΔηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti,
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basis
Obtain WA(j,i)/WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、Δ ηS→nS(i)、ΔηnS→S(i)、bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters.
The training method of the learning machine includes online mode and offline mode in one of the embodiments,.
The online mode includes to n frame video using original HM model based coding, output face in one of the embodiments, To the characteristic vector X of the learning machine of every a kind of coding unit depth iiAnd the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
The coding unit depth prediction trained support vector machines learning machine being used in coding video frames.
The offline mode includes choose several particular sequences and each sequence several in one of the embodiments, Frame is encoded using original HM model, exports the study towards every a kind of coding unit depth i in these encoded video frames The characteristic vector X of machineiAnd the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
It is pre- that trained support vector machines learning machine is used for video sequence, the coding of video frame and coding unit depth It surveys.
The above-mentioned efficient video coding method based on study establishes morning by setting the classifier of different niceties of grading Final only policy construction.The early-abort strategy structure can be converted by adjusting the ratio of output becomes the several classics of tradition Decision structure.Therefore, it can be switched over according to actual needs with the characteristic of classifier.It is preparatory for the feature vector of extraction The learning machine model succeeded in school can correspond to output predicted value, and different coding modes is selected according to predicted value.It can basis Rate distortion costs and computation complexity in efficient video coding, which correspond to, exports optimal cataloged procedure, to improve classifier Study and classification performance, and then improve Video coding code efficiency.
Detailed description of the invention
Fig. 1 is the flow chart of the efficient video coding method based on study;
Fig. 2 is the Fractionation regimen figure of coding unit in coding tree unit;
Fig. 3 is coding tree unit cataloged procedure schematic diagram;
Fig. 4 is that coding unit decision divides schematic diagram;
Fig. 5 (a) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (b) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (c) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 5 (d) is recursive coding unit size decision process Pn(i) one of structure chart;
Fig. 6 is three output category device flow charts;
Fig. 7 is Δ ηT,iAverage computation complexity under configuring condition reduces and the increased curve graph of code rate;
Fig. 8 is online mode training schematic diagram;
Fig. 9 is offline mode training schematic diagram.
Specific embodiment
As shown in Figure 1, being the flow chart of the efficient video coding method based on study.
A kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature of each coding unit block to Amount.
Described eigenvector include feature, motion information, contextual information, quantization parameter of current coded unit block etc. with And forced coding unit size.
The feature of current coded unit block includes encoding block marker xCBF_Meg(i), rate distortion costs value xRD_Meg(i), it loses True xD_Meg(i) and number of coded bits xBit_Meg(i);Wherein, i is the depth of current coded unit.
The calculation formula of motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicate movement and Vertical movement amplitude, i are the depth of current coded unit.
Specifically, the information of current coded unit, main includes using SKIP mode and MERGE pattern-coding present encoding The output information of unit process, including encoding block marker (Coded Block Flag, identification code residual error coefficient number), rate Distortion cost value, distortion and number of coded bits, this four information are expressed as xCBF_Meg(i)、xRD_Meg(i)、xD_Meg(i)、 xBit_Meg(i), wherein i is the depth of current coded unit, is 0,1,2,3.In addition, after being coding there are one SKIP marker bit Output information position, 0 or 1, be denoted as xSKIP(i)。
Motion information is mainly used to characterize the motion intense degree of current coded unit, bigger general of general motion intense Rate uses the coding unit of small size, that is, the probability divided is higher.Using the movement arrow after merge pattern-coding in the present embodiment Scale levies the movement of current coded unit, is specifically calculated as xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy are respectively indicated Movement and vertical movement amplitude.
When usually having stronger due to the macroblock mode of video, spatial coherence, as a result, when, space rate be distorted generation The characteristic quantity of valence and coding unit depth as current coded unit depth decision, is denoted as x respectivelyNB_RD(i) and xCU_depth (i).In the present embodiment, xNB_RDIt (i) is the evaluation of estimate on the adjacent left side and the rate distortion costs value of top coding unit;xCU_depth (i) it is the mean depth of adjacent encoder unit, is calculated by the following formula to obtain
Wherein, djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABV(i) it is 4 × 4 pieces of numbers in the left side and top coding unit.
The quantization parameter for encoding current coded unit, is denoted as xQP, usual quantization parameter is bigger, and current coded unit more has can It can be encoded using biggish piece.
The features above amount enumerated in the present embodiment is main feature amount, the including but not limited to above content, furthermore It can be to wherein deleting.
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration are learned Parameter and mode of learning are practised, learning model is established.
Incorporated by reference to Fig. 6.Specifically, three value output category devices are made of the classifier that a two-value of m (m >=2) exports, structure As shown in figure 5, these two-value output category devices can be well known support vector machines, neural network learning machine or Bayes's classification Device etc..Finally, merging different output results by combining unit, form the output of three values, be positive respectively (+1), it is negative (- 1) and it is uncertain (U), A, B and C in Fig. 5 (d) are corresponded respectively to, amalgamation mode is embodied as.
Wherein TAAnd TBFor 0 to m two threshold values, work as TA=m, then the A in Fig. 5 (d) is 0, works as TB=0 three output point The B of class device is 0, TA=m while TB=0, then exporting C is 100%.If TA<TB, then C output is 0.In practical operation, general TA> TB
The classifier of one three value output can be made of m different learning machines, can also be by the same learning machine but ginseng Number composition different with mode of learning, or different parameters form different learning machines simultaneously.In the present embodiment, using multiple supports Vector machine, different weighting coefficients, m=2, TA=2, TB=1.
Support vector machines learning machine is used in the present embodiment, training process can be divided into two major classes, offline mode and online Mode.
As shown in figure 8, for online mode training schematic diagram.
In online mode, by using original HM model based coding to n frame video, export towards each class coding unit depth The characteristic vector X of the learning machine of iiThe best macroblock mode Y of (including 9 above-mentioned characteristic quantities) and each coding unit, by Xi With Y input support vector machines learning machine training, trained support vector machines learning machine is then used for subsequent video frame coding In coding unit depth prediction;One sequence of every coding requires re -training in which.
As shown in figure 9, for offline mode training schematic diagram.
In offline mode, several frames of several particular sequences and each sequence are chosen, are encoded using original HM model, Export the characteristic vector X of the learning machine towards every a kind of coding unit depth i in these encoded video framesi(include above-mentioned 9 A characteristic quantity) and each coding unit best macroblock mode Y, by XiWith Y input support vector machines learning machine training.Then Trained support vector machines learning machine is used for other video sequences, the coding of video frame and coding unit depth prediction.It should Mode can be trained the selection of video as required and training updates.
Early-abort strategy is added in step 130, the selection course that coding unit size is carried out in efficient video coding device Structure, wherein in each coding unit depth layer i, SKIP mode and MERGE mode current block, extraction and step is first carried out The corresponding feature vector of 110 corresponding present encodings.
In efficient video coding device HEVC Video coding, each image is made of the CTU of some column, coding therein Final coding unit segmentation is obtained in tree unit and ameristic decision is not an individual two-value decision problem, Er Qieyou Multiple decision problem compositions.In efficient video coding device HEVC coding, the coding tree unit luma unit of each 64x64 will be first The coding unit size coding for first using 64x64, calculates rate distortion costs;Then, it is divided into the coding list of 4 32x32 Elemental size calculates separately the rate distortion costs of 4 units.In the process, the coding unit size of each 32x32 can be divided again For the size coding unit of 4 16x16, successively recurrence, until the coding unit unit of 8x8.As shown in figure 3, wherein DnIt indicates Coding current coded unit simultaneously calculates rate distortion costs, and n 0,1,2,3 respectively corresponds coding size 64x64 to 8x8 size, Pn It (i) is then recursive coding unit size decision process, i is the call number of four sub-blocks.Finally, from the coding unit of small size Recurrence upwards in turn, the rate distortion costs compared with upper one layer, cost it is small be set to better model selection;Successively recurrence, simultaneously Compare to the end, thus obtains the optimum code unit size partitioning scheme of coding tree unit.
In coding unit size selection process, current coded unit size or smaller 4 coding unit rulers are selected It is very little can be defined as one point with regardless of the problem of.The coding unit size decision problem of entire coding tree unit can be with as a result, It is described as the two-value decision problem of 3 levels.As shown in figure 4, the selection of 64x64 or 4 32x32 can be defined as one Point with regardless of two-value decision problem, need a classifier, be defined as L1 first layer;Then, for second, there will be 4 Point with regardless of the problem of, thus need 4 classifiers (L2);Similar, third layer needs 16 classifiers.These classifiers, Same layer belongs to the classifier of the same attribute, can be used for multiple times with a classifier or multiple identical classifiers.
Due to the P in Fig. 3n(i) structure is a recursive structure, is repeated in different levels, thus for convenience's sake, We use Pn(i) optimization of structure describes the optimization of entire coding unit size decision process.
For the P in Fig. 3n(i), in the Knowledge Verification Model of efficient video coding device HEVC (i.e. original video encoder) It can be described as by a DnWith 4 Pn+1(i) the advantages of composition, process are that structure is executed such as the sequence of Fig. 5 (a), the structure It is that selection optimal mode code efficiency is high, the disadvantage is that there is a large amount of non-essential calculating, computation complexity is high.During this, Since optimal mode can select DnOr 4 Pn+1, it is the process of alternative.
Thus multiple pre- geodesic structures, Fig. 5 (b) are one of premature termination scheme, are finishing DnLater, pre- using one Classifier is surveyed, predicts current best mode and judge whether to skip to execute subsequent Pn+1, thus reduce computation complexity.However This method will execute D to all blocksnOperation, for selecting Pn+1It is obviously wasted for optimal mode unnecessary.
It is proposed shown in premature termination scheme such as Fig. 5 (c) as a result, i.e., as execution DnAnd Pn+1It is preceding to be predicted first by classifier Current optimal mode, if it is Dn, then it is set as Y and only carries out Dn, otherwise it is set as N and executes Pn+1.The advantages of structure is only Predict accurate, not additional calculating cost;The disadvantage is that it is highly dependent on the precision of prediction of classifier, once prediction is not quasi- enough Really, it will biggish compression efficiency is caused to decline.
It is proposed to this end that as shown in Fig. 5 (d), classifier will make prediction A, B or C according to current video content characteristic, right In A, i.e. prediction DnFor optimal mode, then D is only carried outn;It then predicts that Pn+1 is optimal mode for B, then only carries out Pn+1;For C Then indicate uncertain DnOr Pn+1, then D is executednWith 4 Pn+1.The structure is very flexible, has multiple advantages: due to that can pass through Tri- output items of ABC in sorting algorithm and parameter adjustment classifier, as a result, when it is 100% that A and B, which is 0, C, the structure It can be exchanged into Fig. 5 (a), there is highest compression efficiency maximum complexity.When B is 0, then the structure becomes Fig. 5 (b);Finally, When C is 0, then the structure is converted to structure shown in Fig. 5 (c).The structure can realize code efficiency and meter by adjusting A, B, C The conversion of complexity is calculated, code efficiency and encoder complexity, can be according to reality between structure chart 5 (a) and Fig. 5 (c) The adjustment of application system demand.
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if Predicted value is not divide, then executes and test current coded unit size, at the same skip the test of partition encoding unit size with Coding;If predicted value be segmentation, skip test current coded unit size, directly execute segmentation coding size test and Coding;If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested.
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit has all encoded At.
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames.
Efficient video coding method based on study further includes by the rate distortion costs of the time of video macroblock mode, space And characteristic quantity of the coding unit depth as current coded unit depth decision, it is denoted as x respectivelyNB_RD(i) and xCU_depth(i); Wherein, rate distortion costs xNB_RDIt (i) is the evaluation of estimate on the adjacent left side and the rate distortion costs value of top coding unit;Coding unit Depth xCU_depthIt (i) is the mean depth of adjacent encoder unit.
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABVIt (i) is a left side 4 × 4 pieces of numbers in side and top coding unit.
The feature vector by extraction and forced coding unit size input the learning machine of three values output, configuration study ginseng Several and mode of learning, the step of establishing learning model include:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value Oi,+1 or -1, wherein i indicates the label of learning machine, is 1 to m.
To m output OiIt is merged, obtaining final output is QALL
Wherein, TAAnd TBFor 0 to m two threshold values.
Learning machine is support vector machines, but weighting parameters are different.Preferably, m is set as 2, TA=2, TB=1.
Efficient video coding method based on study further includes that feature is inputted to classifier, and the classifier will be according to current Video content characteristic and the model parameter learned make prediction A (+1), B (- 1) or C (U is uncertain).
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1
Efficient video coding method based on study further includes to configuration learning parameter and mode of learning setting Optimal Learning Parameter.
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、 ΔηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basis
Obtain WA(j,i)/WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、Δ ηS→nS(i)、ΔηnS→S(i)、bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters.
Specifically, needing in the support vector machines learning machine training process of three values output for two support vector machines Habit machine configures reasonable parameter, has reached optimal prediction effect.Mainly by adjusting support vector machines study in the present embodiment Weighting coefficient W in machineAAnd WB, carry out adjusting training process.WAAnd WBThe importance of positive and negative samples in sample is respectively indicated, it is bigger It indicates easier and is divided into such, false acceptance rate will increase, and false rejection rate will be reduced.The knot of these mistake classification or prediction Fruit will lead to the coding of the code efficiency and computation complexity that finally encode.
Define Δ ηS→nS(i) do not divide those coding units for selecting Fractionation regimen as optimal mode of pattern-coding to use The incrementss of caused rate distortion costs are represented by Δ ηS→nS(i)=(1-JnS(i)/JBest× 100%, (i)) wherein i For coding unit decision-making level, i ∈ { 1,2,3 }, JnS(i) and JBest(i) it respectively indicates and is compiled using not Fractionation regimen and optimal mode Rate distortion costs when code current coded unit.Equally, Δ η is definednS→S(i) for using Fractionation regimen encode those selection regardless of The incrementss for cutting rate distortion costs caused by the coding unit that mode is optimal mode, are represented by Δ ηnS→S(i)=(1-JS (i)/JBest(i)) × 100%, wherein JSIt (i) is the rate distortion costs that current coded unit is encoded using Fractionation regimen.As a result, In the decision process of each layer of coding unit, the incrementss of the rate distortion costs as caused by misprediction can be indicated are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALLIt (i) is to divide (B) and do not divide (A) Misprediction rate, NBA,1(i) and NAB,2It (i) is respectively B misprediction is A in No.1 classifier number of samples, A in No. 2 classifiers Misprediction is the number of samples of B, NALL(i) number of samples predicted for i-th layer of coding unit, wherein NALL(1) i.e. For the coding unit number of whole image.
Define qS,1(i)=NS,1(i)/NALLIt (i) is the percentage that segmentation is predicted as by the classifier that three values export, definition qnS,2(i)=NnS,2(i)/NALL(i) to be predicted as ameristic percentage, wherein NS,1(i) and NnS,2It (i) is respectively that three values are defeated The segmentation and ameristic number of samples that classifier 1 and 2 is predicted respectively in classifier out.As a result, by segmentation with do not divide it is excellent Computation complexity after change may be calculated
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to pass through segmentation and ameristic prediction in coding unit depth layer i Caused by computation complexity reduce percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i), Wherein TS(i)、TnS(i) and TALLIt (i) is respectively that segmentation (skips DnOperation), do not divide and (omits 4 Pn+1Operation), and former behaviour Make (DnWith 4 Pn+1Be carried out) computation complexity.
It is encoded and is counted for 20 frames of BQMall, FourPeople etc. 5 different cycle tests, by above-mentioned parameter It is fitted to
Wherein bi、ai、ti、BiAnd AiFor fitting parameter;
Wherein ui、viAnd TiFor fitting parameter, these fitting parameters by with the cycle tests of selection, coding frame number etc. no Together, it is varied.In addition, Δ TnS(i)、ΔTS(i)、ΔηS→nS(i) and Δ ηnS→SIt (i) can also be by partial test sequence The coding of column and count to obtain.
In the present embodiment, above-mentioned fitting parameter is as follows:
Wherein R2Indicate fitting precision, it is better closer to 1.
Thus the performance optimized in order to obtain, function of setting objectives, i.e., under the conditions of rate distortion costs increase limited, The computation complexity 1- Δ T (i) for minimizing encoder, is represented by
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two models ginseng of i-th layer of three value output category device Number, is expressed asWA(j, i) and WBIt is single that (j, i) respectively indicates i-th layer of coding In member in j-th of classifier positive and negative samples weighted value.Above problem conversion are as follows:
To above-mentioned formula respectively to xi、yiAnd λiLocal derviation is sought, and is set to 0, is obtained
Above-mentioned formula is solved, available:
In the present embodiment, only λiWith Δ ηT,iFor parameter, other coefficients are known constant, although can not be dominant It indicates, but as long as giving a Δ ηT,iCorresponding λ can be obtained by the methods of known least square methodi, then by λiIt brings into X can be calculated in formulaiAnd yi, finally obtain WA(j,i)/WBThe ratio of (j, i), the instruction as support vector machines learning machine Practice parameter.
Based on above-mentioned all embodiments, it is based on using the reference software platform HM12.0 verifying of efficient video coding device is above-mentioned The efficient video coding method of study.Configuration information includes low delay B frame class, and coded sequence first frame is I frame, remaining frame is P frame, coding unit size support 64 × 64 to 8 × 8, and motion estimation range 64, other parameters are default parameters.Coding experiments Implement on computers.
Encoding verification experiment is divided into two parts, encodes 5 cycle tests Basketballpass (416 × 240) first, Partyscene (832 × 480), Johnny (1280 × 720), Kimono (1920 × 1080), Traffic (2560 × 1600), using different user configuration parameter, Δ ηT,i, it can be expressed as { 0.1%, 0.1%, 0.1% } respectively, 0.3%, 0.3%, 0.3% }, { 0.5%, 0.5%, 0.5% }, { 0.7%, 0.7%, 0.7% } is denoted as Para_111, Para_333, Para_555 and Para_777.In addition, { 0.3%, 0.2%, 0.1% }, { 0.6%, 0.4%, 0.2% }, 0.9%, 0.6%, 0.3% }, it is denoted as Para_321, Para_642, Para_963, and thus obtains training parameter WAAnd WBTraining learning machine, then Learning machine is used for coding unit depth prediction in cataloged procedure.
It is illustrated in figure 7 the volume that the encoder encoded video before 5 videos and optimization is encoded using the encoder after optimization Code efficiency and computation complexity compare, the saving degree of Y-axis position computation complexity, compared to the code rate before optimization after the what happened latter of X-axis position Increase percentage, as seen from the figure, in different configuration parameter Δ ηT,iAdjusting under, average computation complexity can reduce by 42% ~56%.
In addition, using parameter Para_642 training learning machine, and the model that training obtains is used for the CU depth in coding Prediction optimization complexity encodes all frames of 21 sequences in this experiment, while compared existing state-of-the-art three kinds of codings Method ShenEVIP, ShenTMM and XiongTMM compare, and the present invention is soft compared to former efficient video coding device HEVC Knowledge Verification Model Part platform HM can reduce computation complexity 28.82% to 70.93%, and average 51.45%, average BDPSNR and BDBR is respectively- 0.061dB and 1.98%, it is almost the same with the compression efficiency of former HM.Due to currently advanced from compression efficiency and computation complexity Three schemes ShenEVIP, ShenTMM and XiongTMM.
It is that code efficiency and computation complexity compare such as following table:
It is above-mentioned it is all it is used in the examples by two Support vector machines form three values output learning machines, Learning machine therein could alternatively be other kinds of learning machine, such as Bayes, neural network, decision tree, while learning machine Quantity can be with more than two.Furthermore the learning machine of three values output both can export learning machine by multiple two-values and form, can also be by Classification learning machine is directly realized by more than one.
The feature vector of input learning machine includes, but are not limited to four classes mentioned in above-described embodiment, may also include figure As texture edge, brightness etc.;Meanwhile the characteristic quantity in the present invention can have many forms, such as current coded unit Feature, the motion information of block are calculated as xMV_Meg(i)=| MVx |+| MVy |, contextual information, quantization parameter etc. and optimal encoding Code unit size etc. can actually be replaced using other forms of expression.
Classification method based on study is used for the decision of coding unit depth, complicated to reduce the model selection in coding Degree.Process with a variety of " multiselects one " selected similar to size/depth as coding in actual video cataloged procedure, is removed Outside coding unit depth decision, there are also predicting unit model selection predicting unit, converter unit model selection converter unit join more Frame selection is examined, the classification method in above-described embodiment based on study can be used in the processes such as estimation, solves related " multiselect One " the problem of.
The above-mentioned efficient video coding method based on study establishes morning by setting the classifier of different niceties of grading Final only policy construction.The early-abort strategy structure can be converted by adjusting the ratio of output becomes the several classics of tradition Decision structure.Therefore, it can be switched over according to actual needs with the characteristic of classifier.It is preparatory for the feature vector of extraction The learning machine model succeeded in school can correspond to output predicted value, and different coding modes is selected according to predicted value.It can basis Rate distortion costs and computation complexity in efficient video coding, which correspond to, exports optimal cataloged procedure, to improve classifier Study and classification performance, and then improve Video coding code efficiency.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited In contradiction, all should be considered as described in this specification.
The embodiments described above only express several embodiments of the present invention, and the description thereof is more specific and detailed, but simultaneously It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art It says, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to protection of the invention Range.Therefore, the scope of protection of the patent of the invention shall be subject to the appended claims.

Claims (9)

1. a kind of efficient video coding method based on study, comprising the following steps:
Step 110, using efficient video coding device encoded video sequence, extract the corresponding feature vector of each coding unit block;
Step 120, the learning machine that the feature vector of extraction and forced coding unit size are inputted to the output of three values, configuration study ginseng Several and mode of learning, establishes learning model;
Early-abort strategy structure is added in step 130, the selection course that coding unit size is carried out in efficient video coding device, Wherein, in each coding unit depth layer i, SKIP mode and MERGE mode is first carried out in current coded unit block, is extracted and is walked The corresponding feature vector of rapid 110 corresponding current coded unit block;
Feature vector in step 130 is inputted the learning machine model succeeded in school by step 140, exports predicted value, if prediction Value is not divide, then executes and test current coded unit size, while skipping test and the coding of partition encoding unit size; If predicted value is segmentation, skip test current coded unit size directly executes test and the coding of the coding size of segmentation; If uncertain, then current coded unit size is tested, the coding unit size of segmentation is then tested;
Step 150 repeats step 130 and step 140 until coding unit layer all in coding tree unit all encodes completion;
Step 160 repeats step 130- step 150 until coding tree unit all encodes completion in all video frames;
The feature vector of extraction and forced coding unit size, are inputted the learning machine of three values output by the step 120, and configuration is learned The step of practising parameter and mode of learning, establishing learning model include:
Characteristic vector is inputted to the learning machine of m two-value output, learning machine passes through the model learnt, exports predicted value Oi,+1 Or -1, wherein i indicates the label of learning machine, is 1 to m;
To m output OiIt is merged, obtaining final output is QALL
Wherein, TAAnd TBFor 0 to m two threshold values.
2. the efficient video coding method according to claim 1 based on study, which is characterized in that described eigenvector packet Include feature, motion information, contextual information, quantization parameter of current coded unit block etc. and forced coding unit size.
3. the efficient video coding method according to claim 2 based on study, which is characterized in that the present encoding list The feature of first block includes encoding block marker xCBF_Meg(i), rate distortion costs value xRD_Meg(i), it is distorted xD_Meg(i) it and encodes Bit number xBit_Meg(i);Wherein, i is the depth of current coded unit;
The calculation formula of the motion information is xMV_Meg(i)=| MVx |+| MVy |, wherein MVx and MVy respectively indicate movement and Vertical movement amplitude, i are the depth of current coded unit;
Using the time of video macroblock mode, the rate distortion costs in space and coding unit depth as current coded unit depth The characteristic quantity of decision, is denoted as x respectivelyNB_RD(i) and xCU_depth(i);Wherein, rate distortion costs xNB_RDIt (i) is the adjacent left side and upper The evaluation of estimate of the rate distortion costs value of side coding unit;Coding unit depth xCU_depthIt (i) is the average depth of adjacent encoder unit Degree;
It is calculated by the following formula to obtain:
Wherein djIt is in the left side and top coding unit with 4 × 4 pieces of depth values for unit, NLFT(i) and NABV(i) for the left side and 4 × 4 pieces of numbers in the coding unit of top.
4. the efficient video coding method according to claim 1 based on study, which is characterized in that further include that feature is defeated Enter classifier, the classifier will make prediction A, B or C according to current video content characteristic and the model parameter learned;Wherein A It indicates just, B indicates negative, and C indicates uncertain;
When being predicted as A, then by Direct Model DnAs optimal mode, then D is only carried outn
When being predicted as B, then by fault-tolerant mode Pn+1As for optimal mode, then P is only carried outn+1
When being predicted as C, then it represents that uncertain DnOr Pn+1, then D is executednWith 4 Pn+1
5. the efficient video coding method according to claim 4 based on study, which is characterized in that rate caused by misprediction The incrementss of distortion cost can indicate are as follows:
ΔηRD(i)=Δ ηnS→S(i)×pBA(i)+ΔηS→nS(i)×pAB(i);
Wherein Δ ηnS→S(i) for use Fractionation regimen coding select not Fractionation regimen for caused by the coding unit of optimal mode The incrementss of rate distortion costs, Δ ηS→nS(i) do not divide the coding that pattern-coding selects Fractionation regimen as optimal mode to use The incrementss of rate distortion costs caused by unit, pBA(i)=NBA,1(i)/NALL(i), pAB(i)=NAB,2(i)/NALL(i) it is Segmentation B and the misprediction rate for not dividing A, NBA,1(i) and NAB,2(i) be respectively in No.1 classifier B misprediction be A sample Number, A misprediction is the number of samples of B, N in No. 2 classifiersALL(i) sample predicted for i-th layer of coding unit Number, wherein NALL(1) be whole image coding unit number;
Divide and is calculated as with the computation complexity that do not divide after optimizing
Δ T (i)=Δ TS(i)×qS,1(i)+ΔTnS(i)×qnS,2(i);
Wherein Δ TS(i) and Δ TnSIt (i) is respectively to be led in coding unit depth layer i by segmentation and ameristic prediction The computation complexity of cause reduces percentage, Δ TS(i)=1-TS(i)/TALL(i), Δ TnS(i)=1-TnS(i)/TALL(i), wherein TS(i)、TnS(i) and TALL(i) it is respectively segmentation, does not divide and the computation complexity of origin operation;
The calculating complexity after optimization is not divided according to the incrementss of rate distortion costs caused by the misprediction and the segmentation and Degree setting objective function;
The objective function is expressed as
Wherein, Δ ηT,iPercentage, x are reduced for compression efficiencyiAnd yiIt is two model parameters of i-th layer of three value output category device, It is expressed asWA(j, i) and WB(j, i) is respectively indicated in i-th layer of coding unit The weighted value of positive and negative samples in j-th of classifier.
6. the efficient video coding method according to claim 4 based on study, which is characterized in that further include learning configuration It practises parameter and Optimal Learning parameter is arranged in mode of learning;
To the partial video frame in several cycle tests, count to obtain Δ T using efficient video codingnS(i)、ΔTS(i)、Δ ηS→nS(i) and Δ ηnS→S(i), and fitting obtains parameter bi、ai、ti、Bi、Ai、ui、viAnd Ti,
Given Δ ηT,iLeast square method can be used and obtain parameter lambdai, then by λiBring formula into
X can be calculatediAnd yi, final basisObtain WA(j,i)/ WBThe ratio of (j, i), the training parameter as learning machine, wherein Δ TnS(i)、ΔTS(i)、ΔηS→nS(i)、ΔηnS→S(i)、 bi、ai、ti、Bi、Ai、ui、vi、Ti、xiAnd yiIt is intermediate parameters, h1(i), h2(i), k1(i), k2It (i) is known constant.
7. the efficient video coding method according to claim 1 based on study, which is characterized in that the instruction of the learning machine The mode of white silk includes online mode and offline mode.
8. the efficient video coding method according to claim 7 based on study, which is characterized in that the online mode packet It includes to n frame video using original HM model based coding, exports the characteristic vector X of the learning machine towards each class coding unit depth ii And the best macroblock mode Y of each coding unit;
By XiWith Y input support vector machines learning machine training;
The coding unit depth prediction trained support vector machines learning machine being used in coding video frames.
9. the efficient video coding method according to claim 7 based on study, which is characterized in that the offline mode packet Several frames for choosing several particular sequences and each sequence are included, is encoded using original HM model, exports these encoded views The characteristic vector X of learning machine towards every a kind of coding unit depth i in frequency frameiAnd the best piecemeal mould of each coding unit Formula Y;
By XiWith Y input support vector machines learning machine training;
Trained support vector machines learning machine is used for video sequence, the coding of video frame and coding unit depth prediction.
CN201510137157.3A 2015-03-26 2015-03-26 Efficient video coding method based on study Active CN106162167B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510137157.3A CN106162167B (en) 2015-03-26 2015-03-26 Efficient video coding method based on study

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510137157.3A CN106162167B (en) 2015-03-26 2015-03-26 Efficient video coding method based on study

Publications (2)

Publication Number Publication Date
CN106162167A CN106162167A (en) 2016-11-23
CN106162167B true CN106162167B (en) 2019-05-17

Family

ID=57340278

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510137157.3A Active CN106162167B (en) 2015-03-26 2015-03-26 Efficient video coding method based on study

Country Status (1)

Country Link
CN (1) CN106162167B (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10382770B2 (en) * 2017-02-06 2019-08-13 Google Llc Multi-level machine learning-based early termination in partition search for video encoding
CN106713929B (en) * 2017-02-16 2019-06-28 清华大学深圳研究生院 A kind of video inter-prediction Enhancement Method based on deep neural network
CN108737841B (en) * 2017-04-21 2020-11-24 腾讯科技(深圳)有限公司 Coding unit depth determination method and device
CN107690069B (en) * 2017-08-28 2021-01-01 中国科学院深圳先进技术研究院 Data-driven cascade video coding method
US11436471B2 (en) * 2017-10-13 2022-09-06 Panasonic Intellectual Property Corporation Of America Prediction model sharing method and prediction model sharing system
KR102535361B1 (en) 2017-10-19 2023-05-24 삼성전자주식회사 Image encoder using machine learning and data processing method thereof
US20200344474A1 (en) * 2017-12-14 2020-10-29 Interdigital Vc Holdings, Inc. Deep learning based image partitioning for video compression
CN108012150B (en) * 2017-12-14 2020-05-05 湖南兴天电子科技有限公司 Video interframe coding method and device
CN108200442B (en) * 2018-01-23 2021-11-12 北京易智能科技有限公司 HEVC intra-frame coding unit dividing method based on neural network
CN108174208B (en) * 2018-02-12 2020-05-12 杭州电子科技大学 Efficient video coding method based on feature classification
CN108495129B (en) * 2018-03-22 2019-03-08 北京航空航天大学 The complexity optimized method and device of block partition encoding based on deep learning method
CN108924558B (en) * 2018-06-22 2021-10-22 电子科技大学 Video predictive coding method based on neural network
EP3743855A1 (en) * 2018-09-18 2020-12-02 Google LLC Receptive-field-conforming convolution models for video coding
CN109089114B (en) * 2018-09-28 2022-01-28 河海大学 Complexity-adaptive screen content coding method for virtual reality equipment
CN109769119B (en) * 2018-12-18 2021-01-19 中国科学院深圳先进技术研究院 Low-complexity video signal coding processing method
CN110072119B (en) * 2019-04-11 2020-04-10 西安交通大学 Content-aware video self-adaptive transmission method based on deep learning network
CN110401834B (en) * 2019-08-06 2021-07-27 杭州微帧信息科技有限公司 Self-adaptive video coding method based on deep learning
CN111083480B (en) * 2019-12-10 2022-11-04 四川新视创伟超高清科技有限公司 High-speed coding method for 8K ultra-high definition video based on CPU
CN113347415A (en) * 2020-03-02 2021-09-03 阿里巴巴集团控股有限公司 Coding mode determining method and device
CN111988628B (en) * 2020-09-08 2023-10-17 福州大学 VVC rapid intra-frame coding method based on reinforcement learning
CN114584771B (en) * 2022-05-06 2022-09-06 宁波康达凯能医疗科技有限公司 Method and system for dividing intra-frame image coding unit based on content self-adaption

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101217663A (en) * 2008-01-09 2008-07-09 上海华平信息技术股份有限公司 A quick selecting method of the encode mode of image pixel block for the encoder
CN102387356A (en) * 2010-09-06 2012-03-21 索尼公司 Image processing apparatus and method
CN103517069A (en) * 2013-09-25 2014-01-15 北京航空航天大学 HEVC intra-frame prediction quick mode selection method based on texture analysis
WO2015034061A1 (en) * 2013-09-06 2015-03-12 三菱電機株式会社 Video encoding device, video transcoding device, video encoding method, video transcoding method and video stream transmission system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101217663A (en) * 2008-01-09 2008-07-09 上海华平信息技术股份有限公司 A quick selecting method of the encode mode of image pixel block for the encoder
CN102387356A (en) * 2010-09-06 2012-03-21 索尼公司 Image processing apparatus and method
WO2015034061A1 (en) * 2013-09-06 2015-03-12 三菱電機株式会社 Video encoding device, video transcoding device, video encoding method, video transcoding method and video stream transmission system
CN103517069A (en) * 2013-09-25 2014-01-15 北京航空航天大学 HEVC intra-frame prediction quick mode selection method based on texture analysis

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Fast transrating for high efficiency video coding based on machine learning;Yun Zhang,Sam Kwong,Xu Wang,Hui Yuan,Zhaoqing Pan;《2013 IEEE International Conference on Image Processing》;20140213;第26卷(第3期);全文 *

Also Published As

Publication number Publication date
CN106162167A (en) 2016-11-23

Similar Documents

Publication Publication Date Title
CN106162167B (en) Efficient video coding method based on study
CN105306947B (en) video transcoding method based on machine learning
CN103873861B (en) Coding mode selection method for HEVC (high efficiency video coding)
Jin et al. CNN oriented fast QTBT partition algorithm for JVET intra coding
CN106454342B (en) A kind of the inter-frame mode fast selecting method and system of video compression coding
CN106937118B (en) A kind of bit rate control method combined based on subjective area-of-interest and time-space domain
CN106131546B (en) A method of determining that HEVC merges and skip coding mode in advance
CN111654698B (en) Fast CU partition decision method for H.266/VVC
CN108737841A (en) Coding unit depth determination method and device
CN107371022A (en) The quick division methods of interframe encode unit applied to HEVC medical image lossless codings
CN111355956A (en) Rate distortion optimization fast decision making system and method based on deep learning in HEVC intra-frame coding
CN105959611A (en) Adaptive H264-to-HEVC (High Efficiency Video Coding) inter-frame fast transcoding method and apparatus
CN109040764A (en) Fast coding algorithm in a kind of HEVC screen content frame based on decision tree
CN110049338B (en) HEVC (high efficiency video coding) rapid inter-frame coding method based on multi-level classification
CN107852492A (en) Method and apparatus for being coded and decoded to image
Chen et al. A novel fast intra mode decision for versatile video coding
CN108989799A (en) A kind of selection method, device and the electronic equipment of coding unit reference frame
CN103313058B (en) The HEVC Video coding multimode optimization method realized for chip and system
CN116489386A (en) VVC inter-frame rapid coding method based on reference block
CN106937116A (en) Low-complexity video coding method based on random training set adaptive learning
CN107690069B (en) Data-driven cascade video coding method
CN108200431A (en) A kind of video frequency coding rate controls frame-layer Bit distribution method
CN103888770B (en) A kind of video code conversion system efficiently and adaptively based on data mining
CN109361920A (en) A kind of interframe quick predict algorithm of the adaptive decision-making tree selection towards more scenes
Huang et al. Modeling acceleration properties for flexible INTRA HEVC complexity control

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant