CN107197260A - Video coding post-filter method based on convolutional neural networks - Google Patents

Video coding post-filter method based on convolutional neural networks Download PDF

Info

Publication number
CN107197260A
CN107197260A CN201710439132.8A CN201710439132A CN107197260A CN 107197260 A CN107197260 A CN 107197260A CN 201710439132 A CN201710439132 A CN 201710439132A CN 107197260 A CN107197260 A CN 107197260A
Authority
CN
China
Prior art keywords
frame
video
neural networks
convolutional neural
post
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710439132.8A
Other languages
Chinese (zh)
Other versions
CN107197260B (en
Inventor
张永兵
林荣群
王兴政
王好谦
戴琼海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Weilai Media Technology Research Institute
Shenzhen Graduate School Tsinghua University
Original Assignee
Shenzhen Weilai Media Technology Research Institute
Shenzhen Graduate School Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Weilai Media Technology Research Institute, Shenzhen Graduate School Tsinghua University filed Critical Shenzhen Weilai Media Technology Research Institute
Priority to CN201710439132.8A priority Critical patent/CN107197260B/en
Publication of CN107197260A publication Critical patent/CN107197260A/en
Application granted granted Critical
Publication of CN107197260B publication Critical patent/CN107197260B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression

Abstract

Video coding post-filter method based on convolutional neural networks, including convolutional neural networks model training step and filter step, training step include:Set the quantization parameter of video compress to carry out coding compression for 20 to 51 pairs of original videos, obtain compressing video;All videos are carried out with frame and extracts the frame pair for obtaining multiple compressed video frame original video frames;It will extract and obtain frame to being multiple groups by the different demarcation of frame type and quantization parameter;Convolutional neural networks framework and initialization network parameter are built, neutral net is trained respectively using the group of foregoing division, obtains corresponding to different quantization parameters and multiple neural network models of frame type.Filter step includes:Obtained multiple neural network models are embedded in the post-filtering link of video encoder;The coding compression foregoing to pending original video execution and frame extract and obtain pending frame pair, and the quantization parameter and the corresponding neural network model of frame type selecting according to pending frame pair are filtered processing.

Description

Video coding post-filter method based on convolutional neural networks
Technical field
The present invention relates to computer vision and field of video encoding, and in particular to a kind of video based on convolutional neural networks Encode post-filter method.
Background technology
With the development of science and technology and enriching for various video display apparatus, video has been increasingly becoming can not in people's life The part lacked, highly important effect is played in every field.The past video resolution that witnessed in decades is with showing Show the great development of device screen, and the video council of ultrahigh resolution produces huge data volume, this can be produced to the network bandwidth Greatly bear.Accordingly, it would be desirable to which efficient Video coding and transmission technology ensure that user watches the experience of video, while to the greatest extent may be used The data volume of energy ground reduction video, is network bandwidth Reduction of Students' Study Load.In consideration of it, researcher is in decades, constantly research is efficiently regarded Frequency coding method.Video coding technique, mainly reduced by removing the redundancy in video the data volume of video with up to To effectively storing and transmitting multitude of video data, it is intended on the premise of keeping former video quality as far as possible, with lower code check Compress video.
However, current video encoding standard is mainly all based on the hybrid video coding technology of block, in this encoder block It is block-based to predict within the frame/frames, change and slightly quantify all cause the decline of video quality, especially in low code under frame In the case of rate.Therefore, the distortion reduced in Video coding turns into one of study hotspot of current video coding field.Although, Some algorithms are also taken in current video coding standards to reduce blocking effect and improve subjective quality, but its effect is also not Enough ideals, the video after processing still suffers from obvious blocking effect and edge blurry, and loss in detail is still than more serious.
The content of the invention
It is a primary object of the present invention to using strong capability of fitting of the convolutional neural networks to nonlinear transformation, propose a kind of Video coding post-filter method based on convolutional neural networks, this method, which is set up, damages frame of video reflecting to lossless video frame Penetrate, so that the inverse transformation of Distortion course in Video coding is approximately obtained, to reach the purpose for reducing distortion.
The present invention is as follows for the technical scheme provided up to above-mentioned purpose:
A kind of Video coding post-filter method based on convolutional neural networks, includes the training of convolutional neural networks model Step and post-filtering process step, wherein:
The training step includes S1 to S4:
S1, the quantization parameter of setting video compress are 20 to 51, carry out coding compression to original video, obtain compression and regard Frequently;
S2, the extraction that frame is carried out to the compression video and the original video, obtain multiple frames pair, each frame pair Include a compressed video frame and an original video frame;
S3, the frame for extracting step S2 are multiple groups to the different demarcation according to frame type and quantization parameter;
S4, convolutional neural networks framework and initialization network parameter are built, the group divided using step S3 is respectively to nerve Network is trained, and obtains corresponding to different quantization parameters and multiple neural network models of frame type;
The post-filtering process step includes S5 and S6:
S5, the post-filtering link that multiple neural network models that step S4 is obtained are embedded in video encoder;
S6, step S1 and S2 are performed to pending original video obtain pending frame pair, and according to pending frame pair Quantization parameter and the corresponding neural network model of frame type selecting are filtered processing.
The difference of quantization parameter and frame type causes their nature of distortion also different, and the present invention is in training and actual treatment During be all the frame pair for extracting original video and compression video composition, set up with this and damage frame of video and lossless video frame Mapping;And post-filtering in encoder is embedded into by the neural network model trained for different quantization parameters and frame type Link is filtered processing, takes into full account the influence of quantization parameter and frame type to distortion level, first using the method for the present invention Differentiate the quantization parameter and frame type of frame, reselection quantization parameter and all corresponding neural network model of frame type are carried out Filtering, so as to effectively suppress distortion.
Brief description of the drawings
Fig. 1 is the flow chart for the Video coding post-filter method based on convolutional neural networks that the present invention is provided;
Fig. 2 is convolutional neural networks frame diagram;
Fig. 3 is the training process schematic diagram of convolutional neural networks.
Embodiment
The invention will be further described with specific embodiment below in conjunction with the accompanying drawings.
The present invention is using the powerful nonlinear fitting ability of convolutional neural networks, with reference to the characteristic of Video coding, fully divides In analysis Video coding compression process the reason for cause distortion, it is proposed that based on convolutional neural networks and be relevant to quantization parameter and frame The Video coding post-filter method of type, sets up the mapping damaged between frame of video and lossless video frame, and training obtains difference Convolutional neural networks model under quantization parameter (abbreviation QP) and different frame type, for the post-filtering link in encoder Processing is filtered for the frame of different quantization parameters and frame type, can effectively suppress distortion.
In mixed video coding framework, frame type mainly has I frames, P frames and B frames, wherein:
I frames use intra prediction mode, also referred to as intracoded frame (intra picture), and I frames are typically every The start frame of one picture group (GOP), its information independent of front and rear frame, therefore shuffle can be used as in broadcasting Point of penetration.For intra prediction mode, in decoding end, it is the block and the block of top on the adjacent left side based on decoding, uses Algorithm carries out the prediction of current block, and then along with residual error data carrys out the raw information of reconstruction image, the residual error data is by solving Code value and predicted value are subtracted each other and obtained;In coding side, it is necessary to using identical infra-frame prediction, because predicted value and raw image data It is more similar, so the energy for the remaining data to be transmitted is low-down, so as to realize the compression of code stream.
P frames and B frames use inter-frame forecast mode, P frames are forward predicted frame (predictive-frame), with reference to above I frames or B frames remove temporal redundancy, realize Efficient Compression.B frames are bi-directional predicted interpolation coding frame, also known as (bi-directional interpolated prediction frame), it needs to refer to front and rear frame of video, utilizes two The frame on side removes redundancy with the similitude of B frames.Inter prediction is mainly based upon higher similarity between consecutive frame, utilizes The information of given frame rebuilds present frame, and commonly used approach is pre- to carry out interframe by way of motor-function evaluation Survey.The method i.e. estimation method of the macro block most matched is found, this method is to be directed to current macro, first finds most phase in former frame As macro block (i.e. match block), poor are asked to them, many null values are produced, carrying out the coding flow similar to frame in can with regard to that can save The bit number of sight.
In mixed video coding framework, the frame of video of different frame type, its coding method is different, just because such as This, its nature of distortion of the frame of video of different frame type is also different.Because I frames are relatively independent, therefore master the reason for its distortion To be quantizing process and dct transform (discrete cosine transform) process, quantizing process is substantially that many numerical value are mapped into a small number of values Process, relatively, de-quantization process recovers the process of many numerical value by a small number of values, and its recovery inevitably has number According to error, therefore frame of video occurs some by quantifying the high-frequency noise that loss is caused.Producing the major way of P frames and B frames is By way of inter prediction, the source of both frames prediction is from I frames, therefore the nature of distortion that I frames have in itself Also P frames and B frames can be broadcast to;In addition, during motor-function evaluation, there is also the factor for causing distortion.Such as, in frame Between the motion compensation link predicted, two neighboring piece of prediction block of present frame is possible to be not from the phase of same reference frame The prediction of adjacent two blocks, it may be possible to from the prediction of two blocks in different frame.That is, that two of adjacent block reference The edge of block does not just possess continuity in itself, and this also leads to certain blocking effect phenomenon, so as to cause distortion.
The general principle of Video coding is exactly the correlation that has using video sequence to compress redundancy, and realization is regarded The high efficiency of transmission of frequency sequence.The correlation that wherein video sequence has is broadly divided into room and time correlation, spatially Correlation has correlation between being mainly reflected in same frame of video neighbor pixel, and temporal correlation is mainly reflected in Adjacent frame has higher similarity on time.Therefore, in mixed video coding framework, the generation side of each frame of video Formula, is not the result of infra-frame prediction, is exactly the result of inter prediction.Its generation mechanism of different types of frame is also differed, meaning Needs frame type being also contemplated for into the video encoding quality enhancing algorithm based on convolutional neural networks, to obtain more preferably Effect.
Based on foregoing principle, it is rearmounted that the specific embodiment of the invention proposes the Video coding based on convolutional neural networks Filtering method, this method includes the training step and post-filtering process step of convolutional neural networks model, as shown in Figure 1:
The training step includes S1 to S4:
S1, the quantization parameter of setting video compress are 20 to 51, carry out coding compression to original video, obtain compression and regard Frequently;
S2, the extraction that frame is carried out to the compression video and the original video, obtain multiple frames pair, each frame pair Include a compressed video frame and an original video frame;
S3, the frame for extracting step S2 are multiple groups to the different demarcation according to frame type and quantization parameter;
S4, convolutional neural networks framework and initialization network parameter are built, the group divided using step S3 is respectively to nerve Network is trained, and obtains corresponding to different quantization parameters and multiple neural network models of frame type;
The post-filtering process step includes S5 and S6:
S5, the post-filtering link that multiple neural network models that step S4 is obtained are embedded in video encoder;
S6, step S1 and S2 are performed to pending original video obtain pending frame pair, and according to pending frame pair Quantization parameter and the corresponding neural network model of frame type selecting are filtered processing.
The method of the present invention is filtered place firstly the need of training neutral net, then with the neural network model trained Reason.Illustrate that the above-mentioned Video coding based on convolutional neural networks that the present invention is provided is rearmounted with a kind of specific embodiment below Filtering method.
The original video for training is chosen, coding compression is carried out, quantization parameter QP during compression is set to 20 to 51 (i.e. Continuous integer 20,21,22,23,24 ..., 50,51), obtain the compression video under different Q P.Preferably, can be using normal JM10 and/or HM12 video encoding standards software carries out coding compression to original video.To original video and corresponding pressure Contracting video carries out frame extraction, obtains many frames pair, and frame is to that can be expressed as " original video frame-compressed video frame ", i.e., each Frame is to including an original video frame and a corresponding compressed video frame.By obtained all frames to according to frame type and quantization parameter QP different demarcation is multiple groups, and specific partition process is, for example,:All frames are drawn to elder generation according to QP difference (32) Point, then to the frame under each QP to being divided into I frames, P frames and B frames according to frame type, so as to obtain described multiple groups (in this example In, obtain), the frame in each group is to identical frame type and quantization parameter QP.
The framework of neutral net to be trained includes convolutional layer and Relu layers, in a kind of specific embodiment, builds The framework of neutral net refers to Fig. 2, including 3 convolutional layers (Convolution layer1, Convolution in figure Layer2, Convolution layer3) and 2 Relu layers (ReLU layer1 and the ReLU layer2 in figure), convolutional layer Alternately connected successively with Relu layers, for example, connecting convolutional layer 1 (Convolution layer1), convolution after input layer input The output connection Relu layers 1 (ReLU layer1) of layer 1, the output connection convolutional layer 2 (Convolution of Relu layers 1 Layer2), convolutional layer 2 (Convolution layer2) output connection Relu layers 2 (ReLU layer2), Relu layers 2 it is defeated Go out to connect convolutional layer 3 (Convolution layer3), the output connection output layer output of convolutional layer 3.It is to be appreciated that Fig. 2 What shown convolutional neural networks were merely exemplary, it is not construed as limiting the invention.To in the neutral net shown in Fig. 2, The filtering core size (filtering core is convolution kernel) for setting convolutional layer 1,2,3 is respectively 9,1,5, neuron number is respectively 64,32, 1;Relu layers of neuron number with and the Relu layers of neuron number of last layer convolutional layer being connected be consistent.
Then, using foregoing obtained many framings to separately go training it is above-mentioned put up for example shown in Fig. 2 Neutral net.First, by the frame in each group to dividing training set, checking collection and test set, example according to resolution ratio and scene Such as by QP be 20 and frame type be I frames all frames to be divided into training set, checking collection and test set, by QP be 20 and frame type Collect and test set to being divided into training set, checking for all frames of P frames, wherein, the frame in test set is to resolution ratio and field The characteristics of scape is various.Secondly, by all frames to according to certain pel spacing interception image block, obtaining " primitive frame image block- The image block pair of condensed frame image block ", is used as the input of neutral net in training process.
Illustrate the process for training neutral net so that QP is the framing pair that 20, type is I frames as an example, for example, to the group The frame of interior all frame centerings is that spacing carrys out interception image block by 28 pixels, such as the size of each image block is 33 × 33, In the training process, all convolutional layers all avoid boundary effect (because using padding phases without using padding operations When the pixel on border in imagineering, so as to expand the size of image, the information of boundary is just inaccurate), in order to The penalty values between output image block and original picture block are calculated, original picture block is also cut into output image block by the present invention It is in the same size.Then, all image blocks of interception are divided into multiple batches at random, the image block of a batch is sequentially input Neutral net, it is then an iteration to complete a batch, and how many batch is to carry out how many times iteration, sharp in an iterative process Back-propagating is carried out with stochastic gradient descent method, network parameter is updated, constantly loss function is minimized.Per the certain number of times of iteration The network parameter trained is then corrected using checking collection, to prevent over-fitting.
Specific training process is as shown in figure 3, operation of the neuron to the image block of input of convolutional layer is expressed asWherein n represents the quantity of input picture block, wiRepresent i-th of convolution kernel, xiRepresent i-th of input picture Block, b represents biasing coefficient, and * represents convolution operation, and image block is represented with picture element matrix;Relu layers of neuron is to image block Operation be expressed as M=max (N, 0), wherein N and M represent the pixel value of input picture block and output image block respectively.
When parameter restrains, training is completed, it is the framing that 20, type is I frames to corresponding neutral net to obtain QP Model, is embedded into encoder post-filtering link and is used as filtering.Similarly, QP is a framing that 20, type is P frames to also adopting Neutral net is trained in aforementioned manners, obtains corresponding to the neural network model that QP is the framing pair that 20, type is P frames; Frame under other QP values using same method to also training neutral net.
Assuming that the compression video employed in training process has two kinds of softwares of JM10 and HM12, then two kinds of Software Compressions are obtained Compression video should separate and handled, separate training.So as to by above-mentioned training method, then obtain following Multiple neural network models:
32 different I frame models that the lower correspondence QP of JM10 standard softwares coding is 20 to 51, JM10 standard softwares coding 32 different P frame models that lower correspondence QP is 20 to 51, the lower correspondence QP of JM10 standard softwares coding for 32 of 20 to 51 not Same B frame models, 32 different I frame models that the lower correspondence QP of HM12 standard softwares coding is 20 to 51, HM12 standard softwares 32 different P frame models that the lower correspondence QP of coding is 20 to 51, the lower correspondence QP of HM12 standard softwares coding is the 32 of 20 to 51 Individual different B frame models.
In actual use, coding compression and frame are carried out to extracting it to the pending encoded device of original video Afterwards, obtained frame is for example judged to learn that the QP of currently pending frame pair is to carrying out the judgement of quantization parameter and frame type 23rd, type is B frames, then it is that the image block that 23, type is B frames trains obtained neural network model come after carrying out that selection, which belongs to QP, Put filtering process.In a preferred embodiment, the encoding software of use can be also contemplated for into, i.e., in training always according to coding The difference of software is trained (as described by the preceding paragraph) respectively, so that it is which kind of that first resolution is further contemplated in actual use Encoding software compressed encoding is obtained, and considers further that quantization parameter and frame type to select corresponding neural network model to be filtered Processing.
Above content is to combine specific preferred embodiment further description made for the present invention, it is impossible to assert The specific implementation of the present invention is confined to these explanations.For those skilled in the art, do not taking off On the premise of from present inventive concept, some equivalent substitutes or obvious modification can also be made, and performance or purposes are identical, all should When being considered as belonging to protection scope of the present invention.

Claims (8)

1. a kind of Video coding post-filter method based on convolutional neural networks, includes the training step of convolutional neural networks model Rapid and post-filtering process step, wherein:
The training step includes S1 to S4:
S1, the quantization parameter of setting video compress are 20 to 51, and coding compression is carried out to original video, obtain compressing video;
S2, the extraction that frame is carried out to the compression video and the original video, obtain multiple frames pair, each frame to comprising One compressed video frame and an original video frame;
S3, the frame for extracting step S2 are multiple groups to the different demarcation according to frame type and quantization parameter;
S4, convolutional neural networks framework and initialization network parameter are built, the group divided using step S3 is respectively to neutral net It is trained, obtains corresponding to different quantization parameters and multiple neural network models of frame type;
The post-filtering process step includes S5 and S6:
S5, the post-filtering link that multiple neural network models that step S4 is obtained are embedded in video encoder;
S6, step S1 and S2 are performed to pending original video obtain pending frame pair, and according to the quantization of pending frame pair Parameter and the corresponding neural network model of frame type selecting are filtered processing.
2. the Video coding post-filter method as claimed in claim 1 based on convolutional neural networks, it is characterised in that:Step By frame to being divided into I frames, P frames and B frames according to frame type in S3.
3. the Video coding post-filter method as claimed in claim 1 based on convolutional neural networks, it is characterised in that:Step S3 is also included the frame in each group to dividing training set, checking collection and test set according to resolution ratio and scene;It is wherein described Checking collection is used for the parameter trained in the training process per iteration pre-determined number post-equalization, to prevent over-fitting.
4. the Video coding post-filter method as claimed in claim 3 based on convolutional neural networks, it is characterised in that:It is described Frame pair including a variety of resolution ratio and several scenes in test set.
5. the Video coding post-filter method as claimed in claim 1 based on convolutional neural networks, it is characterised in that:Step When S4 carries out neural metwork training, by frame to according to predetermined pel spacing interception image block, obtaining image block to as training During neutral net input;Wherein, each image block is to including a primitive frame image block and a condensed frame image block.
6. the Video coding post-filter method as claimed in claim 5 based on convolutional neural networks, it is characterised in that:Step The framework for the convolutional neural networks built in S4 includes convolutional layer and Relu layers;Wherein:
Operation of the neuron of convolutional layer to the image block of input is expressed asWherein n represents input picture block Quantity, wiRepresent i-th of convolution kernel, xiI-th of input picture block is represented, b represents to bias coefficient, and * represents convolution operation, schemes As block is represented with picture element matrix;
Operation of Relu layers of the neuron to image block is expressed as M=max (N, 0), and wherein N and M represent input picture block respectively With the pixel value of output image block.
7. the Video coding post-filter method as claimed in claim 6 based on convolutional neural networks, it is characterised in that:Convolution Layer and Relu layers of alternately connection successively.
8. the Video coding post-filter method as claimed in claim 5 based on convolutional neural networks, it is characterised in that:Nerve During network training, the size of the original picture block of input is cut in the same size with output image block.
CN201710439132.8A 2017-06-12 2017-06-12 Video coding post-filter method based on convolutional neural networks Active CN107197260B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710439132.8A CN107197260B (en) 2017-06-12 2017-06-12 Video coding post-filter method based on convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710439132.8A CN107197260B (en) 2017-06-12 2017-06-12 Video coding post-filter method based on convolutional neural networks

Publications (2)

Publication Number Publication Date
CN107197260A true CN107197260A (en) 2017-09-22
CN107197260B CN107197260B (en) 2019-09-13

Family

ID=59876636

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710439132.8A Active CN107197260B (en) 2017-06-12 2017-06-12 Video coding post-filter method based on convolutional neural networks

Country Status (1)

Country Link
CN (1) CN107197260B (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107743235A (en) * 2017-10-27 2018-02-27 厦门美图之家科技有限公司 Image processing method, device and electronic equipment
CN108134932A (en) * 2018-01-11 2018-06-08 上海交通大学 Filter achieving method and system in coding and decoding video loop based on convolutional neural networks
CN108174225A (en) * 2018-01-11 2018-06-15 上海交通大学 Filter achieving method and system in coding and decoding video loop based on confrontation generation network
CN108289224A (en) * 2017-12-12 2018-07-17 北京大学 A kind of video frame prediction technique, device and neural network is compensated automatically
CN108307193A (en) * 2018-02-08 2018-07-20 北京航空航天大学 A kind of the multiframe quality enhancement method and device of lossy compression video
CN109451308A (en) * 2018-11-29 2019-03-08 北京市商汤科技开发有限公司 Video compression method and device, electronic equipment and storage medium
WO2019141193A1 (en) * 2018-01-19 2019-07-25 杭州海康威视数字技术股份有限公司 Method and apparatus for processing video frame data
WO2019141255A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Image filtering method and device
WO2019141258A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Video encoding method, video decoding method, device, and system
CN110351568A (en) * 2019-06-13 2019-10-18 天津大学 A kind of filtering video loop device based on depth convolutional network
WO2019205871A1 (en) * 2018-04-25 2019-10-31 杭州海康威视数字技术股份有限公司 Image decoding and encoding methods and apparatuses, and device thereof
CN110740319A (en) * 2019-10-30 2020-01-31 腾讯科技(深圳)有限公司 Video encoding and decoding method and device, electronic equipment and storage medium
CN110971915A (en) * 2018-09-28 2020-04-07 杭州海康威视数字技术股份有限公司 Filtering method and device
CN111083482A (en) * 2019-12-31 2020-04-28 合肥图鸭信息科技有限公司 Video compression network training method and device and terminal equipment
CN111107357A (en) * 2018-10-25 2020-05-05 杭州海康威视数字技术股份有限公司 Image processing method, device and system
CN111212288A (en) * 2020-01-09 2020-05-29 广州虎牙科技有限公司 Video data encoding and decoding method and device, computer equipment and storage medium
CN111556337A (en) * 2020-05-15 2020-08-18 腾讯科技(深圳)有限公司 Media content implantation method, model training method and related device
CN111699686A (en) * 2018-01-26 2020-09-22 联发科技股份有限公司 Method and device for packet neural network for video coding and decoding
CN111726613A (en) * 2020-06-30 2020-09-29 福州大学 Video coding optimization method based on just noticeable difference
CN111742553A (en) * 2017-12-14 2020-10-02 交互数字Vc控股公司 Deep learning based image partitioning for video compression
CN111937392A (en) * 2018-04-17 2020-11-13 联发科技股份有限公司 Neural network method and device for video coding and decoding
CN112019843A (en) * 2019-05-30 2020-12-01 富士通株式会社 Encoding and decoding program, encoding and decoding device, encoding and decoding method
CN112470472A (en) * 2018-06-11 2021-03-09 无锡安科迪智能技术有限公司 Blind compression sampling method and device and imaging system
WO2021064291A1 (en) * 2019-10-02 2021-04-08 Nokia Technologies Oy Guiding decoder-side optimization of neural network filter
WO2021109846A1 (en) * 2019-12-06 2021-06-10 华为技术有限公司 Bit stream data processing method and apparatus
CN113259671A (en) * 2020-02-10 2021-08-13 腾讯科技(深圳)有限公司 Loop filtering method, device and equipment in video coding and decoding and storage medium
CN113542867A (en) * 2020-04-20 2021-10-22 声音猎手公司 Content filtering in a media playback device
CN114125449A (en) * 2021-10-26 2022-03-01 阿里巴巴新加坡控股有限公司 Video processing method, system and computer readable medium based on neural network
CN115460463A (en) * 2018-01-04 2022-12-09 三星电子株式会社 Video playing device and control method thereof
WO2023116173A1 (en) * 2021-12-21 2023-06-29 腾讯科技(深圳)有限公司 Data processing method and apparatus, and computer device and storage medium

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114501031B (en) * 2020-11-13 2023-06-02 华为技术有限公司 Compression coding and decompression method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6075884A (en) * 1996-03-29 2000-06-13 Sarnoff Corporation Method and apparatus for training a neural network to learn and use fidelity metric as a control mechanism
US7822698B1 (en) * 2007-03-23 2010-10-26 Hrl Laboratories, Llc Spike domain and pulse domain non-linear processors
CN105160678A (en) * 2015-09-02 2015-12-16 山东大学 Convolutional-neural-network-based reference-free three-dimensional image quality evaluation method
CN105894046A (en) * 2016-06-16 2016-08-24 北京市商汤科技开发有限公司 Convolutional neural network training and image processing method and system and computer equipment
CN106485661A (en) * 2016-11-15 2017-03-08 杭州当虹科技有限公司 A kind of high-quality image magnification method
CN106682697A (en) * 2016-12-29 2017-05-17 华中科技大学 End-to-end object detection method based on convolutional neural network

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6075884A (en) * 1996-03-29 2000-06-13 Sarnoff Corporation Method and apparatus for training a neural network to learn and use fidelity metric as a control mechanism
US7822698B1 (en) * 2007-03-23 2010-10-26 Hrl Laboratories, Llc Spike domain and pulse domain non-linear processors
CN105160678A (en) * 2015-09-02 2015-12-16 山东大学 Convolutional-neural-network-based reference-free three-dimensional image quality evaluation method
CN105894046A (en) * 2016-06-16 2016-08-24 北京市商汤科技开发有限公司 Convolutional neural network training and image processing method and system and computer equipment
CN106485661A (en) * 2016-11-15 2017-03-08 杭州当虹科技有限公司 A kind of high-quality image magnification method
CN106682697A (en) * 2016-12-29 2017-05-17 华中科技大学 End-to-end object detection method based on convolutional neural network

Cited By (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107743235A (en) * 2017-10-27 2018-02-27 厦门美图之家科技有限公司 Image processing method, device and electronic equipment
CN108289224B (en) * 2017-12-12 2019-10-29 北京大学 A kind of video frame prediction technique, device and neural network is compensated automatically
CN108289224A (en) * 2017-12-12 2018-07-17 北京大学 A kind of video frame prediction technique, device and neural network is compensated automatically
CN111742553A (en) * 2017-12-14 2020-10-02 交互数字Vc控股公司 Deep learning based image partitioning for video compression
CN115460463A (en) * 2018-01-04 2022-12-09 三星电子株式会社 Video playing device and control method thereof
CN108174225B (en) * 2018-01-11 2021-03-26 上海交通大学 Video coding and decoding in-loop filtering implementation method and system based on countermeasure generation network
CN108134932B (en) * 2018-01-11 2021-03-30 上海交通大学 Method and system for realizing video coding and decoding loop internal filtering based on convolutional neural network
CN108174225A (en) * 2018-01-11 2018-06-15 上海交通大学 Filter achieving method and system in coding and decoding video loop based on confrontation generation network
CN108134932A (en) * 2018-01-11 2018-06-08 上海交通大学 Filter achieving method and system in coding and decoding video loop based on convolutional neural networks
WO2019141255A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Image filtering method and device
WO2019141258A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Video encoding method, video decoding method, device, and system
WO2019141193A1 (en) * 2018-01-19 2019-07-25 杭州海康威视数字技术股份有限公司 Method and apparatus for processing video frame data
CN110062246A (en) * 2018-01-19 2019-07-26 杭州海康威视数字技术股份有限公司 The method and apparatus that video requency frame data is handled
CN111699686A (en) * 2018-01-26 2020-09-22 联发科技股份有限公司 Method and device for packet neural network for video coding and decoding
WO2019154152A1 (en) * 2018-02-08 2019-08-15 北京航空航天大学 Multi-frame quality enhancement method and device for lossy compressed video
CN108307193A (en) * 2018-02-08 2018-07-20 北京航空航天大学 A kind of the multiframe quality enhancement method and device of lossy compression video
CN108307193B (en) * 2018-02-08 2018-12-18 北京航空航天大学 A kind of the multiframe quality enhancement method and device of lossy compression video
US10965959B2 (en) 2018-02-08 2021-03-30 Beihang University Multi-frame quality enhancement for compressed video
CN111937392A (en) * 2018-04-17 2020-11-13 联发科技股份有限公司 Neural network method and device for video coding and decoding
CN110401836B (en) * 2018-04-25 2022-04-26 杭州海康威视数字技术股份有限公司 Image decoding and encoding method, device and equipment
WO2019205871A1 (en) * 2018-04-25 2019-10-31 杭州海康威视数字技术股份有限公司 Image decoding and encoding methods and apparatuses, and device thereof
CN110401836A (en) * 2018-04-25 2019-11-01 杭州海康威视数字技术股份有限公司 A kind of image decoding, coding method, device and its equipment
CN112470472A (en) * 2018-06-11 2021-03-09 无锡安科迪智能技术有限公司 Blind compression sampling method and device and imaging system
CN110971915A (en) * 2018-09-28 2020-04-07 杭州海康威视数字技术股份有限公司 Filtering method and device
CN111107357B (en) * 2018-10-25 2022-05-31 杭州海康威视数字技术股份有限公司 Image processing method, device, system and storage medium
CN111107357A (en) * 2018-10-25 2020-05-05 杭州海康威视数字技术股份有限公司 Image processing method, device and system
CN109451308A (en) * 2018-11-29 2019-03-08 北京市商汤科技开发有限公司 Video compression method and device, electronic equipment and storage medium
US11290723B2 (en) 2018-11-29 2022-03-29 Beijing Sensetime Technology Development Co., Ltd. Method for video compression processing, electronic device and storage medium
CN109451308B (en) * 2018-11-29 2021-03-09 北京市商汤科技开发有限公司 Video compression processing method and device, electronic equipment and storage medium
CN112019843A (en) * 2019-05-30 2020-12-01 富士通株式会社 Encoding and decoding program, encoding and decoding device, encoding and decoding method
CN112019843B (en) * 2019-05-30 2022-07-15 富士通株式会社 Encoding and decoding method, apparatus, readable storage medium
CN110351568A (en) * 2019-06-13 2019-10-18 天津大学 A kind of filtering video loop device based on depth convolutional network
WO2021064291A1 (en) * 2019-10-02 2021-04-08 Nokia Technologies Oy Guiding decoder-side optimization of neural network filter
US11341688B2 (en) 2019-10-02 2022-05-24 Nokia Technologies Oy Guiding decoder-side optimization of neural network filter
CN110740319B (en) * 2019-10-30 2024-04-05 腾讯科技(深圳)有限公司 Video encoding and decoding method and device, electronic equipment and storage medium
CN110740319A (en) * 2019-10-30 2020-01-31 腾讯科技(深圳)有限公司 Video encoding and decoding method and device, electronic equipment and storage medium
WO2021109846A1 (en) * 2019-12-06 2021-06-10 华为技术有限公司 Bit stream data processing method and apparatus
CN111083482A (en) * 2019-12-31 2020-04-28 合肥图鸭信息科技有限公司 Video compression network training method and device and terminal equipment
CN111212288B (en) * 2020-01-09 2022-10-04 广州虎牙科技有限公司 Video data encoding and decoding method and device, computer equipment and storage medium
CN111212288A (en) * 2020-01-09 2020-05-29 广州虎牙科技有限公司 Video data encoding and decoding method and device, computer equipment and storage medium
CN113259671A (en) * 2020-02-10 2021-08-13 腾讯科技(深圳)有限公司 Loop filtering method, device and equipment in video coding and decoding and storage medium
CN113542867A (en) * 2020-04-20 2021-10-22 声音猎手公司 Content filtering in a media playback device
CN111556337A (en) * 2020-05-15 2020-08-18 腾讯科技(深圳)有限公司 Media content implantation method, model training method and related device
CN111726613A (en) * 2020-06-30 2020-09-29 福州大学 Video coding optimization method based on just noticeable difference
CN111726613B (en) * 2020-06-30 2021-07-27 福州大学 Video coding optimization method based on just noticeable difference
CN114125449A (en) * 2021-10-26 2022-03-01 阿里巴巴新加坡控股有限公司 Video processing method, system and computer readable medium based on neural network
WO2023116173A1 (en) * 2021-12-21 2023-06-29 腾讯科技(深圳)有限公司 Data processing method and apparatus, and computer device and storage medium

Also Published As

Publication number Publication date
CN107197260B (en) 2019-09-13

Similar Documents

Publication Publication Date Title
CN107197260B (en) Video coding post-filter method based on convolutional neural networks
CN106960416A (en) A kind of video satellite compression image super-resolution method of content complexity self adaptation
CN106716997A (en) Video encoding method using in-loop filter parameter prediction and apparatus therefor, and video decoding method and apparatus therefor
CN110351568A (en) A kind of filtering video loop device based on depth convolutional network
CN104093021B (en) Monitoring video compression method
CN103782598A (en) Fast encoding method for lossless coding
CN108184118A (en) Cloud desktop contents encode and coding/decoding method and device, system
CN108900848A (en) A kind of video quality Enhancement Method based on adaptive separable convolution
CN113766249B (en) Loop filtering method, device, equipment and storage medium in video coding and decoding
CN111031315B (en) Compressed video quality enhancement method based on attention mechanism and time dependence
CN101272489B (en) Encoding and decoding device and method for video image quality enhancement
CN105791877A (en) Adaptive loop filter method in video coding and decoding
CN111247797A (en) Method and apparatus for image encoding and decoding
CN103347185A (en) Unmanned aerial vehicle scouting image synthesis compressed encoding method based on selective block transformation
CN109903351A (en) The method for compressing image combined based on convolutional neural networks and traditional code
Li et al. Multi-scale grouped dense network for vvc intra coding
CN112188217B (en) JPEG compressed image decompression effect removing method combining DCT domain and pixel domain learning
CN113055674B (en) Compressed video quality enhancement method based on two-stage multi-frame cooperation
KR100697516B1 (en) Moving picture coding method based on 3D wavelet transformation
CN110677644A (en) Video coding and decoding method and video coding intra-frame predictor
CN103747257B (en) A kind of method of video data high efficient coding
CN112001854A (en) Method for repairing coded image and related system and device
KR101998036B1 (en) Artifact reduction method and apparatus
CN103313064B (en) Based on the time domain error hidden method of inter-frame mode and movement repair
CN115002482B (en) End-to-end video compression method and system using structural preserving motion estimation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant