CN113132735A - Video coding method based on video frame generation - Google Patents

Video coding method based on video frame generation Download PDF

Info

Publication number
CN113132735A
CN113132735A CN201911392082.8A CN201911392082A CN113132735A CN 113132735 A CN113132735 A CN 113132735A CN 201911392082 A CN201911392082 A CN 201911392082A CN 113132735 A CN113132735 A CN 113132735A
Authority
CN
China
Prior art keywords
frame
neural network
target
video
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911392082.8A
Other languages
Chinese (zh)
Inventor
刘家瑛
段凌宇
夏思烽
杨文瀚
胡越予
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University
Original Assignee
Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University filed Critical Peking University
Priority to CN201911392082.8A priority Critical patent/CN113132735A/en
Publication of CN113132735A publication Critical patent/CN113132735A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention discloses a video coding method based on video frame generation, which comprises the following steps: training a neural network: at each training iteration, two frames I of a video segment are extracted from the sample video training setkAnd ItSending into neural network to generate pair ItPrediction of (2)
Figure DDA0002345278860000011
Computing
Figure DDA0002345278860000012
And a target frame ItThe L1 norm in between and back-propagating it to the neural network until the neural network converges; and (3) an encoding stage: the encoding end extracts sparse motion representation between the encoded reference frame and the target to-be-encoded non-key frame by utilizing a neural network to generate a prediction frame; adding the predicted frame into a reference frame list for inter-frame prediction, and then sending inter-frame prediction information and sparse motion representation to a decoding end; and a decoding stage: the decoding end estimates dense motion information of the target frame according to the reconstructed reference frame and the transmitted sparse motion representation and generates the target frame; and then adding the generated target frame into a reference frame list and reconstructing the target frame by utilizing inter-frame prediction information.

Description

Video coding method based on video frame generation
Technical Field
The invention belongs to the field of video coding, relates to an inter-frame prediction module of video coding, and particularly relates to a video coding method based on video frame generation, which can be used for improving the video compression rate.
Background
In the use and transmission process of digital video, video coding and decoding are indispensable key technologies. The video coding and decoding technology greatly reduces the cost of the digital video in the storage and transmission processes by carrying out coding compression on the video at the coding end and decoding recovery on the video at the decoding end, so that the use of the digital video in daily life is realized. The inter-frame prediction module is a key method for improving the video compression rate by using inter-frame redundant information in the video coding and decoding technology.
In the interframe prediction module, an encoder searches an encoded reference block similar to a current video frame block to be encoded in an encoded video frame, and based on the similar encoded reference block, the encoder can only encode and record a residual error between the block to be encoded and the reference block and corresponding block motion offset information without encoding complete information of the block to be encoded, so that the bit number required by encoding is saved, and the compression rate is improved. However, since the search for the reference block is performed in units of image blocks, when there is a pixel-level motion offset (e.g., rotation, distortion) between the reference block and the block to be encoded, the encoder cannot obtain a good prediction of the block to be encoded, and the performance of inter-coding is affected.
Disclosure of Invention
On the premise of the technical background, the invention provides an inter-frame prediction method based on video frame generation, which generates a target to-be-coded video frame by using the prediction of a coded video frame, so that after the operation is also performed at a decoding end, a reference with a very close pixel level can be provided for the time domain prediction of the target to-be-coded frame, and the effect of removing time domain redundancy is improved.
In order to ensure the generated effect, the motion between frames needs to be well modeled. An intuitive idea is to extract dense motion information between the reference frame and the target frame and transmit it to the decoding side as well for the generation of the target frame. However, the transmission of dense motion information itself adds additional bit overhead. Therefore, another core of the present invention is to extract a sparse representation of inter-frame motion and model inter-frame motion based on such a sparse representation, thereby achieving an ideal target frame generation with less overhead.
The technical scheme of the invention is as follows:
a video encoding method based on video frame generation, comprising the steps of:
1) training a neural network: at each training iteration, two frames I of a video segment are extracted from the sample video training setkAnd ItRespectively used as source frame and target frame, and fed into neural network to extract ItAnd IkThen the neural network estimates I based on the extracted sparse motion characterizationkTo ItDense motion light stream xik→tAnd extracting the source frame IkThen based on ξk→tTransforming the related content features to generate pairs ItPrediction of (2)
Figure BDA0002345278840000021
Computing
Figure BDA0002345278840000022
And a target frame ItL1 norm in between and back-propagating it to the neural network to update the weights of the neural network until it isThe neural network converges;
2) and (3) an encoding stage: the encoding end extracts sparse motion representation between the encoded reference frame and the target to-be-encoded non-key frame by using the trained neural network; then, generating a prediction frame of a target frame by using the extracted sparse motion representation and the reference frame; adding the generated prediction frame into a reference frame list for inter-frame prediction, and then sending the obtained inter-frame prediction information and the sparse motion representation to a decoding end;
3) and a decoding stage: the decoding end estimates dense motion information of the target frame according to the reconstructed reference frame and the transmitted sparse motion representation and generates the target frame; and then adding the generated target frame into a reference frame list and reconstructing the target frame by utilizing the transmitted inter-frame prediction information.
Further, the neural network performs forward calculation of the neural network on the input source frame and the input target frame to obtain the sparse motion representation.
Further, the frame number of the source frame is smaller than the frame number of the target frame.
Furthermore, each time of training iteration, a first frame is extracted from the video segment as a source frame, and another frame randomly selected from the video segment is used as a target frame.
And further, the sparse motion representation is quantized and subjected to lossless compression and then transmitted to a decoding end.
Further, in the encoding stage, an encoder firstly divides frames in a video sequence to be encoded into key frames and non-key frames; for the key frame, the traditional coding and decoding method is still used for compressing and reconstructing; and (3) encoding and decoding non-key frames by adopting the steps 2) to 3).
A target frame generation network training generation method comprises the following steps:
1) selecting a sample video training set;
2) at each training iteration, two frames I of a video segment are extracted from the sample video training setkAnd ItRespectively used as source frame and target frame, and fed into neural network to extract ItAnd IkSparse motion characterization in between;
3) neural network estimation I based on extracted sparse motion characterizationkTo ItDense motion light stream xik→tAnd from the source frame IkExtracting content features;
4) dense motion-based optical flow xik→tTransforming the extracted content features to generate pairs ItPrediction of (2)
Figure BDA0002345278840000023
5) Computing
Figure BDA0002345278840000031
And a target frame ItThe L1 norm in between and back-propagating it to the neural network to update the weights of the neural network until the neural network converges; the converged neural network is used as a target frame generation network.
The invention adds a new inter-frame prediction reference under the existing video coding frame, and designs two convolution neural networks for sparse motion representation and target frame generation to generate a better inter-frame prediction reference.
The main steps of the method of the invention are described next. The method mainly comprises two parts, wherein one part is the training of sparse motion representation and a target frame generation network, and the other part is a newly designed video coding frame.
Firstly, describing the training of the network, wherein the network structure is as shown in fig. 1, and for each training sample, firstly, extracting the inter-frame sparse motion representation. Then another group of networks extracts the content of the input source video frame from the sparse motion representation, and simultaneously estimates dense motion information between frames from the sparse motion representation. And finally, realizing the transformation from the source frame to the target frame by the guidance of the estimated dense motion information, and realizing the generation of the target frame. The network training process is as follows:
step 1: extracting video clips of a plurality of continuous frames from a batch of videos to construct training data pairs; each training data pair includes a source frame and a target frame.
Step 2: in each training iteration, a source frame and a target frame are formed by extracting a first frame and a random other frame from each video segment, and a group of two video frames are sent into a network for forward calculation of the network. The first frame is selected as the source frame as the key frame to correspond to the relationship between the key frame and the non-key frame in the actual encoding process, that is, one key frame corresponds to a plurality of subsequent non-key frames in the actual encoding.
And step 3: and 2, obtaining a calculation result, namely a generation result from the source frame to the target frame, and calculating an L1 norm between the generated target frame and the original target frame.
And 4, step 4: and reversely transmitting the calculated norm to each layer of the neural network to update the weight of each layer, so that the result is closer to the target effect in the next iteration.
And 5: repeating steps 1-4 until the L1 norm of the neural network converges.
And after the trained network model is obtained, applying the model to an interframe prediction module of an encoder. Specifically, at the encoding end, frames in a video sequence to be encoded are distinguished into key frames and non-key frames. For key frames, the conventional video coding method is still used. For non-key frames, the video coding method of the present invention is used. The video coding of the invention additionally adds a generated reference frame on the basis of the traditional video coding method to improve the inter-frame prediction performance. Specifically, for each non-key frame to be encoded, an encoded reference frame and the non-key frame to be encoded are input into a trained network, the encoded reference frame is used for generating the non-key frame to be encoded (namely, a predicted frame), the generated non-key frame to be encoded is added into a reference frame list of the non-key frame to be encoded, and the inter-frame prediction process of the traditional encoder is also carried out. And after the flow is finished, in addition to the relevant information of inter-frame prediction, additionally storing the sparse motion representation of the transmission network extracted from the inter-frame. At the decoder side, dense motion estimation can be performed and an object can be generated by using the reconstructed reference frame and the transmitted sparse motion representation. And after the generation is finished, adding the generated result into a reference frame list, and then reconstructing the target frame by using the inter-frame prediction related information transmitted by coding.
Compared with the prior art, the invention has the following positive effects:
the invention constructs a convolutional neural network based on sparse motion sampling and motion generation to generate a reference frame with smaller pixel-level motion offset with a target frame, and adds the generated reference frame into a reference frame list, thereby further compressing inter-frame time domain redundant information and improving the video coding performance. The invention can better model the complex inter-frame motion and obtain better time domain prediction effect in the inter-frame prediction process, thereby further avoiding the coding overhead of the complex motion and better removing the time domain redundancy.
Drawings
Fig. 1 is a diagram of a network architecture of the present invention.
Detailed Description
For further explanation of the technical method of the present invention, the video encoding method based on video frame generation according to the present invention is further described in detail below with reference to the drawings and specific examples of the specification.
The present example will focus on a detailed description of the training and specific encoding processes of the neural network part of the technical approach, respectively. Suppose now that the present invention has constructed the convolutional neural network model of FIG. 1 and has N video segments { I }1,I2,…,INAs training data.
The present example method in conjunction with FIG. 1 is as follows:
firstly, a training process:
step 1: from the training set I in each training iteration1,I2,…,INExtracting two frames I of a video clipkAnd ItAs source and target frames, respectively.
Step 2: the network first extracts ItAnd IkThen the network estimates I based on the extracted sparse motion characterizationkTo ItDense motion light stream xik→t. Simultaneously with source frame IkExtraction of video frame content features for input (content features extracted through network adaptive learning) followed by ξ -based extractionk→tTo ItIs transformed to generate a pair ItPrediction of (2)
Figure BDA0002345278840000041
Sparse motion characterization refers to the method that can be used for characterizing motion between video frames but at the same time, information itself is sparse, that is, the amount of information is small, and the encoding overhead is small, such as some key points that can be selected.
And step 3: calculating the estimation result of step 2
Figure BDA0002345278840000042
And the original target frame ItThe L1 norm in between as a loss function.
And 4, step 4: after obtaining the loss function, the error value is propagated reversely to the network to train the network to update the network weight.
And 5, repeating the steps 1-4 until the neural network converges.
Secondly, an encoding process:
after training is finished and sparse motion representation extraction and a target frame generation network are completed, in the actual test of an encoder, a video frame is divided into a key frame and a non-key frame, and for the key frame, the original encoding and decoding method of the encoder is still used for compression and reconstruction. For non-key frames, the operation steps are as follows, firstly at the encoding end:
step 1: extracting sparse motion representation between an encoded reference frame and a target to-be-encoded non-key frame by using a network;
step 2: generating a target frame by using the extracted sparse motion representation and the reference frame;
and step 3: adding the generated target frame into a reference frame list, and continuing the interframe prediction process of the traditional coding method;
and 4, step 4: and additionally carrying out appropriate optional quantization and lossless compression on the extracted sparse motion representation, and transmitting the sparse motion representation to a decoding end.
At the decoding end:
the method comprises the following steps: reconstructing a sparse motion representation of the transmission;
step two: estimating dense motion information of the target frame by using the reconstructed reference frame and the sparse motion representation and generating a prediction frame of the target frame;
step three: and adding the predicted frame of the target frame into a reference frame list and reconstructing the target frame by utilizing the inter-frame prediction related information transmitted by the encoder.
Fig. 1 summarizes the network architecture of the present invention. It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (10)

1. A video encoding method based on video frame generation, comprising the steps of:
1) training a neural network: at each training iteration, two frames I of a video segment are extracted from the sample video training setkAnd ItRespectively used as source frame and target frame, and fed into neural network to extract ItAnd IkThen the neural network estimates I based on the extracted sparse motion characterizationkTo ItDense motion light stream xik→tAnd extracting the source frame IkThen based on ξk→tTransforming the related content features to generate pairs ItPrediction of (2)
Figure FDA0002345278830000011
Computing
Figure FDA0002345278830000012
And a target frame ItThe L1 norm in between and back-propagating it to the neural network to update the weights of the neural network until the neural network converges;
2) and (3) an encoding stage: the encoding end extracts sparse motion representation between the encoded reference frame and the target to-be-encoded non-key frame by using the trained neural network; then, generating a prediction frame of a target frame by using the extracted sparse motion representation and the reference frame; adding the generated prediction frame into a reference frame list for inter-frame prediction, and then sending the obtained inter-frame prediction information and the sparse motion representation to a decoding end;
3) and a decoding stage: the decoding end estimates dense motion information of the target frame according to the reconstructed reference frame and the transmitted sparse motion representation and generates the target frame; and then adding the generated target frame into a reference frame list and reconstructing the target frame by utilizing the transmitted inter-frame prediction information.
2. The method of claim 1, wherein the neural network performs a forward calculation of the neural network on the input source frames and target frames to obtain the sparse motion characterization.
3. The method of claim 1, wherein the frame sequence number of the source frame is less than the frame sequence number of the destination frame.
4. The method of claim 1, wherein at each training iteration, a first frame is extracted from the video segment as a source frame and another frame randomly selected from the video segment is taken as a target frame.
5. The method of claim 1, wherein the sparse motion representation is quantized and lossless compressed before being transmitted to a decoding end.
6. The method of claim 1, wherein in the encoding stage, the encoder first distinguishes frames in the video sequence to be encoded into key frames and non-key frames; for the key frame, the traditional coding and decoding method is still used for compressing and reconstructing; and (3) encoding and decoding non-key frames by adopting the steps 2) to 3).
7. A target frame generation network training generation method comprises the following steps:
1) selecting a sample video training set;
2) training from sample video each time training iterationsTwo frames I of a video segment are collectively extractedkAnd ItRespectively used as source frame and target frame, and fed into neural network to extract ItAnd IkSparse motion characterization in between;
3) neural network estimation I based on extracted sparse motion characterizationkTo ItDense motion light stream xik→tAnd from the source frame IkExtracting content features;
4) dense motion-based optical flow xik→tTransforming the extracted content features to generate pairs ItPrediction of (2)
Figure FDA0002345278830000021
5) Computing
Figure FDA0002345278830000022
And a target frame ItThe L1 norm in between and back-propagating it to the neural network to update the weights of the neural network until the neural network converges; the converged neural network is used as a target frame generation network.
8. The method of claim 7, wherein the frame sequence number of the source frame is less than the frame sequence number of the target frame.
9. The method of claim 7, wherein the neural network performs a forward calculation of the neural network on the input source frames and target frames to obtain the sparse motion characterization.
10. The method of claim 7, wherein at each training iteration, a first frame is extracted from the video segment as a source frame and another frame is randomly selected from the video segment as a target frame.
CN201911392082.8A 2019-12-30 2019-12-30 Video coding method based on video frame generation Pending CN113132735A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911392082.8A CN113132735A (en) 2019-12-30 2019-12-30 Video coding method based on video frame generation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911392082.8A CN113132735A (en) 2019-12-30 2019-12-30 Video coding method based on video frame generation

Publications (1)

Publication Number Publication Date
CN113132735A true CN113132735A (en) 2021-07-16

Family

ID=76767759

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911392082.8A Pending CN113132735A (en) 2019-12-30 2019-12-30 Video coding method based on video frame generation

Country Status (1)

Country Link
CN (1) CN113132735A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114466199A (en) * 2022-04-12 2022-05-10 宁波康达凯能医疗科技有限公司 Reference frame generation method and system applicable to VVC (variable valve timing) coding standard
WO2023050431A1 (en) * 2021-09-30 2023-04-06 浙江大学 Encoding method, decoding method, decoder, encoder and computer-readable storage medium
WO2023143349A1 (en) * 2022-01-25 2023-08-03 阿里巴巴(中国)有限公司 Facial video encoding method and apparatus, and facial video decoding method and apparatus
WO2023143331A1 (en) * 2022-01-25 2023-08-03 阿里巴巴(中国)有限公司 Facial video encoding method, facial video decoding method, and apparatus

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107105278A (en) * 2017-04-21 2017-08-29 中国科学技术大学 The coding and decoding video framework that motion vector is automatically generated
CN107968946A (en) * 2016-10-18 2018-04-27 深圳万兴信息科技股份有限公司 Video frame rate method for improving and device
WO2019009452A1 (en) * 2017-07-06 2019-01-10 삼성전자 주식회사 Method and device for encoding or decoding image
CN109379550A (en) * 2018-09-12 2019-02-22 上海交通大学 Video frame rate upconversion method and system based on convolutional neural networks

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107968946A (en) * 2016-10-18 2018-04-27 深圳万兴信息科技股份有限公司 Video frame rate method for improving and device
CN107105278A (en) * 2017-04-21 2017-08-29 中国科学技术大学 The coding and decoding video framework that motion vector is automatically generated
WO2019009452A1 (en) * 2017-07-06 2019-01-10 삼성전자 주식회사 Method and device for encoding or decoding image
CN109379550A (en) * 2018-09-12 2019-02-22 上海交通大学 Video frame rate upconversion method and system based on convolutional neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
HYOMIN CHOI等: "Deep Frame Prediction for Video Coding", 《IEEE TRANSACTIONS CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY》 *
ZHAO LEI等: "Enhanced Motion-Compensated Video Coding With Deep Virtual Reference Frame Generation", 《IEEE TRANSACTIONS ON IMAGE PROCESSING》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023050431A1 (en) * 2021-09-30 2023-04-06 浙江大学 Encoding method, decoding method, decoder, encoder and computer-readable storage medium
WO2023143349A1 (en) * 2022-01-25 2023-08-03 阿里巴巴(中国)有限公司 Facial video encoding method and apparatus, and facial video decoding method and apparatus
WO2023143331A1 (en) * 2022-01-25 2023-08-03 阿里巴巴(中国)有限公司 Facial video encoding method, facial video decoding method, and apparatus
CN114466199A (en) * 2022-04-12 2022-05-10 宁波康达凯能医疗科技有限公司 Reference frame generation method and system applicable to VVC (variable valve timing) coding standard

Similar Documents

Publication Publication Date Title
Wang et al. Wireless deep video semantic transmission
WO2021164176A1 (en) End-to-end video compression method and system based on deep learning, and storage medium
CN113132735A (en) Video coding method based on video frame generation
CN106973293B (en) Light field image coding method based on parallax prediction
CN112203093B (en) Signal processing method based on deep neural network
Campos et al. Content adaptive optimization for neural image compression
CN110248190B (en) Multilayer residual coefficient image coding method based on compressed sensing
CN108174218B (en) Video coding and decoding system based on learning
CN101626512A (en) Method and device of multiple description video coding based on relevance optimization rule
CN102256133B (en) Distributed video coding and decoding method based on side information refining
CN104539961B (en) Gradable video encoding system based on the gradual dictionary learning of hierarchy
CN110062239B (en) Reference frame selection method and device for video coding
CN110290386B (en) Low-bit-rate human motion video coding system and method based on generation countermeasure network
CN112218072A (en) Video coding method based on deconstruction compression and fusion
CN114501013A (en) Variable bit rate video compression method, system, device and storage medium
CN113132727B (en) Scalable machine vision coding method and training method of motion-guided image generation network
CN111711815A (en) Fast VVC intra-frame prediction method based on integrated learning and probability model
Li et al. Multiple description coding based on convolutional auto-encoder
WO2023024115A1 (en) Encoding method, decoding method, encoder, decoder and decoding system
CN113068041B (en) Intelligent affine motion compensation coding method
Djelouah et al. Content adaptive optimization for neural image compression
CN115529457B (en) Video compression method and device based on deep learning
WO2023225808A1 (en) Learned image compress ion and decompression using long and short attention module
KR102613527B1 (en) Microdosing for low bitrate video compression
CN111726636A (en) HEVC (high efficiency video coding) coding optimization method based on time domain downsampling and frame rate upconversion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210716

RJ01 Rejection of invention patent application after publication