CN113507607B - Compressed video multi-frame quality enhancement method without motion compensation - Google Patents
Compressed video multi-frame quality enhancement method without motion compensation Download PDFInfo
- Publication number
- CN113507607B CN113507607B CN202110654128.XA CN202110654128A CN113507607B CN 113507607 B CN113507607 B CN 113507607B CN 202110654128 A CN202110654128 A CN 202110654128A CN 113507607 B CN113507607 B CN 113507607B
- Authority
- CN
- China
- Prior art keywords
- block
- feature map
- module
- quality
- residual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000012549 training Methods 0.000 claims abstract description 7
- 230000004927 fusion Effects 0.000 claims description 39
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 32
- 238000013507 mapping Methods 0.000 claims description 28
- 238000010586 diagram Methods 0.000 claims description 23
- 238000000605 extraction Methods 0.000 claims description 21
- 230000004913 activation Effects 0.000 claims description 11
- 230000007246 mechanism Effects 0.000 claims description 11
- 238000011176 pooling Methods 0.000 claims description 9
- 238000007781 pre-processing Methods 0.000 claims description 9
- 238000005070 sampling Methods 0.000 claims description 3
- 230000006835 compression Effects 0.000 abstract description 10
- 238000007906 compression Methods 0.000 abstract description 10
- 230000003287 optical effect Effects 0.000 abstract description 8
- 238000013527 convolutional neural network Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 11
- 230000000694 effects Effects 0.000 description 9
- 230000002708 enhancing effect Effects 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 238000013139 quantization Methods 0.000 description 5
- 238000004590 computer program Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- UIFGGABIJBWRMG-FMQUCBEESA-N (4-chlorophenyl)methyl (ne)-n-[(4-chlorophenyl)methoxycarbonylimino]carbamate Chemical compound C1=CC(Cl)=CC=C1COC(=O)\N=N\C(=O)OCC1=CC=C(Cl)C=C1 UIFGGABIJBWRMG-FMQUCBEESA-N 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 230000015556 catabolic process Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000006731 degradation reaction Methods 0.000 description 2
- 230000001747 exhibiting effect Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000002427 irreversible effect Effects 0.000 description 1
- 238000013441 quality evaluation Methods 0.000 description 1
- 238000011158 quantitative evaluation Methods 0.000 description 1
- ZFMRLFXUPVQYAU-UHFFFAOYSA-N sodium 5-[[4-[4-[(7-amino-1-hydroxy-3-sulfonaphthalen-2-yl)diazenyl]phenyl]phenyl]diazenyl]-2-hydroxybenzoic acid Chemical compound C1=CC(=CC=C1C2=CC=C(C=C2)N=NC3=C(C=C4C=CC(=CC4=C3O)N)S(=O)(=O)O)N=NC5=CC(=C(C=C5)O)C(=O)O.[Na+] ZFMRLFXUPVQYAU-UHFFFAOYSA-N 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/154—Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/48—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using compressed domain processing techniques other than decoding, e.g. modification of transform coefficients, variable length coding [VLC] data or run-length data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
- H04N19/86—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
The invention discloses a compressed video multi-frame quality enhancement method without motion compensation, belonging to the technical field of video compression. The invention adopts a multi-frame quality enhancement network without optical flow estimation, solves the contradiction between the definition and color richness of the video and the limited network bandwidth in the multimedia application, enhances the low-quality frame by fully utilizing the time information between adjacent frames, and enhances the subjective and objective quality of the compressed video. At the same time, the invention does not need to utilize optical flow estimation to compensate the motion between adjacent frames in a displaying way, thereby simplifying the network training. The application of the invention can ensure that the high-definition video is normally transmitted in the network after being compressed under the same code rate, and can also ensure better subjective and objective quality.
Description
Technical Field
The invention belongs to the technical field of video compression, and particularly relates to a compressed video multi-frame quality enhancement method without motion compensation.
Background
Since the international standard of video compression was proposed, the compressed video quality enhancement method has been widely studied in the industry and academia. Before deep learning is proposed, the method for enhancing the quality of compressed video is mainly a spatial domain method and a frequency domain method for enhancing single-frame images based on mathematical derivation. After deep learning is successfully applied to the field of image enhancement, various new methods have been proposed for enhancing the quality of compressed video, resulting in better results and more generalization than conventional methods.
The most commonly used H.265/HEVC standard now adopts a mixed coding framework based on blocks, and the core process comprises the following steps: predictive coding, transform coding, quantization and entropy coding, and block-based prediction. The transformation and quantization operations ignore block-to-block correlation, resulting in the encoded reconstructed image exhibiting blockiness, i.e., significant discontinuities in block boundaries that can be perceived by the human eye (these effects are more pronounced when the step size is larger and the bit rate is lower); meanwhile, quantization is based on block expansion in the transform domain, and this quantization process is irreversible. In addition, high precision interpolation in motion compensation is prone to ringing effects. Due to the accumulation of errors in the inter-frame coding process, the above effects also affect the coding quality of the subsequent frames, thereby resulting in a degradation of the objective evaluation quality of the video image and a degradation of the visual perception quality of the human eye.
In the existing solutions, chinese patent application publication No. CN107481209a proposes an enhancement scheme named "an image or video quality enhancement method based on convolutional neural network", in which two convolutional neural networks for video (or image) quality enhancement are first designed, the two networks having different computational complexity; then selecting a plurality of training images or videos to train parameters in the two convolutional neural networks; according to actual needs, a convolutional neural network with proper computational complexity is selected, and an image or video with the quality to be enhanced is input into the selected network; finally, the network outputs the quality enhanced image or video. The invention can effectively enhance the video quality; the user can select the convolutional neural network with proper calculation complexity according to the calculation capacity or the residual electric quantity of the equipment to enhance the quality of the image or the video. In the scheme, two convolutional neural networks with different complexity are designed, a user selects the networks according to the condition of equipment, the difference between the two networks is only that the depths of the convolutional neural networks are different, the scheme of improving the quality enhancement effect by only deepening the depths of the networks is not feasible, and the network is not designed aiming at the characteristics of the image video, namely, the network cannot utilize the time correlation between video frames, so the quality enhancement effect of the method is limited.
And publication No. CN108900848A proposes an enhancement scheme called "an adaptive separable convolution-based video quality enhancement method", in which an adaptive separable convolution is applied as a first module to a network model, each two-dimensional convolution is converted into a pair of one-dimensional convolution kernels in the horizontal and vertical directions, and the parameters are represented by n 2 Becomes n + n. Secondly, the self-adaptive change convolution kernels learned by the network for different inputs are utilized to realize motion vector estimation, two continuous frames are selected as network inputs, a pair of separable two-dimensional convolution kernels can be obtained for every two continuous inputs, then the 2-dimensional convolution kernels are expanded into four 1-dimensional convolution kernels, the obtained 1-dimensional convolution kernels are changed along with the change of the inputs, and the network self-adaptability is improved. The invention replaces the two-dimensional convolution kernel with the one-dimensional convolution kernel, so that the parameters of the network training model are reduced, and the execution efficiency is high. The scheme adopts five coding modules, four decoding modules, a separation convolution module and an image prediction module, and the structure is that the last decoding module is replaced by the separation convolution module on the basis of the traditional symmetrical coding and decoding module network, so that the parameters of a model are effectively reduced, and the quality enhancement effect is still to be further improved.
In addition, chinese patent application publication No. CN108307193a proposes an enhancement scheme named as "a multi-frame quality enhancement method and apparatus for lossy compressed video", in which, for an i-th frame of a decompressed video stream, m frames associated with the i-th frame are adopted to perform quality enhancement on the i-th frame, so as to play the i-th frame after quality enhancement; the m frames belong to frames in the video stream, and each frame in the m frames and the i frame respectively have the same or corresponding pixel number which is larger than a preset threshold value; m is a natural number greater than 1. In a particular application, a peak quality frame may be utilized to enhance a non-peak quality frame between two peak quality frames. The scheme reduces the quality fluctuation among multiple frames in the video stream playing process, and simultaneously enhances the quality of each frame in the video after lossy compression. Although this scheme takes into account the temporal information between adjacent frames, the multi-frame convolutional neural network (MF-CNN) is designed to be divided into a motion compensation sub-network (MC-sub-network) and a quality enhancement sub-network (QE-sub-network), where the motion compensation sub-network relies heavily on optical flow estimation to compensate for motion between non-peak quality frames and peak quality frames to achieve alignment, any error in optical flow calculation can introduce artifacts around the image structure in the aligned adjacent frames. However, accurate optical flow estimation is inherently challenging and time consuming, and therefore the quality enhancement effect of the invention is still limited.
Disclosure of Invention
The invention aims to enhance a low-quality frame by fully utilizing time information between adjacent frames by adopting a multi-frame quality enhancement network without optical flow estimation, so that the subjective and objective quality of a compressed video is enhanced.
The embodiment of the invention provides a compressed video multi-frame quality enhancement method needing motion compensation, which comprises the following steps:
to-be-enhanced compressed video, forming an input sequence by a low-quality frame of the compressed video sequence and two adjacent high-quality frames of the compressed video sequence;
inputting each input sequence into a quality enhancement network to obtain an enhanced low-quality frame of the current input sequence;
the quality enhancement network comprises a preprocessing module and an enhancement module, wherein the preprocessing module comprises a feature extraction module and a feature fusion module, and the feature extraction module is used for extracting the spatial feature of each frame of an input sequence to obtain a feature map of each frame; the fusion module is used for fusing event information between adjacent frames to obtain a first fusion feature map and inputting the first fusion feature map into the enhancement module;
the enhancement module comprises a nonlinear mapping module and a reconstruction module, wherein the nonlinear mapping module is used for carrying out nonlinear mapping on the first fusion feature map to obtain a second fusion feature and inputting the second fusion feature into the reconstruction module; the reconstruction module is formed by stacking at least two layers of convolution layers and is used for predicting an enhanced residual image, and then the enhanced low-quality frame is obtained based on fusion of the enhanced residual image and the low-quality frame of the input sequence.
The embodiment of the invention improves the quality of the compressed video by fully utilizing the time information between frames. Meanwhile, the quality enhancement network adopted by the embodiment of the invention does not need to utilize optical flow estimation to compensate the motion between adjacent frames in a displaying way, thereby simplifying the network structure and simplifying the network training.
In one possible implementation, the feature extraction module is a feature extraction network based on a multi-scale feature extraction strategy to extract rich feature information.
In one possible implementation, the nonlinear mapping module is a nonlinear mapping module based on a layered residual and a channel attention mechanism.
Further, the nonlinear mapping module includes a plurality of layered residual modules hr_block, where the hr_block includes: the method comprises a downsampling layer, a residual block R, a residual block RA, a convolution layer and an upsampling layer, wherein the residual block R comprises at least two layers of convolution layers, and the residual block RA is as follows: adding a channel attention mechanism module CA_block into the residual block R to extract information of residual errors of layered characteristics on different channels, wherein the residual block RA comprises two paths of outputs and defines the residual error output as a first output of the residual block RA, and the CA_block output of the residual block RA is a second output of the residual block RA; the characteristic diagram of the input HR_block is expressed as Z, and the characteristic diagram Z is obtained through a first residual block S The method comprises the steps of carrying out a first treatment on the surface of the Meanwhile, the feature map Z obtains a feature map ZD through a downsampling layer and inputs the feature map ZD into a second residual block, and obtains the feature map based on the first output of the second residual blockObtaining a feature map based on a second output of the second residual block>The characteristic diagram->And->Obtaining a sign map by an up-sampling layer with the same structure>And->Feature map +.>And after being spliced with ZS according to the channel, the ZS is input into a convolution layer to obtain a characteristic diagram HR (Z) output by a layered residual error module.
Further, the nonlinear mapping module comprises three hr_blocks and two channel attention mechanism modules ca_blocks, and sequentially defines the three hr_blocks as a first hr_block, a second hr_block and a third hr_block according to the forward propagation direction of the network, and sequentially defines the two ca_blocks as a first ca_block and a second ca_block; the input feature map of the first HR_block is a feature map Z, the feature map HR (Z) output by the first HR_block is subtracted from the feature map Z and then input into a second HR_block, the feature map HR (Z) output by the second HR_block is added with the feature map Z and then input into a first CA_block, and the input feature map of the first CA_block is added with the output feature map of the first CA_block and then input into a third HR_block; the feature map Z, the feature map HR (Z) output by the first HR_block and the feature map input by the third HR_block are subjected to splicing layer to obtain a third fusion feature map; and outputting a feature map of the first hr_blockAn output characteristic diagram of the first CA_block and a characteristic diagram of the third HR_block output +.>Obtaining a fourth fusion feature map through the splicing layer; and finally, adding the third fusion characteristic diagram passing through one convolution layer and the four fusion characteristic diagrams passing through one convolution layer and one second CA_block in sequence to obtain an output characteristic diagram of the nonlinear mapping module.
In one possible implementation, the network structure of the ca_block is: the method comprises the steps of sequentially connecting a pooling layer, at least 1 convolution block, a convolution layer and a sigmoid activation function, wherein the convolution block comprises the convolution layer and the activation function thereof.
The technical scheme provided by the embodiment of the invention at least has the following beneficial effects:
in the embodiment of the invention, the low-quality frames are enhanced by fully utilizing the adjacent high-quality information frames, so that the subjective and objective quality of the compressed video is obviously enhanced, and the contradiction between the quality (definition and color richness) of the video in the multimedia application and the limited network bandwidth is solved, thereby ensuring that the multimedia application based on the invention can ensure the normal transmission of the high-definition video in the network after compression under the same code rate and also ensure better subjective and objective quality.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and that other drawings can be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic diagram of a quality enhancement network used in a method for enhancing quality of a compressed video multi-frame without motion compensation according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a feature extraction network employed in an embodiment of the invention;
FIG. 3 is a schematic diagram of a network structure of a nonlinear mapping module according to an embodiment of the present invention;
fig. 4 is a schematic diagram of a network structure of a layered residual module (Hierarchical Residual block, hr_block) according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a network architecture of a channel attention mechanism module (Channel Attention block, CA_block) employed in an embodiment of the present invention;
FIG. 6 is a graph showing the result of enhancement processing of the present invention compared with the conventional comparative scheme in the embodiment of the present invention;
fig. 7 is a graph showing PSNR fluctuation curves after enhancement processing of frames 6 to 36 in a video sequence bqsquad according to the present invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the present invention more apparent, the embodiments of the present invention will be described in further detail with reference to the accompanying drawings.
With the continuous progress of hardware and algorithms, digital video contains an increasing amount of data while resolution is higher and colors are richer. The video compression coding can effectively reduce the video data volume, but inevitably leads to the reduction of subjective and objective quality, and in order to enhance the subjective and objective quality of the compressed video, the embodiment of the invention provides a multi-frame quality enhancement method of the compressed video without motion compensation. Before describing embodiments of the present invention, related terms and nouns are annotated as follows:
265/HEVC: is a new video coding standard formulated after H.264, retains some techniques of the original H.264 coding standard, and improves some techniques at the same time. The new technique used is to improve the relationship between the code stream, the coding quality, the delay and the algorithm complexity in order to achieve an optimal setting.
GOP, group of pictures (group of pictures): refers to the distance between two I frames.
I-frame, intra-coded picture (Intra-coded image frame): the other image frames are not referred to, and only the information of the present frame is used for encoding.
P-frame, predictive-coded picture (Predictive-coded image frame): and performing inter-frame prediction coding by using a previous I frame or P frame in a motion prediction mode.
Low Delay P (LDP): only the first frame is I-frame encoded, while the others are P-frame encoded.
Peak Signal to Noise Ratio (PSNR): peak signal to noise ratio, an objective criterion for evaluating images.
Structural Similarity (SSIM): the structural similarity is a full-reference image quality evaluation index, and measures the image similarity from three aspects of brightness, contrast and structure respectively.
Ringing effect: for strong edges in the image, moire is generated around the edges after decoding due to quantization distortion of the high frequency ac coefficients, which is called ringing effect.
PQF: peak quality frames, i.e., high quality frames in a GOP, may also be considered I frames in a GOP.
non-PQF: non-peak quality frames, i.e., low quality frames in a GOP, can also be considered P-frames in a GOP.
In one possible implementation manner, the method for enhancing the multi-frame quality of the compressed video without motion compensation provided by the embodiment of the invention comprises the following steps:
to enhance compressed video, one low quality frame of the compressed video sequence and two adjacent high quality frames thereof are formed into an input sequence, e.g. the input sequence is defined as X, where x= { X -1 ,x 0 ,x +1 },x 0 Low quality frames representing the input sequence (i.e. original frames y 0 Compressed frame x of (2) 0 ),x -1 And x +1 Representation and x 0 Adjacent high quality frames, such that the currently input low quality frames are enhanced with the aid of their adjacent high quality frames;
inputting each input sequence into a quality enhancement network to obtain an enhanced low-quality frame of the current input sequence;
the quality enhancement network includes a preprocessing module (Pre-processing Module) and an enhancement module (Enhancement Module), as shown in fig. 1;
the preprocessing module comprises a feature extraction module (Feature Extraction Module) and a feature fusion module (Feature Fusion Module), wherein the feature extraction module is used for extracting the spatial features of each frame of an input sequence to obtain a feature map of each frame; the fusion module is used for fusing event information between adjacent frames to obtain a first fusion feature map and inputting the first fusion feature map into the enhancement module;
the enhancement module comprises a nonlinear mapping module (Non-linear Mapping Module) and a reconstruction module (Reconstruction Module), wherein the nonlinear mapping module is used for carrying out nonlinear mapping on the first fusion feature map to obtain a second fusion feature and inputting the second fusion feature into the reconstruction module; the reconstruction module is formed by stacking convolution layers (at least comprising two layers of convolution layers) and is used for predicting an enhanced residual image and obtaining an enhanced low-quality frame based on fusion of the enhanced residual image and the low-quality frame of the input sequence.
The object of the embodiment of the invention is to compress the frame x according to the original frame y0 0 Deducing frames of high qualityI.e. enhanced low quality frames. Extracting and fusing input x by preprocessing module -1 、x 0 And x +1 Is then reduced by an enhancement module 0 Compression artifact of (a) to get->It is thus possible to achieve that the quality of low quality frames in compressed video is enhanced by making full use of the information of its neighboring high quality frames.
In one possible implementation manner, in the embodiment of the present invention, the adopted feature extraction module is a feature extraction network based on a multi-scale feature extraction policy.
Referring to fig. 2, in the embodiment of the present invention, a multi-layer convolutional feature extraction network, for example, a 4-layer structure is adopted, a ReLU activation function is set after each of the first three layers of convolutional layers, and according to the forward propagation direction of the network, the convolution kernel sizes of the convolutional layers are sequentially set as follows: 3×3,5×5,7×7,3×3.
The feature extraction network is used to extract each frame X in the input sequence X t To obtain a feature map T (x t ):
T(x t )=W t *M(S(x t );W M ,B M )+b t ;
S(x t )=ReLU(Conv 1 (x t ));
Wherein Conv 1 Representing the first convolution layer in the feature extraction network, M representing a weight W M Offset to B M The sign 'x' refers to convolution operation after feature splicing of different scales, W t And b t Representing the weight and bias of its convolutional layers, respectively.
In one possible implementation, the feature fusion module obtains a fused feature, i.e. a first fused feature map F (x 0 ):
F(x 0 )=Conv([T(x -1 ),T(x 0 ),T(x +1 )]),
Wherein the number of the components is [. Cndot. ],]representing the splicing operation, T (x -1 ),T(x 0 ),T(x +1 ) Respectively represent x -1 、x 0 And x +1 Is a feature map of (1).
In the enhancement processing module, a nonlinear mapping module is first used to calculate a first fused feature map F (x 0 ) Is a more useful representation of U (x 0 ) Obtaining a second fusion profile U (x 0 ). Then, U (x 0 ) Input into a reconstruction module consisting of at least 2 convolutions of layers to predict (learn) an enhanced residual map R (x) 0 ). Note that the enhanced residual map refers to the difference between the original frame and the enhanced frame. Finally, the learned enhanced residual picture (x 0 ) X with input 0 Enhanced low quality frame after element-wise summationI.e. < ->
In a possible implementation, the nonlinear mapping module is constructed based on a layered residual and a channel attention mechanism, i.e. the nonlinear mapping module is a nonlinear mapping module based on a layered residual and a channel attention mechanism.
Further, in the embodiment of the present invention, the nonlinear mapping module includes a plurality of layered residual modules (Hierarchicai Residual block, hr_block) and a channel attention mechanism module (Channel Attention block, ca_block), and referring to fig. 3 and fig. 4, the network structure of the hr_block includes: the network structure of the hr_block used is introduced based on the input feature map and its processing output.
For ease of representation, the feature map of the input hr_block is represented as Z (i.e., U (x 0 )). First, Z is input to a Residual block consisting of at least 2 convolutional layers and a Relu-activated function to obtain Z S :
Z S =Conv(ReLU(Conv(ReLU(Z))))+Z;
Then, the feature map is subjected to layering operation through a downsampling layer to obtain a downsampled feature map Z D And then the characteristic diagram Z D Is input into a Residual block, but the convolution operation adopts bottleneck convolution and adds a CA_block to the Residual block to extract information (expressed as) This Residual block to which the CA_block is added is referred to as RA_block. The output of RA_blcok is denoted +.>Will then->Obtaining a characteristic diagram after an up-sampling layer>Finally, the feature map->And Z is S After being spliced according to the channel (i.e. passing through a Concat layer) is input into aThe convolution layers of the number obtain a characteristic diagram HR (Z) output by the layered residual module:
when needing to be noted, feature mapThe feature map is also obtained via the same upsampling layer +.>As input to the subsequent network.
Referring to fig. 3, as a possible implementation manner, in the embodiment of the present invention, a preferred network structure of the nonlinear mapping module includes 3 hr_blocks and 2 ca_blocks, where an input feature map of a first hr_block is Z, an output feature map HR (Z) of the first hr_block is subtracted from the feature map Z, and then a second hr_block is input, and after the feature map HR (Z) of the second hr_block is added to the feature map Z, a first ca_block is input, and after the input feature map of the first ca_block is added to the output feature map thereof, a third hr_block is input; then fusing the feature map Z, the feature map HR (Z) output by the first HR_block and the feature map input by the third HR_bloc through a Concat layer to obtain a third fused feature map; and outputting a feature map of the first hr_blockAn output characteristic diagram of the first CA_block and a characteristic diagram of the third HR_block output +.>A fourth fusion feature map is obtained after Concat layer fusion (channel splicing); and finally, adding the third fusion feature map passing through one convolution layer and the four fusion feature maps passing through one convolution layer and one second CA_block in sequence to obtain an output feature map of the nonlinear mapping module.
Preferably, the convolution kernel size of the convolution layer employed in the Residual block is 5×5, and the convolution kernel size of the convolution layer outputting the feature map HR (Z) is 3×3.
Preferably, the downsampling strategy of the downsampling layer may be described in the literature "Sajjadi M S M, vemulapalli R, brown M.frame-recurrent video super-resolution [ C ]]v/Proceedings of the IEEE Conference on Computer Vision and Pattern Reconnaist.2018: 6626-6634 "or" Sun W, he X, chen H, et a1.A quality enhancement framework with noisedistribution characteristics for high efficiency video coding [ J ]]Neuroompuying, 2020, 411:428-441", the present invention does not specifically limit the downsampling strategy of the downsampling layer. Likewise, the upsampling strategy of the upsampling layer is not particularly limited, and the literature "Shi W, cabellero J, husz r F, et al real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network [ C ]]/(Proceedings of the IEEE conference on computer vision and pattern recoganation.2016): 1874-188", e.g. for a special group graphThe up-sampled ad hoc graph thereof can be expressed as:
In one possible implementation, referring to fig. 5, the structure of ca_block used includes a pooling layer, a convolution layer and its activation function (layer number adjustable), a convolution layer and a sigmoid activation function connected in sequence. The output of the sigmoid activation function is multiplied by the output of CA_block to obtain the output of CA_block.
To facilitate description of the hierarchy, v= [ V ] is defined 1 ,v 2 ,...,v c ]To represent the input feature map of CA_block, which is composed of C (number of input channels) feature maps v with the space size H×W i Composition is prepared. The feature graphs are subjected to dimension reduction in the space dimension through a pooling layer which is an average pooling (average pooling) in a pooling mode, so that a pooling result of a c-th channel is obtained:
wherein v is i (i, j) is the value of the c-th channel at position (i, j), e c The value after the c-th channel average is shown. A sigmoid activation function is then used as a gating mechanism to derive the most relevant feature map:
α c =σ(Conv(ReLU(Conv(e c ))))
wherein σ represents a sigmoid activation function, defining E= [ E 1 ,e 2 ,...,e c ]. Finally based on channel weight alpha= [ alpha ] 1 ,α 2 ,...,α C ]To readjust the interdependence between the channels of the input profile V:
in a preferred manner, the number of input channels of the CA_block is C, and the number of output channels of the first convolution layer of the CA_block is set to beThe core size of the pooling layer of CA_block is 1×1, and the convolution core size is 1×1.
In the embodiment of the invention, the training of the network parameters of the adopted quality enhancement network can be performed based on the conventional training mode of the neural network parameters, and as a preferable mode, the preprocessing module and the enhancement module of the quality enhancement network provided by the embodiment of the invention are jointly trained in an end-to-end mode, and the network does not need to firstly train and converge a certain sub-network, so that the loss function only consists of one item. For example, using L 2 Norm as a loss function of the network:
in an exemplary embodiment, the present invention also provides a computer device including a processor and a memory having at least one computer program stored therein. The at least one computer program is loaded and executed by one or more processors to implement any of the methods for compressed video multi-frame quality enhancement described above without motion compensation.
In an exemplary embodiment, the present invention further provides a computer readable storage medium having at least one computer program stored therein, the at least one computer program being loaded and executed by a processor of a computer device to implement any of the above methods for compressed video multi-frame quality enhancement without motion compensation.
In one possible implementation, the computer readable storage medium may be a Read-Only Memory (ROM), a random access Memory (Random Access Memory, RAM), a Read-Only Memory (CD-ROM), a magnetic tape, an optical data storage device, and the like.
It should be noted that the terms "first," "second," and the like in the description and in the claims are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It should be understood that the data so used may be interchanged where appropriate such that the description of the embodiments of the invention herein described is not intended to represent all the embodiments consistent with the invention. The idea is that they are merely examples of apparatuses and methods consistent with some aspects of the invention as detailed in the appended claims.
In order to qualitatively and quantitatively evaluate the enhancement performance of the multi-frame quality enhancement method for compressed video without motion compensation provided by the embodiment of the present invention, the quality enhancement network shown in fig. 1 to 5 is adopted in the present embodiment to perform quality enhancement processing on the selected compressed video to be enhanced, and the quantitative evaluation is compared with the existing enhancement schemes DCAD, DS-CNN, MFQE and MFQE 2.0 according to Δpsnr and Δssim, as shown in tables 1 and 2; qualitative assessment was compared to MFQE 2.0, as shown in fig. 6. Among them, enhancement scheme DCAD can be referred to as "Wang T, chen M, chao H.A novel deep learning-based method of improving coding efficiency from the decoder-end for HEVC [ C ]//2017Data Compression Conference (DCC) & lt/EN & gt, 2017:410-419", enhancement scheme DS-CNN can be referred to as "Yang R, xu M, wang Z.Decoder-side HEVC quality enhancement with scalable convolutional neural network [ C ]//201 2017IEEE International Conference on Multimedia and Expo (ICME) & lt/EN & gt, IEEE,2017:817-822", enhancement discard MFQE can be referred to as "Yang R, xu M, wang Z, et al, multi-frame quality enhancement for compressed video [ C ]// Proceedings of the IEEE Conference on Computer Vision and Pattern Recognity.2018:6664-6673".
TABLE 1
TABLE 2
Table 1 gives the average results of Δpsnr and Δssim over all frames of each test sequence, Δpsnr (dB) and Δssim for the HEVC standard test sequence at a total of five QP points ((×10) -4 ) Is a comparison of the population of (c). It can be seen that the enhancement scheme (outer) provided by the embodiment of the present invention is always superior to other video quality enhancement methods. Specifically, at qp=37, the maximum Δpsnr value of the present embodiment reaches 1.154dB, the average Δpsnr is 0.632dB, 12.5% higher than MFQE 2.0 (0.562 dB), 38.9% higher than MFQE (0.455 dB), 96.3% higher than DCAD (0.322 dB), and 110.7% higher than DS-CNN (0.300 dB). At other QP points, the embodiments of the present invention are also superior to other methods in both ΔPSNR and ΔSSIM. In addition, the decrease in BD-rate was used to compare the performance of the network, as shown in Table 2, with an average decrease in BD-rate of 17.18% for our network, which is superior to the currently best MFQE 2.0 (14.06%).
Table 2 gives the BD-rate (%) for the test sequence as compared to the HEVC reference drop, calculated at five points qp=22, 27, 32, 37 and 42.
Fig. 6 shows subjective quality performance of sequences BasketballPass, partyScene and BQMall at qp=37. As can be seen from the figure, compared with the MFQE 2.0 method, the enhancement scheme provided by the embodiment of the invention can reduce more compression artifacts, and better visual experience is realized.
For lossless video, after compression, there will be a certain quality fluctuation between video frames, as shown in fig. 7 (line corresponding to HEVC), and the PSNR fluctuation between high quality frames and low quality frames is obvious in the enhancement scheme MFQE 2.0 (see "Guan Z, xing Q, xu M, et al MFQE 2.0:A new approach for multi-frame quality enhancement on compressed video [ J ]. IEEE transactions on pattern analysis and machine intelligence, 2019") after using adjacent high quality frames to enhance middle low quality frames, as shown in fig. 7. Meanwhile, the PSNR fluctuation of the scheme (outer) provided by the embodiment of the present invention is also shown in fig. 7. It can be seen that the enhancement method provided by the embodiments of the present invention better utilizes the information of the high quality frames than the existing MFQE 2.0, exhibiting lower enhancement quality fluctuations. And enhancement scheme MFQE 2.0 is a modification of the enhancement scheme of publication No. CN108307193 a. Namely, the compressed video multi-frame quality enhancement network without motion compensation provided by the embodiment of the invention enhances the low-quality frames by fully utilizing the adjacent high-quality information frames, so that the subjective quality and the objective quality of the compressed video are obviously enhanced.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and are not limiting; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.
What has been described above is merely some embodiments of the present invention. It will be apparent to those skilled in the art that various modifications and improvements can be made without departing from the spirit of the invention.
Claims (7)
1.A method of motion compensation-free multi-frame quality enhancement of compressed video, comprising:
to-be-enhanced compressed video, forming an input sequence by a low-quality frame of the compressed video sequence and two adjacent high-quality frames of the compressed video sequence;
inputting each input sequence into a quality enhancement network to obtain an enhanced low-quality frame of the current input sequence;
the quality enhancement network comprises a preprocessing module and an enhancement module, wherein the preprocessing module comprises a feature extraction module and a feature fusion module, and the feature extraction module is used for extracting the spatial feature of each frame of an input sequence to obtain a feature map of each frame; the fusion module is used for fusing event information between adjacent frames to obtain a first fusion feature map and inputting the first fusion feature map into the enhancement module;
the enhancement module comprises a nonlinear mapping module and a reconstruction module, wherein the nonlinear mapping module is used for carrying out nonlinear mapping on the first fusion feature map to obtain a second fusion feature and inputting the second fusion feature into the reconstruction module; the reconstruction module is formed by stacking at least two layers of convolution layers and is used for predicting an enhanced residual error map, and then an enhanced low-quality frame is obtained based on fusion of the enhanced residual error map and the low-quality frame of the input sequence;
the nonlinear mapping module is a nonlinear mapping module based on a layered residual and a channel attention mechanism, the nonlinear mapping module comprises a plurality of layered residual modules hr_blocks, the hr_blocks comprise: the method comprises a downsampling layer, a residual block R, a residual block RA, a convolution layer and an upsampling layer, wherein the residual block R comprises at least two layers of convolution layers, and the residual block RA is as follows: adding a channel attention mechanism module CA_block into the residual block R to extract information of residual errors of layered characteristics on different channels, wherein the residual block RA comprises two paths of outputs and defines the residual error output as a first output of the residual block RA, and the CA_block output of the residual block RA is a second output of the residual block RA; feature map to be input to hr_blockDenoted as Z, the feature map Z is obtained by a first residual block S The method comprises the steps of carrying out a first treatment on the surface of the At the same time, the feature map Z is obtained through a downsampling layer D And input into a second residual block, and obtain a feature map based on the first output of the second residual blockObtaining a feature map based on a second output of the second residual block>The characteristic diagram->And->Obtaining a sign map by an up-sampling layer with the same structure>And->Feature map +.>And Z is S And (3) inputting the characteristic images into a convolution layer after channel splicing to obtain a characteristic image HR (Z) output by a layered residual error module.
2. The method of claim 1, wherein the feature extraction module is a feature extraction network based on a multi-scale feature extraction strategy.
3. The method of claim 1, wherein the convolution kernel size of the convolution layer employed in the residual block R is 5 x 5 and the convolution kernel size of the last layer of the hr_block is 3 x 3.
4. The method of claim 1, wherein the nonlinear mapping module comprises three hr_blocks and two channel attention mechanism modules ca_blocks, and wherein the three hr_blocks are sequentially defined as a first hr_block, a second hr_block, and a third hr_block, and the two ca_blocks are sequentially defined as a first ca_block and a second ca_block according to a forward propagation direction of the network;
the input feature map of the first HR_block is a feature map Z, the feature map HR (Z) output by the first HR_block is subtracted from the feature map Z, then a second HR_block is input, the feature map HR (Z) output by the second HR_block is added with the feature map Z, then a first CA_block is input, and the input feature map of the first CA_block is added with the output feature map of the first CA_block and then a third HR_block is input; the feature map Z, the feature map HR (Z) output by the first HR_block and the feature map input by the third HR_block are subjected to splicing layer to obtain a third fusion feature map; and outputting a feature map of the first hr_blockAn output characteristic diagram of the first CA_block and a characteristic diagram of the third HR_block output +.>Obtaining a fourth fusion feature map through the splicing layer; and finally, adding the third fusion characteristic diagram passing through one convolution layer and the four fusion characteristic diagrams passing through one convolution layer and one second CA_block in sequence to obtain an output characteristic diagram of the nonlinear mapping module. />
5. The method according to any one of claims 1 to 4, wherein the network structure of the ca_block is: the method comprises the steps of sequentially connecting a pooling layer, at least 1 convolution block, a convolution layer and a sigmoid activation function, wherein the convolution block comprises the convolution layer and the activation function thereof.
7. The method of claim 2, wherein the feature extraction network is a 4-layer convolution structure, and a ReLU activation function is set after each of the first three convolution layers, and according to the forward propagation direction of the network, the convolution kernel sizes of the convolution layers are set as follows: 3×3,5×5,7×7,3×3.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110654128.XA CN113507607B (en) | 2021-06-11 | 2021-06-11 | Compressed video multi-frame quality enhancement method without motion compensation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110654128.XA CN113507607B (en) | 2021-06-11 | 2021-06-11 | Compressed video multi-frame quality enhancement method without motion compensation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113507607A CN113507607A (en) | 2021-10-15 |
CN113507607B true CN113507607B (en) | 2023-05-26 |
Family
ID=78009890
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110654128.XA Active CN113507607B (en) | 2021-06-11 | 2021-06-11 | Compressed video multi-frame quality enhancement method without motion compensation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113507607B (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108307193A (en) * | 2018-02-08 | 2018-07-20 | 北京航空航天大学 | A kind of the multiframe quality enhancement method and device of lossy compression video |
CN108989731A (en) * | 2018-08-09 | 2018-12-11 | 复旦大学 | A method of improving video spatial resolution |
US10701394B1 (en) * | 2016-11-10 | 2020-06-30 | Twitter, Inc. | Real-time video super-resolution with spatio-temporal networks and motion compensation |
CN111524068A (en) * | 2020-04-14 | 2020-08-11 | 长安大学 | Variable-length input super-resolution video reconstruction method based on deep learning |
CN112381866A (en) * | 2020-10-27 | 2021-02-19 | 天津大学 | Attention mechanism-based video bit enhancement method |
CN112613356A (en) * | 2020-12-07 | 2021-04-06 | 北京理工大学 | Action detection method and device based on deep attention fusion network |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9195903B2 (en) * | 2014-04-29 | 2015-11-24 | International Business Machines Corporation | Extracting salient features from video using a neurosynaptic system |
EP3259911B1 (en) * | 2015-02-19 | 2021-04-07 | Magic Pony Technology Limited | Enhancing visual data using updated neural networks |
-
2021
- 2021-06-11 CN CN202110654128.XA patent/CN113507607B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10701394B1 (en) * | 2016-11-10 | 2020-06-30 | Twitter, Inc. | Real-time video super-resolution with spatio-temporal networks and motion compensation |
CN108307193A (en) * | 2018-02-08 | 2018-07-20 | 北京航空航天大学 | A kind of the multiframe quality enhancement method and device of lossy compression video |
CN108989731A (en) * | 2018-08-09 | 2018-12-11 | 复旦大学 | A method of improving video spatial resolution |
CN111524068A (en) * | 2020-04-14 | 2020-08-11 | 长安大学 | Variable-length input super-resolution video reconstruction method based on deep learning |
CN112381866A (en) * | 2020-10-27 | 2021-02-19 | 天津大学 | Attention mechanism-based video bit enhancement method |
CN112613356A (en) * | 2020-12-07 | 2021-04-06 | 北京理工大学 | Action detection method and device based on deep attention fusion network |
Also Published As
Publication number | Publication date |
---|---|
CN113507607A (en) | 2021-10-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109309834B (en) | Video compression method based on convolutional neural network and HEVC compression domain significant information | |
Zhang et al. | Machine learning based video coding optimizations: A survey | |
CN111260560B (en) | Multi-frame video super-resolution method fused with attention mechanism | |
CN113055674B (en) | Compressed video quality enhancement method based on two-stage multi-frame cooperation | |
CN113592746B (en) | Method for enhancing quality of compressed video by fusing space-time information from coarse to fine | |
CN110852964A (en) | Image bit enhancement method based on deep learning | |
CN1695381A (en) | Sharpness enhancement in post-processing of digital video signals using coding information and local spatial features | |
CN112381866B (en) | Attention mechanism-based video bit enhancement method | |
CN110177282B (en) | Interframe prediction method based on SRCNN | |
Ma et al. | CVEGAN: a perceptually-inspired gan for compressed video enhancement | |
CN117730338A (en) | Video super-resolution network and video super-resolution, encoding and decoding processing method and device | |
CN114173131B (en) | Video compression method and system based on inter-frame correlation | |
CN116916036A (en) | Video compression method, device and system | |
CN115442613A (en) | Interframe information-based noise removal method using GAN | |
CN113507607B (en) | Compressed video multi-frame quality enhancement method without motion compensation | |
CN114827616B (en) | Compressed video quality enhancement method based on space-time information balance | |
CN114511485B (en) | Compressed video quality enhancement method adopting cyclic deformable fusion | |
Cui et al. | Convolutional neural network-based post-filtering for compressed YUV420 images and video | |
CN113691817B (en) | Cross-frame information fusion screen content video quality enhancement method | |
Ulas et al. | Flexible luma-chroma bit allocation in learned image compression for high-fidelity sharper images | |
CN115002482A (en) | End-to-end video compression method and system using structural preservation motion estimation | |
CN112819707B (en) | End-to-end anti-blocking effect low-illumination image enhancement method | |
Jung | Comparison of video quality assessment methods | |
CN114554213B (en) | Motion adaptive and detail-focused compressed video quality enhancement method | |
Ding et al. | Blind Quality Enhancement for Compressed Video |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |