CN112616052B - Method for reconstructing video compression signal - Google Patents

Method for reconstructing video compression signal Download PDF

Info

Publication number
CN112616052B
CN112616052B CN202011461038.0A CN202011461038A CN112616052B CN 112616052 B CN112616052 B CN 112616052B CN 202011461038 A CN202011461038 A CN 202011461038A CN 112616052 B CN112616052 B CN 112616052B
Authority
CN
China
Prior art keywords
image
key frame
residual
frame
key
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011461038.0A
Other languages
Chinese (zh)
Other versions
CN112616052A (en
Inventor
周涛
李琛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai IC R&D Center Co Ltd
Shanghai IC Equipment Material Industry Innovation Center Co Ltd
Original Assignee
Shanghai IC R&D Center Co Ltd
Shanghai IC Equipment Material Industry Innovation Center Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai IC R&D Center Co Ltd, Shanghai IC Equipment Material Industry Innovation Center Co Ltd filed Critical Shanghai IC R&D Center Co Ltd
Priority to CN202011461038.0A priority Critical patent/CN112616052B/en
Publication of CN112616052A publication Critical patent/CN112616052A/en
Application granted granted Critical
Publication of CN112616052B publication Critical patent/CN112616052B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Abstract

The invention discloses a reconstruction method of a video compression signal, which comprises the following steps: according to the image observation values of the key frame and the non-key frame output by the acquisition end, carrying out preliminary prediction on the image of the non-key frame to generate an image prediction value of the non-key frame; calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end; calculating the average energy of residual error signals of the observation residual errors of the non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, and reconstructing the predicted residual errors to generate reconstructed residual errors; and calculating an image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame. The reconstruction method of the video compression signal can effectively improve the accuracy of residual error reconstruction, thereby improving the differentiation retention capacity of the video image and finally realizing the high-quality reconstruction of the video image.

Description

Method for reconstructing video compression signal
Technical Field
The present invention relates to the technical field of video compression processing, and in particular, to a method for reconstructing a video compression signal.
Background
The traditional representative video compression coding technology such as H.26x and MPEG is based on the Nyquist sampling theorem, sampling is carried out at a sampling rate which is at least twice as large as the bandwidth of a sampling signal, then high-complexity compression coding is carried out on the video signal, and finally the signal is transmitted and stored.
The compressed sensing theory proposed by Donoho, candes, tao, et al, 2006 provides a completely new signal sampling technique, which indicates that: for sparse or nearly sparse signals in a transform domain, the signal can be sampled at a frequency well below the nyquist sampling rate and reconstructed with great probability accurately. The compression sensing theory is innovative in that on one hand, the bottleneck of the traditional sampling rate is broken through, on the other hand, the compression sensing theory directly samples signals by designing an observation matrix, and the signals obtained by sampling are compressed, so that the sampling and compression are completed synchronously, the waste of sampling resources is effectively avoided, meanwhile, the signal reconstruction with high complexity is performed at a decoding end, and the equipment performance bottleneck of a sampling section can be effectively solved. The compressive sensing technology has attracted wide attention of academic circles at the beginning of its birth, and is also favored in the fields of wireless sensing, multimedia sensing networks and the like in the aspect of practical application.
With the development of the compressive sensing theory, numerous scholars deeply research the video compressive sensing technology and obtain various remarkable results, wherein a video reconstruction distributed compressive sensing framework based on prediction-residual errors is widely adopted, but the current video reconstruction algorithm of the prediction-residual errors only focuses on the improvement of the precision of the prediction algorithm, and the influence of residual error reconstruction on the final prediction performance is ignored. Since the reconstruction of most image frames is closely related to the residual error, if the accuracy of the residual error reconstruction is not high, it is difficult to improve the overall performance of the video compressed sensing algorithm.
Disclosure of Invention
The present invention is directed to overcoming the above-mentioned drawbacks of the prior art and providing a method for reconstructing a compressed video signal.
In order to realize the purpose, the technical scheme of the invention is as follows:
a method of reconstructing a compressed video signal, comprising:
s01: according to the image observation values of the key frame and the non-key frame output by the acquisition end, carrying out preliminary prediction on the image of the non-key frame to generate an image prediction value of the non-key frame;
s02: calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end;
s03: calculating the average energy of residual error signals of the observation residual errors of the non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, reconstructing the predicted residual errors and generating reconstructed residual errors;
s04: and calculating the image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame.
Further, in step S01, the preliminary predicting the image of the non-key frame, and generating the predicted image value of the non-key frame includes:
the image prediction value for the non-key frame is calculated using the following formula:
Figure BDA0002831782380000021
wherein the content of the first and second substances,
Figure BDA0002831782380000022
image prediction values for non-key frames, H p,q Is a component block, w, for image target block estimation p,q Is the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is the weight coefficient p,q Calculated using the following formula:
Figure BDA0002831782380000023
wherein, y k,j Is the image observed value of the non-key frame, lambda is the weighting factor of time prior, is used for adjusting the influence between the associated frames, phi is the random sampling matrix, p1 is the frame index of the image block of the associated frame, and the associated frame comprises the adjacent frame of the non-key frame and the same imageThe key frame in the image group, p2 is the frame index of the image block of the non-associated frame.
Further, λ > =0.5 for static class images, and λ <0.5 for dynamic class images.
Further, in step S02, calculating the observation residual of the non-key frame includes:
performing domain transformation processing on the image prediction value of the non-key frame based on the random sampling matrix to obtain an image prediction value of a transformation domain;
and subtracting the image predicted value of the transform domain from the image observed value of the non-key frame to obtain an observation residual error.
Further, in step S02, calculating the prediction residual of the non-key frame includes:
and performing domain inverse transformation processing on the observation residual error based on the random sampling matrix to obtain a prediction residual error of the non-key frame.
Further, in step S03, calculating the average energy of the residual signal of the observation residual of the non-key frame includes:
calculating the residual signal average energy according to the following formula:
Figure BDA0002831782380000031
wherein R is the average energy of the residual signal,
Figure BDA0002831782380000032
observation residuals for non-key frames->
Figure BDA0002831782380000033
R is the average energy of the residual signal, k is the frame index of the image block, j is the image block index within the frame, L is the index of the non-zero residual signal, and L is the number of non-zero components in the residual signal.
Further, in step S03, selecting an adaptive residual reconstruction algorithm according to the magnitude of the residual signal average energy, and reconstructing the prediction residual of the non-key frame, wherein generating the reconstructed residual of the non-key frame includes:
according to the relation between preset judgment thresholds T1 and T2 and the average energy of the residual error signals, one of the following formulas is selected to calculate the sparse expression coefficient of the reconstructed residual error:
if R is less than or equal to T1, then use
Figure BDA0002831782380000034
/>
If T1 < R ≦ T2, use
Figure BDA0002831782380000035
If R > T2, use
Figure BDA0002831782380000036
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002831782380000037
for reconstructing sparse representation coefficients of the residual error, <' >>
Figure BDA0002831782380000038
Sparse representation coefficient for prediction residue of non-key frame, < > or >>
Figure BDA0002831782380000039
Lambda 1 and lambda 2 are weight factors for balancing the similarity and difference effects between the predicted image and the observed image,
the reconstructed residual of the non-key frame is calculated using the following formula:
Figure BDA0002831782380000041
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002831782380000042
Ψ is a redundant dictionary for the reconstructed residual of the non-key frame.
Further, step S04 includes:
the image reconstruction value of the non-key frame is calculated using the following formula:
Figure BDA0002831782380000043
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002831782380000044
image reconstruction value for a non-key frame, <' >>
Figure BDA0002831782380000045
Is the image prediction value of the non-key frame.
Further, the method further comprises:
and calculating an image reconstruction value of the key frame according to the image observation value of the key frame.
Further, the method further comprises:
respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
and combining the key frame image and the non-key frame image according to the sequence of the frame sequence to generate a reconstructed video image.
In the reconstruction method of the video compression signal, an adaptive residual error reconstruction algorithm is selected based on the average energy of residual error signals in the reconstruction process of non-key frames, and prediction residual errors are subjected to
Figure BDA0002831782380000046
And reconstructing, wherein the difference between the predicted information and the observed information is reflected by the average energy of the residual error signals, and different reconstruction algorithms are selected according to the difference degree, so that the accuracy of residual error reconstruction can be effectively improved, the differentiation retention capacity of the video image is improved, and the high-quality reconstruction of the video image is finally realized.
Drawings
FIG. 1 is a schematic diagram of the overall architecture of the video compression reconstruction algorithm of the present invention;
FIG. 2 is a schematic diagram of a group of pictures in successive frames of a video in accordance with the present invention;
fig. 3 is a flowchart of a method of reconstructing a video compressed signal according to the present invention.
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
In the following detailed description of the embodiments of the present invention, in order to clearly illustrate the structure of the present invention and to facilitate explanation, the structure shown in the drawings is not drawn to a general scale and is partially enlarged, deformed and simplified, so that the present invention should not be construed as limited thereto.
For the convenience of description of the technical scheme, the following part reference numbers are explained as follows:
image observation value: the image signal detected directly by the acquisition-side detector is used as y in the present invention k,j An image observation representing a non-key frame;
image prediction value: image signals predicted based on image observations, used in the invention
Figure BDA0002831782380000051
A picture prediction value representing a non-key frame;
observed residual of non-key frames
Figure BDA0002831782380000052
The difference between the image observation signal and the image prediction signal after the transformation is used in the invention>
Figure BDA0002831782380000053
Represents the observed residual of the non-key frame;
prediction residual of non-key frames
Figure BDA0002831782380000054
The difference between the image observation signal and the image prediction signal after transformation is used in the present invention/>
Figure BDA0002831782380000055
Represents the prediction residual of the non-key frame;
reconstructed residual of non-key frames
Figure BDA0002831782380000056
Observation residual->
Figure BDA0002831782380000057
And combines the signal obtained by the segmented reconstruction algorithm>
Figure BDA0002831782380000058
And then the following formula is used for calculation: />
Figure BDA0002831782380000059
Wherein it is present>
Figure BDA00028317823800000510
For predicting residual non-key frames
Figure BDA00028317823800000511
Sparse representation coefficients of (a);
image reconstruction value: the video signal that finally needs to be presented, i.e. the image signal after decompression, is used in the present invention
Figure BDA00028317823800000512
Representing image reconstruction values of non-key frames.
Referring to fig. 1, it is a schematic diagram of the overall architecture of the video compression reconstruction algorithm of the present invention. The whole process of video compression and reconstruction can be divided into video compression processing at an acquisition end and video reconstruction processing at a reconstruction end. The acquisition end and the reconstruction end can be distributed on different devices, for example, the acquisition end can be arranged on a video server, the reconstruction end can be arranged on a mobile terminal, video signals are compressed at the acquisition end and then transmitted to the mobile terminal through the internet, and the reconstruction end on the mobile terminal reconstructs video data to generate video frames and plays the video frames.
Referring to fig. 2, it is a schematic diagram of the group of pictures in the video consecutive frames of the present invention. At the acquisition end, the video signal is split into continuous image groups, each image group consists of a fixed number of continuous image frames, the first frame of each image group is a key frame, and other frames are non-key frames. At the acquisition end, a block compression sensing algorithm is adopted to compress the key frames and the non-key frames, and the block compression sensing algorithm has the advantages of being less in storage space requirement and capable of well retaining image characteristics.
Taking a non-key frame as an example, assuming that the size of the original image to be sampled is N × N, the image block size may be set to N 1/2 ×N 1/2 If k represents the frame index and j represents the image block index in the frame, the image signal value after sampling the original image is represented as x k,j . For image signal value x k,j After compressed sensing sampling is carried out, an image observation value is obtained and is expressed as y k,j
The sampling process can be expressed by the following equation (1):
y k,j =φx k,j formula (1)
Phi is a random sampling matrix of M multiplied by N, M < < N, namely the observation quantity is far smaller than the quantity of discrete components of the image signal, the sampling rate is M/N, and the random sampling matrix can adopt a Gaussian random matrix, a Hadamard random matrix and the like. The key frames and non-key frames are processed according to the same principle, but at different sampling rates, and a higher M is usually used for the key frames as a reference signal in the corresponding image group to provide a reference for estimation of the image blocks of the non-key frames. The non-key frame adopts lower M, so that higher compression rate is realized, and the load of an acquisition end is reduced. And outputting the image observation value generated after sampling by the sampling end as image compression data of the sampling end.
According to the compressed sensing theory, the image observed value y k,j Can be expressed as the following equation (2):
y k,j =φx k,j =φΨθ k,j formula (2)
φΨCalled the observation matrix, Ψ is a redundant dictionary, θ k,j Is x k,j Sparse representation coefficients under a dictionary Ψ, wherein the redundant dictionary Ψ and the random sampling matrix Φ are both known quantities, and the sampling process of the above formula (1) and formula (2) is equivalent to using Φ to x k,j Performing domain transformation, and distinguishing two domains before and after transformation by using x k,j The domain in which y is located is called the signal domain k,j The domain in which is referred to as the transform domain. The redundant dictionary Ψ is used in the image reconstruction process at the reconstruction end.
And at a reconstruction end, reconstructing the key frame and the non-key frame in different modes. The key frame can be reconstructed by adopting the conventional compressed sensing algorithm, and because the key frame has a higher sampling rate, the key frame can obtain a better effect by adopting a K-SVD (K-singular value decomposition), L1-Norm optimization algorithm, MH-BCS-SPL (hybrid binary coded decimal) algorithm and the like. In the invention, in the process of reconstructing the key frames, the influence of residual energy on the reconstruction of the non-key frames is fully considered, and an adaptive residual reconstruction algorithm is selected according to the size of the residual energy to reconstruct the key frames, wherein the specific processing process is detailed later.
Referring to fig. 3, which is a flowchart of a method for reconstructing a video compression signal according to the present invention, the method may include:
s01: according to the image observation value y of the key frame and the image observation value y of the non-key frame output by the acquisition end k,j Performing preliminary prediction on the image of the non-key frame to generate a predicted image value of the non-key frame
Figure BDA0002831782380000071
Specifically, the image of the non-key frame is preliminarily predicted, and the image prediction value->
Figure BDA0002831782380000072
The processing procedure of (2) may include:
calculating an image prediction value of a non-key frame using the following formula (3)
Figure BDA0002831782380000073
Figure BDA0002831782380000074
Wherein H p,q Is for the image target block x k,j Estimated block of components, w p,q Is the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is p,q Calculated using the following equation (4):
Figure BDA0002831782380000075
wherein, λ is a weighting factor of time prior, and is used for adjusting influence between associated frames. p1 is the frame index of the image block of the associated frame, the associated frame comprises the adjacent frame of the non-key frame and the key frame in the same image group, p2 is the frame index of the image block of the non-associated frame, and the non-associated frame comprises other frames in the same image group except the associated frame. The argmin function is the variable value at which the objective function takes a minimum value.
As can be seen from equation (4), the weight coefficient w p,q Is determined by both the energy prior and the time prior of the image block. Determining a weight coefficient w by minimizing an energy difference between an observed value of an image block and an estimated value of the image block using image energy as prior information of image similarity p,q Meanwhile, the time correlation is included in the formula (4), because of the continuity of the video image, the target frame and the associated frame generally have high similarity, in order to enhance the influence of the associated frame, the energy difference of the associated frame is calculated independently, the first term of the formula (4) is formed, the second term of the formula (4) is the energy difference of the non-associated frame, and the influence of the associated frame is adjusted through lambda. In the present invention, the p1 and p2 cores are used for distinguishing the same position in the same image groupThe image blocks influence the size of the target image blocks, the key frames and the associated frames have more reference information on the corresponding image blocks of the target frames, and the target frames of each non-key frame in the same image group use the key frames as references, so that the key frames are used as the associated frames, namely the key frames are classified into p1 instead of p2. The λ is a preset value, and the weighting factor λ can be set to different values according to different video image types, and for still images, such as a shot image for a still or an image with little change between frames, the λ>=0.5, for dynamic images, e.g. high speed moving objects in the image or the entire frame of image is dynamically changing, λ<0.5。
S02: image prediction values from the non-key frames
Figure BDA0002831782380000081
And image observations y of said non-key frames k,j And a random sampling matrix phi used by the acquisition end, calculating the observation residual of the non-key frame->
Figure BDA0002831782380000082
And prediction residue of non-key frame->
Figure BDA0002831782380000083
Wherein, in step S02, the observation residual of the non-key frame is calculated
Figure BDA0002831782380000084
The method can comprise the following steps:
image prediction values for non-key frames based on the random sampling matrix phi
Figure BDA0002831782380000085
Performing domain transformation to obtain the image prediction value in the transformation domain>
Figure BDA0002831782380000086
The domain transformation processing referred to herein is a transformation from a signal domain to a transform domain based on a random sampling matrix phi, and may specifically beThe following formula (5) is used.
Figure BDA0002831782380000087
Then, the image prediction value of the transform domain is calculated
Figure BDA0002831782380000088
Image observation value y of non-key frame k,j Subtract to obtain the observed residual ^ of the non-key frame>
Figure BDA0002831782380000089
Furthermore, in step S02, the prediction residual of the non-key frame is calculated
Figure BDA00028317823800000810
The method can comprise the following steps:
based on the random sampling matrix phi, observing residual errors
Figure BDA00028317823800000811
Performing inverse domain transform to obtain prediction residue ^ of non-key frame>
Figure BDA00028317823800000812
The domain inverse transform processing here refers to transformation from a transform domain to a signal domain based on a random sampling matrix phi, and may specifically adopt the following formula (6):
Figure BDA00028317823800000813
wherein phi is -1 Is the inverse of the random sampling matrix phi.
S03: computing observed residuals for the non-key frames
Figure BDA0002831782380000091
And selecting the adapted residual according to the magnitude of the average energy of the residual signalDifference reconstruction algorithm for predicting residual->
Figure BDA0002831782380000092
Performs a reconstruction to generate a reconstructed residual ≥ of non-key frames>
Figure BDA0002831782380000093
In step S03, the observation residual of the non-key frame is calculated
Figure BDA0002831782380000094
The average energy of the residual signal of (2) includes: calculating the residual signal average energy according to the following formula (7):
Figure BDA0002831782380000095
wherein, R is the average energy of the residual signal, k is the frame index of the image block, j is the index of the image block in the frame, L is the index of the non-zero residual signal, and L is the number of the non-zero components in the residual signal.
In step S03, an adaptive residual error reconstruction algorithm is selected according to the average energy of the residual error signal, and the prediction residual error is subjected to
Figure BDA0002831782380000096
Performs a reconstruction to generate a reconstructed residual ≥ of non-key frames>
Figure BDA0002831782380000097
The method comprises the following steps: />
According to the relation between preset judgment threshold values T1 and T2 and the average energy R of the residual error signal, one of the following formulas is selected, and the sparse expression coefficient of the reconstructed residual error is calculated
Figure BDA0002831782380000098
If R is less than or equal to T1, then calculate using equation (8)
Figure BDA0002831782380000099
Figure BDA00028317823800000910
If T1 < R ≦ T2, then use equation (9) to calculate
Figure BDA00028317823800000911
Figure BDA00028317823800000912
If R > T2, then calculate using equation (10)
Figure BDA00028317823800000913
Figure BDA00028317823800000914
Wherein, in
Figure BDA00028317823800000915
Prediction residue for non-key frame->
Figure BDA00028317823800000916
Sparse representation coefficient of (4), based on the expression value of (4)>
Figure BDA00028317823800000917
For observing residual errors, lambda 1 and lambda 2 are weight factors for balancing the effects of similarity and difference between the predicted image and the observed image, the decision thresholds T1 and T2 are related to the average atomic energy of the dictionaries, and the decision thresholds T1 and T2 are in direct proportion to the average atomic energy of the redundant dictionary Ψ.
R is less than or equal to T1 corresponding to the situation that the average energy of residual signals is lower, which shows that the similarity between the predicted information and the observed information is high, namely the predicted value of the image in the transform domain
Figure BDA00028317823800000918
And the image observation value y k,j And the similarity is high, and at the moment, 2 norms are adopted to inhibit residual high-frequency information (difference between observation and prediction), so that the consistency of the images is ensured.
T1 < R < T2 corresponds to the situation that the average energy of residual signals is higher, which shows that the prediction information and the observation information have obvious difference, namely the image prediction value of the transform domain
Figure BDA0002831782380000101
And the image observed value y k,j A relatively significant difference occurs, at which point the combined 2-norm and 1-norm is used to balance the similar and different parts.
R > T2 corresponds to the situation that the average energy of residual signals is extremely high, and shows that the difference between the predicted information and the observed value is very obvious, namely the predicted value of the image in the transform domain
Figure BDA0002831782380000102
And the image observation value y k,j The difference is very obvious, and at the moment, the difference effect is enhanced by adopting a 1-mode.
Then, the reconstructed residual of the non-key frame is calculated using the following formula (11)
Figure BDA0002831782380000103
Figure BDA0002831782380000104
Where Ψ is a redundant dictionary.
In the residual reconstruction processing in step S03 described above, the features of the sparse representation coefficients of the prediction residual are sufficiently considered. And evaluating the characteristics of sparse expression coefficients of the predicted residual based on the prior information of positive correlation between the difference between the predicted information and the observed information and the average energy value of the residual signal, thereby selecting different residual reconstruction algorithms to realize high-quality reconstruction effect.
S04: image prediction values from the non-key frames
Figure BDA0002831782380000105
And reconstructed residual ≥ of non-key frames>
Figure BDA0002831782380000106
Calculating an image reconstruction value ≥ of non-key frames>
Figure BDA0002831782380000107
Specifically, step S04 may calculate the image reconstruction value ≦ for the non-key frame using equation (12) below>
Figure BDA0002831782380000108
Figure BDA0002831782380000109
Wherein the content of the first and second substances,
Figure BDA00028317823800001010
is the image prediction value of the non-key frame.
Through the above steps S01 to S04, the image reconstruction value of the non-key frame is calculated, and the non-key frame image can be generated based on the image reconstruction value
In addition, the video compression and reconstruction algorithm of the present invention may further include an image reconstruction process of a key frame, and specifically, the video compression and reconstruction algorithm of the present invention may further include:
s05: and calculating an image reconstruction value of the key frame according to the image observation value of the key frame. The process of calculating the image reconstruction values of the key frames may be performed in parallel with the calculation of the image reconstruction values of the non-key frames of the previous steps S01 to S04, or may be performed sequentially as shown in fig. 3. As described above, the key frame reconstruction process can be implemented using the existing methods of K-SVD, L1-Norm optimization algorithm, MH-BCS-SPL, etc.
After the image reconstruction values of the key frame and the non-key frame are calculated, the complete video image reconstruction can be performed, and specifically, the video compression reconstruction algorithm of the invention can further include:
s06: respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
s07: and combining the key frame images and the non-key frame images according to the sequence of the frame sequences to generate reconstructed video images.
The above description is only a preferred embodiment of the present invention, and the embodiments are not intended to limit the scope of the present invention, so that all equivalent structural changes made by using the contents of the specification and the drawings of the present invention should be included in the scope of the present invention.

Claims (10)

1. A method for reconstructing a compressed video signal, comprising:
s01: the method comprises the steps that a video signal is split into a plurality of image groups, the first frame of each image group is a key frame, the rest frames are non-key frames, according to image observation values of the key frame and the non-key frames output by an acquisition end, the images of the non-key frames are subjected to preliminary prediction, an image prediction value of the non-key frames is generated, and the sampling rate of the image observation values of the key frames is greater than that of the image observation values of the non-key frames;
s02: calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end;
s03: calculating the average energy of residual error signals of the observation residual errors of the non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, reconstructing the predicted residual errors and generating reconstructed residual errors;
s04: and calculating the image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame.
2. The method of claim 1, wherein the step S01 of performing preliminary prediction on the image of the non-key frame and generating the predicted image value of the non-key frame comprises:
the image prediction value for the non-key frame is calculated using the following formula:
Figure FDA0003795430400000011
wherein the content of the first and second substances,
Figure FDA0003795430400000012
image prediction values for non-key frames, H p,q Is a component block, w, for image target block estimation p,q Is the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is the weight coefficient p,q Calculated using the following formula:
Figure FDA0003795430400000013
wherein, y k,j The image observation value of the non-key frame is obtained, lambda is a time prior weight factor used for adjusting the influence between the associated frames, phi is a random sampling matrix, p1 is the frame index of the image block of the associated frame, the associated frame comprises the adjacent frame of the non-key frame and the key frame in the same image group, and p2 is the frame index of the image block of the non-associated frame.
3. The method of claim 2, wherein λ > =0.5 for still-like images and λ <0.5 for moving-like images.
4. The method for reconstructing a compressed video signal according to any of claims 1 to 3, wherein in step S02, calculating the observation residuals of the non-key frames comprises:
performing domain transformation processing on the image prediction value of the non-key frame based on the random sampling matrix to obtain the image prediction value of a transformation domain;
and subtracting the image predicted value of the transform domain from the image observed value of the non-key frame to obtain an observation residual error.
5. The method of claim 4, wherein in step S02, calculating the prediction residual of the non-key frame comprises:
and performing domain inverse transformation processing on the observation residual error based on the random sampling matrix to obtain a prediction residual error of the non-key frame.
6. The method of claim 5, wherein the step S03 of calculating the average energy of residual signals of the observation residuals of the non-key frames comprises:
calculating the residual signal average energy according to the following formula:
Figure FDA0003795430400000021
wherein R is the average energy of the residual signal,
Figure FDA0003795430400000022
observed residual->
Figure FDA0003795430400000023
R is the average energy of the residual signal, k is the frame index of the image block, j is the index of the image block within the frame, L is the index of the non-zero residual signal, and L is the number of non-zero components in the residual signal.
7. The method according to claim 6, wherein in step S03, an adaptive residual reconstruction algorithm is selected according to the average energy of the residual signals, and the prediction residual of the non-key frame is reconstructed, and generating the reconstructed residual of the non-key frame comprises:
according to the relation between preset judgment thresholds T1 and T2 and the average energy of the residual error signals, one of the following formulas is selected to calculate the sparse expression coefficient of the reconstructed residual error:
if R is less than or equal to T1, then use
Figure FDA0003795430400000024
If T1 is more than R and less than or equal to T2, the use is carried out
Figure FDA0003795430400000025
If R > T2, use
Figure FDA0003795430400000031
Wherein the content of the first and second substances,
Figure FDA0003795430400000032
for reconstructing sparse representation coefficients of the residual error, <' >>
Figure FDA0003795430400000033
Sparse representation coefficients for the prediction residual of non-key frames,
Figure FDA0003795430400000034
is the observation residual of the non-key frame, lambda 1 and lambda 2 are weight factors for balancing the similarity and difference effects between the predicted image and the observation image,
the reconstructed residual of the non-key frame is calculated using the following formula:
Figure FDA0003795430400000035
wherein the content of the first and second substances,
Figure FDA0003795430400000036
for reconstructed residual of non-key frames, Ψ is redundancyAnd (4) remaining dictionaries.
8. The method of claim 7, wherein step S04 comprises:
calculating an image reconstruction value of a non-key frame using the following formula
Figure FDA0003795430400000037
Figure FDA0003795430400000038
Wherein the content of the first and second substances,
Figure FDA0003795430400000039
image reconstruction values for non-key frames>
Figure FDA00037954304000000310
Is the image prediction value of the non-key frame.
9. The method of reconstructing a compressed video signal according to claim 8, further comprising:
and calculating an image reconstruction value of the key frame according to the image observation value of the key frame.
10. The method for reconstructing a compressed video signal according to claim 9, further comprising:
respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
and combining the key frame images and the non-key frame images according to the sequence of the frame sequences to generate reconstructed video images.
CN202011461038.0A 2020-12-11 2020-12-11 Method for reconstructing video compression signal Active CN112616052B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011461038.0A CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011461038.0A CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Publications (2)

Publication Number Publication Date
CN112616052A CN112616052A (en) 2021-04-06
CN112616052B true CN112616052B (en) 2023-03-28

Family

ID=75234409

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011461038.0A Active CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Country Status (1)

Country Link
CN (1) CN112616052B (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107360426A (en) * 2017-07-13 2017-11-17 福州大学 A kind of video sequence reconstructing method based on compressed sensing

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050078099A (en) * 2004-01-30 2005-08-04 삼성전자주식회사 Video coding apparatus and method for inserting key frame adaptively
EP2202984B1 (en) * 2007-10-25 2017-04-19 Nippon Telegraph and Telephone Corporation Video scalable encoding method and decoding methods using weighted prediction, devices therefor, programs therefor, and recording medium where program is recorded
CN103297782B (en) * 2013-06-08 2016-04-27 河海大学常州校区 Based on the reconstructing method of Region dividing in distributed video compression perceptual system
WO2015180052A1 (en) * 2014-05-28 2015-12-03 北京大学深圳研究生院 Video coding and decoding methods and apparatuses based on dictionary database
CN107155112A (en) * 2017-05-24 2017-09-12 湖北工业大学 A kind of compressed sensing method for processing video frequency for assuming prediction more
CN108347612B (en) * 2018-01-30 2020-09-15 东华大学 Monitoring video compression and reconstruction method based on visual attention mechanism
CN110933429B (en) * 2019-11-13 2021-11-12 南京邮电大学 Video compression sensing and reconstruction method and device based on deep neural network

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107360426A (en) * 2017-07-13 2017-11-17 福州大学 A kind of video sequence reconstructing method based on compressed sensing

Also Published As

Publication number Publication date
CN112616052A (en) 2021-04-06

Similar Documents

Publication Publication Date Title
US10812790B2 (en) Data processing apparatus and data processing method
CN104199627B (en) Gradable video encoding system based on multiple dimensioned online dictionary learning
CN106960420B (en) Image reconstruction method of segmented iterative matching tracking algorithm
Zhou et al. Image compression based on discrete cosine transform and multistage vector quantization
CN112616052B (en) Method for reconstructing video compression signal
Bhatnagar et al. Image compression using dct based compressive sensing and vector quantization
Yu et al. Medical image compression with thresholding denoising using discrete cosine-based discrete orthogonal stockwell transform
Ali et al. Image and audio fractal compression: Comprehensive review, enhancements and research directions
CN113096019B (en) Image reconstruction method, image reconstruction device, image processing equipment and storage medium
CN115361556A (en) High-efficiency video compression algorithm based on self-adaption and system thereof
Upadhyaya et al. Quality parameter index estimation for compressive sensing based sparse audio signal reconstruction
Thepade et al. New clustering algorithm for Vector Quantization using Slant transform
Thepade et al. New clustering algorithm for Vector Quantization using Haar sequence
Thepade et al. New clustering algorithm for vector quantization using hybrid Haar slant error vector
Rahman et al. An integer wavelet transform based lossless image compression technique using arithmetic coding
Kumar et al. Comparative Analysis and Performance Evaluation of Medical Image Compression Method for Telemedicine
Prabhavathi et al. Compressive Sensing and its Application to Speech Signal Processing
Wang et al. Reduced dimension Vector Quantization encoding method for image compression
Chatterjee et al. Image compression and resizing using vector quantization and other efficient algorithms
Yadav et al. A Survey of Image Compression using Neural Network and Wavelet Transformation
Shahriar et al. Image reconstruction via compressed sensing
Prudvi et al. Analysis of Image Compression Techniques for IoT Applications
Zhang et al. Application of Sparse Dictionary Adaptive Compression Algorithm in Transient Signals
SAHNOUN et al. The Fourier transform for satellite image compression
Zhang An image reconstruction algorithm based on classified block

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant