CN112616052A - Method for reconstructing video compression signal - Google Patents

Method for reconstructing video compression signal Download PDF

Info

Publication number
CN112616052A
CN112616052A CN202011461038.0A CN202011461038A CN112616052A CN 112616052 A CN112616052 A CN 112616052A CN 202011461038 A CN202011461038 A CN 202011461038A CN 112616052 A CN112616052 A CN 112616052A
Authority
CN
China
Prior art keywords
image
key frame
residual
value
key
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011461038.0A
Other languages
Chinese (zh)
Other versions
CN112616052B (en
Inventor
周涛
李琛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai IC R&D Center Co Ltd
Shanghai IC Equipment Material Industry Innovation Center Co Ltd
Original Assignee
Shanghai IC R&D Center Co Ltd
Shanghai IC Equipment Material Industry Innovation Center Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai IC R&D Center Co Ltd, Shanghai IC Equipment Material Industry Innovation Center Co Ltd filed Critical Shanghai IC R&D Center Co Ltd
Priority to CN202011461038.0A priority Critical patent/CN112616052B/en
Publication of CN112616052A publication Critical patent/CN112616052A/en
Application granted granted Critical
Publication of CN112616052B publication Critical patent/CN112616052B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention discloses a reconstruction method of a video compression signal, which comprises the following steps: according to the image observation values of the key frame and the non-key frame output by the acquisition end, carrying out preliminary prediction on the image of the non-key frame to generate an image prediction value of the non-key frame; calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end; calculating the average energy of residual error signals of observation residual errors of non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, reconstructing the predicted residual errors and generating reconstructed residual errors; and calculating the image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame. The method for reconstructing the video compression signal can effectively improve the accuracy of residual error reconstruction, thereby improving the differentiation retention capacity of the video image and finally realizing high-quality reconstruction of the video image.

Description

Method for reconstructing video compression signal
Technical Field
The present invention relates to the technical field of video compression processing, and in particular, to a method for reconstructing a video compression signal.
Background
The traditional representative video compression coding technology such as H.26x and MPEG is based on the Nyquist sampling theorem, sampling is carried out at a sampling rate which is at least twice as large as the bandwidth of a sampling signal, then high-complexity compression coding is carried out on the video signal, and finally the signal is transmitted and stored.
The compressed sensing theory proposed by Donoho, Candes, Tao, et al, 2006 provides a completely new signal sampling technique, which indicates that: for sparse or nearly sparse signals in a transform domain, the signal can be sampled at a frequency well below the nyquist sampling rate and reconstructed with great probability accurately. The compression sensing theory is innovative in that on one hand, the bottleneck of the traditional sampling rate is broken through, on the other hand, the compression sensing theory directly samples signals by designing an observation matrix, and the signals obtained by sampling are compressed, so that the sampling and compression are completed synchronously, the waste of sampling resources is effectively avoided, meanwhile, the signal reconstruction with high complexity is performed at a decoding end, and the equipment performance bottleneck of a sampling section can be effectively solved. The compressive sensing technology has attracted wide attention of academic circles at the beginning of its birth, and is also favored in the fields of wireless sensing, multimedia sensing networks and the like in the aspect of practical application.
With the development of the compressive sensing theory, numerous scholars deeply research the video compressive sensing technology and obtain various remarkable results, wherein a video reconstruction distributed compressive sensing framework based on prediction-residual errors is widely adopted, but the current video reconstruction algorithm of the prediction-residual errors only focuses on the improvement of the precision of the prediction algorithm, and the influence of residual error reconstruction on the final prediction performance is ignored. Since the reconstruction of most image frames is closely related to the residual error, if the accuracy of residual error reconstruction is not high, the overall performance of the video compressed sensing algorithm is difficult to improve.
Disclosure of Invention
The present invention is directed to overcoming the above-mentioned drawbacks of the prior art and providing a method for reconstructing a compressed video signal.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a method of reconstructing a compressed video signal, comprising:
s01: according to the image observation values of the key frame and the non-key frame output by the acquisition end, carrying out preliminary prediction on the image of the non-key frame to generate an image prediction value of the non-key frame;
s02: calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end;
s03: calculating the average energy of residual error signals of the observation residual errors of the non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, reconstructing the predicted residual errors and generating reconstructed residual errors;
s04: and calculating the image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame.
Further, in step S01, the performing preliminary prediction on the image of the non-key frame, and generating the image prediction value of the non-key frame includes:
the image prediction value for the non-key frame is calculated using the following formula:
Figure BDA0002831782380000021
wherein the content of the first and second substances,
Figure BDA0002831782380000022
image prediction values for non-key frames, Hp,qIs a component block, w, for image target block estimationp,qIs the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is the weight coefficientp,qCalculated using the following formula:
Figure BDA0002831782380000023
wherein, yk,jThe image observation value of the non-key frame is represented by lambda, the weight factor of the time prior is represented by lambda, the influence between the associated frames is adjusted by phi, the random sampling matrix is represented by phi, p1 is the frame index of the image blocks of the associated frames, the associated frames comprise the adjacent frames of the non-key frame and the key frames in the same image group, and p2 is the frame index of the image blocks of the non-associated frames.
Further, λ > is 0.5 for static class images and λ <0.5 for dynamic class images.
Further, in step S02, calculating the observation residual of the non-key frame includes:
performing domain transformation processing on the image prediction value of the non-key frame based on the random sampling matrix to obtain the image prediction value of a transformation domain;
and subtracting the image predicted value of the transform domain from the image observed value of the non-key frame to obtain an observation residual error.
Further, in step S02, the calculating the prediction residual of the non-key frame includes:
and performing domain inverse transformation processing on the observation residual error based on the random sampling matrix to obtain a prediction residual error of the non-key frame.
Further, in step S03, the calculating the average energy of the residual signal of the observation residual of the non-key frame includes:
calculating the residual signal average energy according to the following formula:
Figure BDA0002831782380000031
wherein R is the average energy of the residual signal,
Figure BDA0002831782380000032
observed residual for non-key frames
Figure BDA0002831782380000033
R is the average energy of the residual signal, k is the frame index of the image block, j is the image block index within the frame, L is the index of the non-zero residual signal, and L is the number of non-zero components in the residual signal.
Further, in step S03, selecting an adaptive residual reconstruction algorithm according to the magnitude of the residual signal average energy, and reconstructing the prediction residual of the non-key frame, wherein generating the reconstructed residual of the non-key frame includes:
according to the relation between preset judgment threshold values T1 and T2 and the average energy of the residual error signal, one of the following formulas is selected to calculate the sparse expression coefficient of the reconstructed residual error:
if R is less than or equal to T1, then use
Figure BDA0002831782380000034
If T1 < R ≦ T2, then use
Figure BDA0002831782380000035
If R > T2, use
Figure BDA0002831782380000036
Wherein the content of the first and second substances,
Figure BDA0002831782380000037
in order to reconstruct the sparse representation coefficients of the residual,
Figure BDA0002831782380000038
sparse representation coefficients of the prediction residual for non-key frames,
Figure BDA0002831782380000039
is the observation residual of the non-key frame, lambda 1 and lambda 2 are weight factors for balancing the similarity and difference effects between the predicted image and the observation image,
the reconstructed residual of the non-key frame is calculated using the following formula:
Figure BDA0002831782380000041
wherein the content of the first and second substances,
Figure BDA0002831782380000042
Ψ is a redundant dictionary for the reconstructed residual of the non-key frame.
Further, step S04 includes:
the image reconstruction value of the non-key frame is calculated using the following formula:
Figure BDA0002831782380000043
wherein the content of the first and second substances,
Figure BDA0002831782380000044
for the image reconstruction values of non-key frames,
Figure BDA0002831782380000045
is the image prediction value of the non-key frame.
Further, the method further comprises:
and calculating an image reconstruction value of the key frame according to the image observation value of the key frame.
Further, the method further comprises:
respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
and combining the key frame images and the non-key frame images according to the sequence of the frame sequences to generate reconstructed video images.
In the reconstruction method of the video compression signal, an adaptive residual error reconstruction algorithm is selected based on the average energy of residual error signals in the reconstruction process of non-key frames, and prediction residual errors are subjected to
Figure BDA0002831782380000046
And reconstructing, wherein the difference between the prediction information and the observation information is reflected by the average energy of the residual error signals, and different reconstruction algorithms are selected according to the difference degree, so that the accuracy of residual error reconstruction can be effectively improved, the differentiation retention capability of the video image is improved, and the high-quality reconstruction of the video image is finally realized.
Drawings
FIG. 1 is a schematic diagram of the overall architecture of the video compression reconstruction algorithm of the present invention;
FIG. 2 is a schematic diagram of a group of pictures in successive frames of a video in accordance with the present invention;
fig. 3 is a flowchart of a method of reconstructing a video compressed signal according to the present invention.
Detailed Description
The following describes embodiments of the present invention in further detail with reference to the accompanying drawings.
In the following detailed description of the embodiments of the present invention, in order to clearly illustrate the structure of the present invention and to facilitate explanation, the structure shown in the drawings is not drawn to a general scale and is partially enlarged, deformed and simplified, so that the present invention should not be construed as limited thereto.
For the convenience of description of the technical scheme, the following part reference numbers are explained as follows:
image observation value: the image signal detected directly by the acquisition-side detector is used as y in the present inventionk,jAn image observation representing a non-key frame;
image prediction value: image signal predicted based on image observation value and used in the invention
Figure BDA0002831782380000051
Representing image predictors of non-key frames;
observed residual of non-key frames
Figure BDA0002831782380000052
The difference between the image observation signal and the image prediction signal after transformation is used in the present invention
Figure BDA0002831782380000053
Represents the observed residual of the non-key frame;
prediction residual of non-key frames
Figure BDA0002831782380000054
The difference between the image observation signal and the image prediction signal after transformation is used in the present invention
Figure BDA0002831782380000055
Represents the prediction residual of the non-key frame;
reconstructed residual of non-key frames
Figure BDA0002831782380000056
Observed residual through non-key frames
Figure BDA0002831782380000057
And combining signals obtained by a segmented reconstruction algorithm
Figure BDA0002831782380000058
And then the following formula is used for calculation:
Figure BDA0002831782380000059
wherein the content of the first and second substances,
Figure BDA00028317823800000510
for predicting residual non-key frames
Figure BDA00028317823800000511
Sparse representation coefficients of (a);
image reconstruction value: the video signal that is finally to be presented, i.e. the image signal after decompression, is used in the present invention
Figure BDA00028317823800000512
Representing image reconstruction values of non-key frames.
Referring to fig. 1, it is a schematic diagram of the overall architecture of the video compression reconstruction algorithm of the present invention. The whole process of video compression and reconstruction can be divided into video compression processing at an acquisition end and video reconstruction processing at a reconstruction end. The acquisition end and the reconstruction end can be distributed on different devices, for example, the acquisition end can be arranged on a video server, the reconstruction end can be arranged on a mobile terminal, video signals are compressed at the acquisition end and then transmitted to the mobile terminal through the internet, and the reconstruction end on the mobile terminal reconstructs video data to generate video frames and plays the video frames.
Referring to fig. 2, it is a schematic diagram of the group of pictures in the video consecutive frames of the present invention. The video signal is split into continuous image groups at the acquisition end, each image group consists of a fixed number of continuous image frames, the first frame of each image group is a key frame, and other frames are non-key frames. At the acquisition end, a block compression sensing algorithm is adopted to compress the key frames and the non-key frames, and the block compression sensing algorithm has the advantages of being less in storage space requirement and capable of well retaining image characteristics.
For example, assuming the size of the original image to be sampled is N × N, the image block size may be set to N1/2×N1/2When k represents the frame index and j represents the image block index in the frame, the image signal value after sampling the original image is represented as xk,j. For image signal value xk,jAfter compressed sensing sampling is carried out, an image observation value is obtained and is expressed as yk,j
The sampling process can be expressed by the following formula (1):
yk,j=φxk,jformula (1)
Phi is a random sampling matrix of M multiplied by N, M < < N, namely the observation quantity is far smaller than the quantity of discrete components of the image signal, the sampling rate is M/N, and the random sampling matrix can adopt a Gaussian random matrix, a Hadamard random matrix and the like. The processing principle adopted for key frames and non-key frames is the same, except that the sampling rate is different, and a higher M is usually adopted for key frames to serve as a reference signal in the corresponding image group to provide a reference for estimation of non-key frame image blocks. The non-key frame adopts lower M, so that higher compression rate is realized, and the load of an acquisition end is reduced. And outputting the image observation value generated after sampling by the sampling end as image compression data of the sampling end.
According to the compressed sensing theory, the image observed value yk,jCan be expressed as the following equation (2):
yk,j=φxk,j=φΨθk,jformula (2)
Phi psi is called the observation matrix, psi is the redundant dictionary, thetak,jIs xk,jSparse representation coefficients under a dictionary Ψ, wherein the redundant dictionary Ψ and the random sampling matrix Φ are both known quantities, and the sampling process of the above formula (1) and formula (2) is equivalent to using Φ to xk,jPerforming domain transformation, and distinguishing two domains before and after transformation by using xk,jThe domain in which y is located is called the signal domaink,jThe domain in which is referred to as the transform domain. The redundant dictionary Ψ is used in the image reconstruction process at the reconstruction end.
And at a reconstruction end, reconstructing the key frame and the non-key frame in different modes. The key frames can be reconstructed by adopting the conventional compressed sensing algorithm, and because the key frames have higher sampling rate, the key frames can obtain better effect by adopting a K-SVD (K-singular value decomposition), an L1-Norm optimization algorithm, an MH-BCS-SPL and the like. In the invention, in the process of reconstructing the key frames, the influence of residual energy on the reconstruction of the non-key frames is fully considered, and an adaptive residual reconstruction algorithm is selected according to the size of the residual energy to reconstruct the key frames, wherein the specific processing process is detailed later.
Referring to fig. 3, which is a flowchart of a method for reconstructing a video compression signal according to the present invention, the method may include:
s01: according to the image observation value y of the key frame and the image observation value y of the non-key frame output by the acquisition endk,jPerforming preliminary prediction on the image of the non-key frame to generate a predicted image value of the non-key frame
Figure BDA0002831782380000071
Specifically, the image of the non-key frame is preliminarily predicted to generate the predicted image value of the non-key frame
Figure BDA0002831782380000072
The processing procedure of (2) may include:
calculating an image prediction value of a non-key frame using the following formula (3)
Figure BDA0002831782380000073
Figure BDA0002831782380000074
Wherein Hp,qIs for the image target block xk,jEstimated block of components, wp,qIs the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is the weight coefficientp,qCalculated using the following equation (4):
Figure BDA0002831782380000075
wherein, λ is a weighting factor of time prior, and is used for adjusting influence between associated frames. p1 is the frame index of the image block of the associated frame, the associated frame includes the adjacent frame of the non-key frame and the key frame in the same image group, p2 is the frame index of the image block of the non-associated frame, the non-associated frame includes other frames in the same image group except the aforementioned associated frame. The argmin function is the variable value at which the objective function takes a minimum value.
As can be seen from equation (4), the weight coefficient wp,qIs determined by both the energy prior and the time prior of the image block. Determining a weight coefficient w by minimizing an energy difference between an observed value of an image block and an estimated value of the image block using image energy as prior information of image similarityp,qMeanwhile, the time correlation is included in the formula (4), because of the continuity of the video image, the target frame and the associated frame generally have high similarity, in order to enhance the influence of the associated frame, the energy difference of the associated frame is calculated independently, the first term of the formula (4) is formed, the second term of the formula (4) is the energy difference of the non-associated frame, and the influence of the associated frame is adjusted through lambda. In the invention, the core of the method is to distinguish the influence of the image blocks at the same position in the same image group on the target image block through the p1 and the p2, the key frame and the associated frame have more reference information on the corresponding image block of the target frame, and the target frame of each non-key frame in the same image group uses the key frame as a reference, so that the key frame is used as the associated frame, namely the key frame is classified into p1 instead of p 2. The λ is a preset value, and the weighting factor λ can be set to different values according to different video image types, and for still images, such as a shot image for a still or an image with little change between frames, the λ>For a dynamic image, for example, a high-speed moving object exists in the image or the whole frame of image is dynamically changed, λ<0.5。
S02: image prediction values from the non-key frames
Figure BDA0002831782380000081
And image observations y of said non-key framesk,jAnd calculating the observation residual error of the non-key frame by using a random sampling matrix phi used by the acquisition end
Figure BDA0002831782380000082
And prediction residual of non-key frames
Figure BDA0002831782380000083
In step S02, the observation residual of the non-key frame is calculated
Figure BDA0002831782380000084
The method can comprise the following steps:
image prediction values for non-key frames based on the random sampling matrix phi
Figure BDA0002831782380000085
Performing domain transformation to obtain predicted image value of transform domain
Figure BDA0002831782380000086
The domain transformation processing referred to herein is a transformation from a signal domain to a transform domain based on a random sampling matrix phi, and may specifically employ the following formula (5).
Figure BDA0002831782380000087
Then, the image prediction value of the transform domain is calculated
Figure BDA0002831782380000088
Image observation value y of non-key framek,jSubtracting to obtain the observed residual error of the non-key frame
Figure BDA0002831782380000089
In addition, in step S02, the prediction residual of the non-key frame is calculated
Figure BDA00028317823800000810
The method can comprise the following steps:
based on the random sampling matrixPhi, for the observed residual error
Figure BDA00028317823800000811
Performing inverse domain transform to obtain the prediction residual of the non-key frame
Figure BDA00028317823800000812
The domain inverse transform processing here refers to transformation from a transform domain to a signal domain based on a random sampling matrix phi, and may specifically adopt the following formula (6):
Figure BDA00028317823800000813
wherein phi is-1Is the inverse of the random sampling matrix phi.
S03: computing observed residuals for the non-key frames
Figure BDA0002831782380000091
And selecting an adaptive residual reconstruction algorithm according to the average energy of the residual signal, and predicting the residual
Figure BDA0002831782380000092
Reconstructing to generate reconstructed residual error of non-key frame
Figure BDA0002831782380000093
In step S03, the observation residual of the non-key frame is calculated
Figure BDA0002831782380000094
The residual signal average energy of (a) includes: calculating the residual signal average energy according to the following formula (7):
Figure BDA0002831782380000095
wherein, R is the average energy of the residual signal, k is the frame index of the image block, j is the image block index within the frame, L is the index of the non-zero residual signal, and L is the number of non-zero components in the residual signal.
In step S03, an adaptive residual reconstruction algorithm is selected according to the average energy of the residual signal, and the prediction residual is subjected to
Figure BDA0002831782380000096
Reconstructing to generate reconstructed residual error of non-key frame
Figure BDA0002831782380000097
The method comprises the following steps:
according to the relationship between preset judgment threshold values T1 and T2 and the average energy R of the residual error signal, one of the following formulas is selected, and the sparse expression coefficient of the reconstructed residual error is calculated
Figure BDA0002831782380000098
If R is ≦ T1, then it is calculated using equation (8)
Figure BDA0002831782380000099
Figure BDA00028317823800000910
If T1 < R ≦ T2, then calculate using equation (9)
Figure BDA00028317823800000911
Figure BDA00028317823800000912
If R > T2, then calculate using equation (10)
Figure BDA00028317823800000913
Figure BDA00028317823800000914
Wherein, in
Figure BDA00028317823800000915
Prediction residual for non-key frames
Figure BDA00028317823800000916
The sparse representation coefficient of (a) is,
Figure BDA00028317823800000917
for observing the residual error, λ 1 and λ 2 are weight factors for balancing the effects of similarity and difference between the predicted image and the observed image, the decision thresholds T1 and T2 are related to the mean atomic energy of the dictionaries, and the decision thresholds T1 and T2 are proportional to the mean atomic energy of the redundant dictionary Ψ.
R is less than or equal to T1, corresponding to the situation that the average energy of residual signals is lower, the similarity between the predicted information and the observed information is high, namely the predicted value of the image in the transform domain
Figure BDA00028317823800000918
And the image observed value yk,jThe method has high similarity, and residual high-frequency information (difference between observation and prediction) is restrained by adopting 2 norms, so that the consistency of the images is ensured.
T1 < R ≦ T2 corresponds to the situation that the average energy of the residual signal is higher, which indicates that the difference between the predicted information and the observed information is more obvious, i.e. the predicted value of the image in the transform domain
Figure BDA0002831782380000101
And the image observed value yk,jA relatively significant difference occurs, at which point the combined 2-norm and 1-norm is used to balance the similar and different parts.
R > T2 corresponds to the situation that the average energy of residual signals is extremely high, which shows that the difference between the prediction information and the observed value is very obvious, namely the predicted value of the image in the transform domain
Figure BDA0002831782380000102
And the image observed value yk,jThe difference is very obvious, and at the moment, the difference effect is enhanced by adopting a 1-mode.
Then, the reconstructed residual of the non-key frame is calculated using the following equation (11)
Figure BDA0002831782380000103
Figure BDA0002831782380000104
Where Ψ is a redundant dictionary.
In the residual reconstruction processing in step S03 described above, the features of the sparse representation coefficients of the prediction residual are fully considered. And evaluating the characteristics of sparse expression coefficients of the predicted residual based on the prior information of positive correlation between the difference between the predicted information and the observed information and the average energy value of the residual signal, thereby selecting different residual reconstruction algorithms to realize high-quality reconstruction effect.
S04: image prediction values from the non-key frames
Figure BDA0002831782380000105
And reconstructed residual of non-key frames
Figure BDA0002831782380000106
Computing image reconstruction values for non-key frames
Figure BDA0002831782380000107
Specifically, step S04 may calculate an image reconstruction value of a non-key frame using the following formula (12)
Figure BDA0002831782380000108
Figure BDA0002831782380000109
Wherein the content of the first and second substances,
Figure BDA00028317823800001010
is the image prediction value of the non-key frame.
Through the above steps S01 to S04, the image reconstruction value of the non-key frame is calculated, and the non-key frame image can be generated based on the image reconstruction value
In addition, the video compression and reconstruction algorithm of the present invention may further include an image reconstruction process of a key frame, and specifically, the video compression and reconstruction algorithm of the present invention may further include:
s05: and calculating an image reconstruction value of the key frame according to the image observation value of the key frame. The process of calculating the image reconstruction values of the key frames may be performed in parallel with the calculation of the image reconstruction values of the non-key frames of the previous steps S01-S04, or may be performed sequentially as shown in fig. 3. As described above, the key frame reconstruction process can be implemented using the existing methods of K-SVD, L1-Norm optimization algorithm, MH-BCS-SPL, etc.
After the image reconstruction values of the key frame and the non-key frame are calculated, the complete video image reconstruction can be performed, and specifically, the video compression reconstruction algorithm of the invention can further include:
s06: respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
s07: and combining the key frame images and the non-key frame images according to the sequence of the frame sequences to generate reconstructed video images.
The above description is only a preferred embodiment of the present invention, and the embodiments are not intended to limit the scope of the present invention, so that all equivalent structural changes made by using the contents of the specification and the drawings of the present invention should be included in the scope of the present invention.

Claims (10)

1. A method for reconstructing a compressed video signal, comprising:
s01: according to the image observation values of the key frame and the non-key frame output by the acquisition end, carrying out preliminary prediction on the image of the non-key frame to generate an image prediction value of the non-key frame;
s02: calculating an observation residual error of the non-key frame and a prediction residual error of the non-key frame according to the image prediction value of the non-key frame, the image observation value of the non-key frame and a random sampling matrix used by an acquisition end;
s03: calculating the average energy of residual error signals of the observation residual errors of the non-key frames, selecting an adaptive residual error reconstruction algorithm according to the energy of the residual error signals, reconstructing the predicted residual errors and generating reconstructed residual errors;
s04: and calculating the image reconstruction value of the non-key frame according to the image prediction value of the non-key frame and the reconstruction residual error of the non-key frame.
2. The method for reconstructing a compressed video signal as claimed in claim 1, wherein the step S01 of performing preliminary prediction on the image of the non-key frame and generating the predicted image value of the non-key frame comprises:
the image prediction value for the non-key frame is calculated using the following formula:
Figure FDA0002831782370000011
wherein the content of the first and second substances,
Figure FDA0002831782370000012
image prediction values for non-key frames, Hp,qIs a component block, w, for image target block estimationp,qIs the weight coefficient corresponding to the component block, k, p are the frame index of the image block, k corresponds to the non-key frame in the same image group, p corresponds to the key frame and the non-key frame in the same image group, j, q are the image block index in the frame, the weight coefficient w is the weight coefficientp,qCalculated using the following formula:
Figure FDA0002831782370000013
wherein, yk,jThe image observation value of the non-key frame is represented by lambda, the weight factor of the time prior is represented by lambda, the influence between the associated frames is adjusted by phi, the random sampling matrix is represented by phi, p1 is the frame index of the image blocks of the associated frames, the associated frames comprise the adjacent frames of the non-key frame and the key frames in the same image group, and p2 is the frame index of the image blocks of the non-associated frames.
3. A method of reconstructing a compressed video signal as claimed in claim 2, wherein λ > is 0.5 for still type pictures and λ <0.5 for motion type pictures.
4. The method for reconstructing a compressed video signal according to any one of claims 1 to 3, wherein in step S02, calculating the observation residuals of the non-key frames comprises:
performing domain transformation processing on the image prediction value of the non-key frame based on the random sampling matrix to obtain the image prediction value of a transformation domain;
and subtracting the image predicted value of the transform domain from the image observed value of the non-key frame to obtain an observation residual error.
5. The method of claim 4, wherein in step S02, the step of calculating the prediction residual of the non-key frame comprises:
and performing domain inverse transformation processing on the observation residual error based on the random sampling matrix to obtain a prediction residual error of the non-key frame.
6. The method for reconstructing a compressed video signal as claimed in claim 5, wherein in step S03, calculating the average energy of the residual signal of the observation residual of the non-key frame comprises:
calculating the residual signal average energy according to the following formula:
Figure FDA0002831782370000021
wherein R is the average energy of the residual signal,
Figure FDA0002831782370000022
observed residual for non-key frames
Figure FDA0002831782370000023
R is the average energy of the residual signal, k is the frame index of the image block, j is the image block index within the frame, L is the index of the non-zero residual signal, and L is the number of non-zero components in the residual signal.
7. The method of claim 6, wherein in step S03, an adaptive residual reconstruction algorithm is selected according to the average energy of the residual signal, and the reconstruction of the prediction residual of the non-key frame is performed, and the generating of the reconstructed residual of the non-key frame comprises:
according to the relation between preset judgment threshold values T1 and T2 and the average energy of the residual error signal, one of the following formulas is selected to calculate the sparse expression coefficient of the reconstructed residual error:
if R is less than or equal to T1, then use
Figure FDA0002831782370000024
If T1 < R ≦ T2, then use
Figure FDA0002831782370000025
If R > T2, use
Figure FDA0002831782370000026
Wherein the content of the first and second substances,
Figure FDA0002831782370000027
for sparse representation of reconstructed residualThe coefficients of which are such that,
Figure FDA0002831782370000028
sparse representation coefficients of the prediction residual for non-key frames,
Figure FDA0002831782370000031
is the observation residual of the non-key frame, lambda 1 and lambda 2 are weight factors for balancing the similarity and difference effects between the predicted image and the observation image,
the reconstructed residual of the non-key frame is calculated using the following formula:
Figure FDA0002831782370000032
wherein the content of the first and second substances,
Figure FDA0002831782370000033
Ψ is a redundant dictionary for the reconstructed residual of the non-key frame.
8. The method for reconstructing a compressed video signal as claimed in claim 7, wherein the step S04 comprises:
calculating an image reconstruction value of a non-key frame using the following formula
Figure FDA0002831782370000034
Figure FDA0002831782370000035
Wherein the content of the first and second substances,
Figure FDA0002831782370000036
for the image reconstruction values of non-key frames,
Figure FDA0002831782370000037
is the image prediction value of the non-key frame.
9. The method of reconstructing a compressed video signal according to claim 8, further comprising:
and calculating an image reconstruction value of the key frame according to the image observation value of the key frame.
10. The method of reconstructing a compressed video signal according to claim 9, further comprising:
respectively generating a key frame image and a non-key frame image by using the image reconstruction value of the non-key frame and the image reconstruction value of the key frame;
and combining the key frame images and the non-key frame images according to the sequence of the frame sequences to generate reconstructed video images.
CN202011461038.0A 2020-12-11 2020-12-11 Method for reconstructing video compression signal Active CN112616052B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011461038.0A CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011461038.0A CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Publications (2)

Publication Number Publication Date
CN112616052A true CN112616052A (en) 2021-04-06
CN112616052B CN112616052B (en) 2023-03-28

Family

ID=75234409

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011461038.0A Active CN112616052B (en) 2020-12-11 2020-12-11 Method for reconstructing video compression signal

Country Status (1)

Country Link
CN (1) CN112616052B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050169371A1 (en) * 2004-01-30 2005-08-04 Samsung Electronics Co., Ltd. Video coding apparatus and method for inserting key frame adaptively
CN101836457A (en) * 2007-10-25 2010-09-15 日本电信电话株式会社 Video scalable encoding method, video scalable decoding method, devices therefor, programs therefor, and recording medium where program is recorded
CN103297782A (en) * 2013-06-08 2013-09-11 河海大学常州校区 Area-partition-based reconstruction method in distributed video compression sensing (CS) system
US20160212448A1 (en) * 2014-05-28 2016-07-21 Peking University Shenzhen Graduate School Method and device for video encoding or decoding based on dictionary database
CN107155112A (en) * 2017-05-24 2017-09-12 湖北工业大学 A kind of compressed sensing method for processing video frequency for assuming prediction more
CN107360426A (en) * 2017-07-13 2017-11-17 福州大学 A kind of video sequence reconstructing method based on compressed sensing
CN108347612A (en) * 2018-01-30 2018-07-31 东华大学 A kind of monitored video compression and reconstructing method of view-based access control model attention mechanism
CN110933429A (en) * 2019-11-13 2020-03-27 南京邮电大学 Video compression sensing and reconstruction method and device based on deep neural network

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050169371A1 (en) * 2004-01-30 2005-08-04 Samsung Electronics Co., Ltd. Video coding apparatus and method for inserting key frame adaptively
CN101836457A (en) * 2007-10-25 2010-09-15 日本电信电话株式会社 Video scalable encoding method, video scalable decoding method, devices therefor, programs therefor, and recording medium where program is recorded
CN103297782A (en) * 2013-06-08 2013-09-11 河海大学常州校区 Area-partition-based reconstruction method in distributed video compression sensing (CS) system
US20160212448A1 (en) * 2014-05-28 2016-07-21 Peking University Shenzhen Graduate School Method and device for video encoding or decoding based on dictionary database
CN107155112A (en) * 2017-05-24 2017-09-12 湖北工业大学 A kind of compressed sensing method for processing video frequency for assuming prediction more
CN107360426A (en) * 2017-07-13 2017-11-17 福州大学 A kind of video sequence reconstructing method based on compressed sensing
CN108347612A (en) * 2018-01-30 2018-07-31 东华大学 A kind of monitored video compression and reconstructing method of view-based access control model attention mechanism
CN110933429A (en) * 2019-11-13 2020-03-27 南京邮电大学 Video compression sensing and reconstruction method and device based on deep neural network

Also Published As

Publication number Publication date
CN112616052B (en) 2023-03-28

Similar Documents

Publication Publication Date Title
CN108960333B (en) Hyperspectral image lossless compression method based on deep learning
US10812790B2 (en) Data processing apparatus and data processing method
CN104199627B (en) Gradable video encoding system based on multiple dimensioned online dictionary learning
CN106960420B (en) Image reconstruction method of segmented iterative matching tracking algorithm
Zhou et al. Image compression based on discrete cosine transform and multistage vector quantization
Abd-Alzhra et al. Image compression using deep learning: methods and techniques
CN112616052B (en) Method for reconstructing video compression signal
Zhang et al. Image primitive coding and visual quality assessment
Zheng et al. An improved distributed compressed video sensing scheme in reconstruction algorithm
Li et al. Image compression using fast transformed vector quantization
CN113096019B (en) Image reconstruction method, image reconstruction device, image processing equipment and storage medium
Upadhyaya et al. Quality parameter index estimation for compressive sensing based sparse audio signal reconstruction
Wahidah et al. A comparative study on video coding techniques with compressive sensing
Wang et al. Recovery error analysis of noisy measurement in compressed sensing
Thepade et al. New clustering algorithm for vector quantization using hybrid Haar slant error vector
Kumar et al. Comparative Analysis and Performance Evaluation of Medical Image Compression Method for Telemedicine
Rahman et al. An integer wavelet transform based lossless image compression technique using arithmetic coding
Prabhavathi et al. Compressive Sensing and its Application to Speech Signal Processing
Chatterjee et al. Image compression and resizing using vector quantization and other efficient algorithms
Wang et al. Reduced dimension Vector Quantization encoding method for image compression
CN114998457B (en) Image compression method, image decompression method, related device and readable storage medium
SAHNOUN et al. The Fourier transform for satellite image compression
Zhang et al. Application of Sparse Dictionary Adaptive Compression Algorithm in Transient Signals
Arunapriya et al. Image compression using single layer linear neural networks
Zhang An image reconstruction algorithm based on classified block

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant