CN111009018A - Image dimensionality reduction and reconstruction method based on deep neural network - Google Patents

Image dimensionality reduction and reconstruction method based on deep neural network Download PDF

Info

Publication number
CN111009018A
CN111009018A CN201911347254.XA CN201911347254A CN111009018A CN 111009018 A CN111009018 A CN 111009018A CN 201911347254 A CN201911347254 A CN 201911347254A CN 111009018 A CN111009018 A CN 111009018A
Authority
CN
China
Prior art keywords
image
network
dimensionality reduction
neural network
deep neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201911347254.XA
Other languages
Chinese (zh)
Inventor
侯兴松
康越
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Tianbiyou Technology Co ltd
Original Assignee
Suzhou Tianbiyou Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Tianbiyou Technology Co ltd filed Critical Suzhou Tianbiyou Technology Co ltd
Priority to CN201911347254.XA priority Critical patent/CN111009018A/en
Publication of CN111009018A publication Critical patent/CN111009018A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/001Model-based coding, e.g. wire frame
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks

Abstract

The invention discloses an image dimensionality reduction and reconstruction method based on a deep neural network, which is used for carrying out image dimensionality reduction at a coding end, reducing bit streams generated by image compression and saving bandwidth; and carrying out image reconstruction at a decoding end. The invention combines discrete wavelet transform and deep learning to improve the potential of image degradation and reconstruction performance. The invention combines the templates of transformation, quantization and entropy coding together, so that the functional modules are mutually influenced, and the coding steps of transformation, quantization and the like are jointly optimized, so that the performance of the coder is optimal as far as possible. The invention not only uses the neural network to realize the dimensionality reduction and reconstruction of the image, but also nestedly uses the discrete wavelet transform. In order to provide more accurate codeword distribution when optimizing rate distortion, the invention combines a context network and a super-prior network, and the super-prior network can correct the prediction information of the context network and jointly generate more accurate mean and standard deviation parameters.

Description

Image dimensionality reduction and reconstruction method based on deep neural network
Technical Field
The invention relates to the field of image reconstruction, in particular to a method for reducing dimension and reconstructing an image based on a deep neural network.
Background
The image is a similar and vivid description of objective objects, is a relatively intuitive representation mode of objective objects and contains related information of the described objects.
With the development of the information age, image information is increasing, and the bandwidth of the network is limited. In this case, it is particularly important to reduce the bandwidth occupied by transmitting the image by reducing the dimension of the image. The image data can be reduced in dimension because of the redundancy in the data. The redundancy of image data is mainly represented by: spatial redundancy due to correlation between adjacent pixels in the image; temporal redundancy caused by correlation between different frames in the image sequence; spectral redundancy due to different band correlations. The goal of image dimensionality reduction is to reduce the number of bits required to represent an image by removing these data redundancies.
The wavelet transform is a hot point direction for researching image degradation and reconstruction, and the invention aims to combine discrete wavelet transform with deep learning to improve the potential of image degradation and reconstruction performance.
Disclosure of Invention
The invention aims to provide an image dimension reduction and reconstruction method based on a deep neural network.
In order to achieve the purpose, the invention designs an image dimension reduction and reconstruction method based on a deep neural network, which is used for constructing an image dimension reduction and reconstruction network framework, wherein the network framework comprises an encoding end and a decoding end; performing image dimensionality reduction at a coding end, reducing bit streams generated by image compression and saving bandwidth; reconstructing an image at a decoding end; the method comprises the following steps:
s1: at a coding end, inputting an image to be coded into a convolutional neural network containing Discrete Wavelet Transform (DWT) to obtain a low-resolution image y stored with structural information;
s2: quantizing the low resolution image y to obtain a code word
Figure BDA0002333725630000021
The code word is processed by an entropy coder to obtain a code stream file after entropy coding;
probability modeling is carried out on the quantized code words through a Gaussian mixture model to control the code rate; context and prior information are introduced, and the prior network learns the mean value and the standard deviation of the probability distribution based on the context;
s3: at a decoding end, inputting the compressed and decoded image into a convolution network containing Integer Wavelet Transform (IWT) to obtain an image reconstructed from an original image; the integer wavelet transform is the inverse of the discrete wavelet transform.
Further, in step S1, the input image [0,255] is normalized to [ -1.0, +1.0], and then the input image is subjected to convolution-DWT-convolution transformation to obtain the transformed feature codeword y.
Further, in step S1, after each layer of convolution operation, the activation function used is GDN;
the positive transformation expression of the GDN transformation is as follows:
Figure BDA0002333725630000022
wherein i and j are channel serial numbers; w is ai(m, n) is the characteristic code word of the ith channel with the plane position at (m, n); βiAnd gammajFor parameters in GDN conversion, ui(m, n) is wiAnd (m, n) the code word is subjected to GDN transformation to obtain a characteristic code word.
Further, in step S2, a superior network and a context model are combined to learn to obtain a mean and a standard deviation of the probability distribution;
the super-prior network provides parameter information for an entropy coder for transforming the code words by sacrificing additional code words so as to further remove redundant information between the code words; the context model predicts the probability model parameters of the entropy coder through the code words obtained by decoding, thereby saving the code words; the combination of the two can provide parameter information for the entropy codec more efficiently.
Further, in step S3, after the input image is subjected to the deconvolution-IWT-deconvolution transform, a decoded image of the original image is obtained, and the range of the decoded image is normalized to the range of [0,255] to obtain a final decoded image.
Further, in step S3, after each layer of deconvolution operation, the activation function used is IGDN;
the IGDN transformation is expressed as:
Figure BDA0002333725630000031
wherein i and j are channel serial numbers; w is ai(m, n) is the characteristic code word of the ith channel with the plane position at (m, n); βiAnd gammajFor a parameter in the IGDN transformation, ui(m, n) is wiAnd (m, n) the code words are subjected to IGDN conversion to obtain the characteristic code words.
Furthermore, the invention carries out effective rate distortion optimization;
parameters in the image dimension reduction and reconstruction network are obtained by network training and learning; in network training, in order to perform rate distortion optimization effectively, the code rate of the codeword needs to be estimated, so as to facilitate the control of the code rate of the encoder in the training. The loss function in training is therefore set to:
L=lamda*D+R
wherein lamda is used to control the model code rate; d is an original image x and a reconstructed image
Figure BDA0002333725630000037
Mean square error MSE between; r is the code rate, which is specifically calculated as follows:
Figure BDA0002333725630000032
wherein the content of the first and second substances,
Figure BDA0002333725630000033
the distribution parameters are obtained by combining a prior network and a context network, and the mean value and the standard deviation are respectively recorded as follows:
Figure BDA0002333725630000034
for the prior information
Figure BDA0002333725630000035
Constraining the distribution to be zero mean value, Gaussian distribution with standard deviation being learnable, and recording the standard deviation as
Figure BDA0002333725630000036
Compared with the prior art, the invention has the following advantages and characteristics:
conventional image downscaling and reconstruction algorithms, such as JPEG, JPEG2000, BPG, which use fixed transforms, i.e. discrete cosine transform and discrete wavelet transform, in combination with quantization and entropy coders, reduce the spatial redundancy of the image. These conventional image encoders mainly optimize each module inside the encoder, such as transform, quantization, and entropy coding, respectively. Aiming at the end-to-end mode of the deep neural network, the invention combines the templates of transformation, quantization and entropy coding together, so that the functional modules are mutually influenced, and the coding steps of transformation, quantization and the like are jointly optimized, so that the performance of the coder is optimal as far as possible.
The invention not only uses the neural network to realize the dimensionality reduction and reconstruction of the image, but also nestedly uses the discrete wavelet transform.
In order to provide more accurate codeword distribution when optimizing rate distortion, the invention combines a context network and a super-prior network, and the super-prior network can correct the prediction information of the context network and jointly generate more accurate mean and standard deviation parameters.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a diagram of a masked convolution for implementing a context network.
Detailed Description
The following description of the embodiments of the present invention will be made with reference to the accompanying drawings. The following examples are only for illustrating the technical solutions of the present invention more clearly, and the protection scope of the present invention is not limited thereby.
Referring to fig. 1, the invention constructs an image dimension reduction and reconstruction network framework based on a deep neural network, and the network framework comprises the following parts: encoder, decoder, quantization, super-a-encoder, super-a-decoder, context network, entropy encoder parameter estimation network, entropy encoder, entropy decoder.
Based on the network framework, the invention provides an image dimension reduction and reconstruction method, which comprises the following steps:
s1: at a coding end, inputting an image to be coded into a convolutional neural network containing Discrete Wavelet Transform (DWT) to obtain a low-resolution image y stored with structural information;
s2: quantizing the low resolution image y to obtain a code word
Figure BDA0002333725630000041
The code word is processed by an entropy coder to obtain a code stream file after entropy coding;
probability modeling is carried out on the quantized code words through a Gaussian mixture model to control the code rate; context and prior information are introduced, and the prior network learns the mean value and the standard deviation of the probability distribution based on the context;
s3: at a decoding end, inputting the compressed and decoded image into a convolution network containing Integer Wavelet Transform (IWT) to obtain an image reconstructed from an original image; the integer wavelet transform is the inverse of the discrete wavelet transform.
The method comprises the following specific steps:
1) preprocessing and standardizing the range of the image to be coded to the range of [ -1, +1] to obtain a preprocessed image x;
2) sending the x into an encoder to obtain an output result y of the encoder; the method comprises the following specific steps:
2.1) x is subjected to convolution operation with a layer of convolution kernel with the size of 5x5, the number of channels of 128 and the step length of 2; then, the result after convolution is sent to GDN transformation of the first layer;
2.2) after the first layer GDN conversion, the output is sent to a DWT conversion layer and is converted by the GDN of the second layer;
2.3) inputting the output result into a second layer convolution operation, wherein the size of a convolution kernel is 5x5, the number of channels is 512, and the step length is 2; then, similarly, the convolution output of the layer is transformed by GDN of a third layer;
2.4) then sending the output result to the convolution layer of the last layer, wherein the size of the convolution kernel is 5x5, the number of channels is 192, the step length is 2, and then obtaining an unquantized coded code word y;
3) sending y into an encoder in a super-prior network to obtain a super-prior code word z;
the encoding process comprises three layers of convolution operation, wherein the convolution kernel size of each layer is 5x5, the channel number is 128, and the step length is 2;
4) the quantization of the rounding of the codeword z is obtained to obtain the quantized codeword
Figure BDA0002333725630000051
5) Passing the information according to an entropy model
Figure BDA0002333725630000052
Writing into a code stream file, assuming
Figure BDA0002333725630000053
The obedience mean value is 0, and the standard deviation is a learnable Gaussian distribution;
6) will be provided with
Figure BDA0002333725630000054
Sending the data to a decoder in a super prior network to obtain a parameter phi;
this decoding process includes three layers of deconvolution operations, where the convolution kernel size of the last layer is 5x5, the number of channels is 192, and the step size is 2;
7) the characteristic code word y is quantized by rounding, and the quantized code word
Figure BDA00023337256300000610
Inputting the information into a context network to obtain parameter information theta;
the context model is realized by mask convolution, when a pixel value of a certain point is generated by the mask convolution layer, the pixel values of the right side and the lower side of the pixel are covered, and the pixel value of the current point is calculated only by the pixel values of the upper side and the left side, as shown in FIG. 2, wherein the gray position is the pixel value to be generated;
8) combining phi and theta in channel dimensions, and inputting the combined phi and theta into an entropy encoder parameter estimation network to obtain estimated mu and delta;
9) will be provided with
Figure BDA0002333725630000061
Entropy coding is carried out under a Gaussian mixture probability model with the mean value of mu and the standard deviation of delta, a code stream file is written in, and once image coding is completed after code stream writing is completed;
10) the decoding process is similar to the encoding process; reading code stream file and according to code word
Figure BDA0002333725630000062
Is obtained by decoding the entropy probability model
Figure BDA0002333725630000063
Then will be
Figure BDA0002333725630000064
Sending the data to a decoder in a super prior network to obtain a parameter phi; and will be
Figure BDA0002333725630000065
The decoding information is input into a context network to obtain a parameter theta;
11) inputting phi and theta into an entropy encoder parameter estimation network to obtain
Figure BDA0002333725630000066
The probability model parameters mu, delta; and sending mu and delta into an entropy decoder according to the writing sequence, and continuously decoding quantized characteristic code words
Figure BDA0002333725630000067
12) After the decoding of the entropy decoder is finished, the entropy decoder will
Figure BDA0002333725630000068
Sending the data into a decoder network; obtaining a decoded image and normalizing the decoded image range to [0,255]]Obtaining a final decoded image; the specific operation is as follows:
12.1) code word after quantization
Figure BDA0002333725630000069
Feeding the data into a first layer of deconvolution, wherein the size of a convolution kernel is 5x5, the number of channels is 192, and the step length is 2; then, the deconvolution result of the first layer is sent to the first layer IGDN for conversion;
12.2) continuously sending the output into a deconvolution layer of a second layer, and then passing through an IGDN of the second layer;
12.3) then sending the output result of the second layer IGDN conversion into an IWT conversion layer;
12.4) subjecting the result to one-layer deconvolution operation, wherein the kernel size is 5x5, the channel number is 1, and the step size is 2, thereby obtaining a decoded image, and normalizing the range of the decoded image to [0,255] to obtain a final decoded image.
Compared with the traditional image encoder structure, the parameters of the encoder and the decoder are obtained by network training and learning; moreover, in the network training, in order to effectively perform rate distortion optimization, the code rate of the codeword needs to be estimated; the existing code rate estimation schemes mainly have two types: one is to directly constrain the number of code words, and this code rate constraint scheme is mostly seen in the self-encoder where the output code words are binary; the other method is to assume the distribution of code words, then solve the self-information quantity according to the probability of the code words to estimate the code rate, and take the estimated code rate as the approximate code rate in the network training; the entropy probability model based on deep learning in the invention is Gaussian distribution.
The foregoing is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, various modifications and decorations can be made without departing from the technical principle of the present invention, and these modifications and decorations should also be regarded as the protection scope of the present invention.

Claims (10)

1. The image dimensionality reduction and reconstruction method based on the deep neural network is characterized in that an image dimensionality reduction and reconstruction network framework is constructed, and the network framework comprises an encoding end and a decoding end; performing image dimensionality reduction at a coding end, reducing bit streams generated by image compression and saving bandwidth; reconstructing an image at a decoding end; the method comprises the following steps:
s1: at a coding end, inputting an image to be coded into a convolutional neural network containing Discrete Wavelet Transform (DWT) to obtain a low-resolution image y stored with structural information;
s2: quantizing the low resolution image y to obtain a code word
Figure FDA0002333725620000011
The code word is processed by an entropy coder to obtain a code stream file after entropy coding;
probability modeling is carried out on the quantized code words through a Gaussian mixture model to control the code rate; context and prior information are introduced, and the prior network learns the mean value and the standard deviation of the probability distribution based on the context;
s3: at a decoding end, inputting the compressed and decoded image into a convolution network containing Integer Wavelet Transform (IWT) to obtain an image reconstructed from an original image; the integer wavelet transform is the inverse of the discrete wavelet transform.
2. The deep neural network-based image dimensionality reduction and reconstruction method according to claim 1, wherein the step S1 includes the following specific steps:
1.1) standardizing the pixel value range of an image to be coded to [ -1.0,1.0], and obtaining a preprocessed image x;
1.2) carrying out a layer of convolution operation on x, extracting relevant characteristic information and obtaining an output result x 1;
x1 ═ F (x * w + b), where F is the activation function, x is the input information, w is the weight, b is the offset, and x1 is the output result;
1.3) carrying out DWT transformation on the x1 to obtain an output result x 2;
1.4) performing convolution operation on x2 n times to obtain a low-resolution image y of x.
3. The deep neural network-based image dimensionality reduction and reconstruction method according to claim 1, wherein the step S2 includes the following specific steps:
2.1) inputting the low-resolution image y into a super-first-order-check network to obtain an output result phi;
2.2) quantization of y by rounding to give
Figure FDA0002333725620000021
Figure FDA0002333725620000022
Wherein U (-0.5,0.5) is value range [ -0.5,0.5 [)]Uniform distribution of (2);
2.3) mixing
Figure FDA0002333725620000023
Inputting a context network to obtain an output result theta;
when the context network generates a pixel value of a certain point, the pixel values of the right side and the lower side of the pixel are covered, and the pixel value of the current point is calculated only through the pixel values of the upper side and the left side;
2.4) simultaneously inputting phi and theta into a parameter estimation network of an entropy coder, and carrying out multilayer convolution operation to obtain parameters of mixed Gaussian distribution, namely a mean value mu and a standard deviation delta;
2.5) entropy encoder based on μ, δ pairs
Figure FDA0002333725620000024
And coding to obtain a compressed bit stream.
4. The deep neural network-based image dimensionality reduction and reconstruction method according to claim 1, wherein the step S3 includes the following specific steps:
3.1) the entropy decoder decodes the compressed bit stream according to mu and delta to obtain
Figure FDA0002333725620000025
3.2) to
Figure FDA0002333725620000026
Performing deconvolution operation m times to obtainGraph with channel number being multiple of 4
Figure FDA0002333725620000027
Figure FDA0002333725620000028
Wherein Q is an activation function, w is a weight, and b is a bias;
3.3) to
Figure FDA0002333725620000029
Performing IWT conversion to obtain multi-channel result
Figure FDA00023337256200000210
3.4) pairs
Figure FDA00023337256200000211
Performing convolution operation with the number of output channels being 1 to obtain a decoded image of the original image x, and normalizing the pixel value range of the decoded image to [0,255]]To obtain the final decoded image
Figure FDA00023337256200000212
5. The deep neural network-based image dimensionality reduction and reconstruction method of claim 2, wherein:
the encoding end combines DWT and convolution operation;
in the step 1.2), the activation function is GDN;
in step 1.4), n is 2.
6. The deep neural network-based image dimensionality reduction and reconstruction method of claim 3, wherein:
in the step 2.1), the super-advanced network provides additional parameter information for the entropy encoder by sacrificing additional code words, and further removes redundant information between the code words;
in step 2.3), the function of the context network is realized by a mask convolution operation.
7. The deep neural network-based image dimensionality reduction and reconstruction method of claim 6, wherein the structure of the super-prior network is based on a convolution operation.
8. The deep neural network-based image dimensionality reduction and reconstruction method of claim 4, wherein:
the decoding end combines IWT and convolution operation;
in the step 3.2), the activation function is IGDN; m is 2.
9. The deep neural network-based image dimensionality reduction and reconstruction method according to claim 1, wherein, compared with a traditional image encoder structure, the encoder and decoder parameters in the image dimensionality reduction and reconstruction network are obtained by network training learning; and when the network is trained, in order to carry out rate distortion optimization effectively, the code rate of the code word is also estimated, so that the code rate of the encoder can be controlled conveniently in the training.
10. The image dimensionality reduction and reconstruction method based on the deep neural network of claim 1, wherein the prior network is complementary to the context network when estimating the code rate of the codeword; adding context information into the prior check network does not cause potential interest rate loss; introducing the super-a priori information in the context network eliminates a certain amount of uncertainty.
CN201911347254.XA 2019-12-24 2019-12-24 Image dimensionality reduction and reconstruction method based on deep neural network Withdrawn CN111009018A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911347254.XA CN111009018A (en) 2019-12-24 2019-12-24 Image dimensionality reduction and reconstruction method based on deep neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911347254.XA CN111009018A (en) 2019-12-24 2019-12-24 Image dimensionality reduction and reconstruction method based on deep neural network

Publications (1)

Publication Number Publication Date
CN111009018A true CN111009018A (en) 2020-04-14

Family

ID=70117755

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911347254.XA Withdrawn CN111009018A (en) 2019-12-24 2019-12-24 Image dimensionality reduction and reconstruction method based on deep neural network

Country Status (1)

Country Link
CN (1) CN111009018A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112019865A (en) * 2020-07-26 2020-12-01 杭州皮克皮克科技有限公司 Cross-platform entropy coding method and decoding method for deep learning coding
CN112702600A (en) * 2020-12-29 2021-04-23 南京大学 Image coding and decoding neural network layered fixed-point method
CN113067832A (en) * 2021-03-29 2021-07-02 郑州铁路职业技术学院 Communication data encryption method based on block chain and artificial intelligence
CN113079377A (en) * 2021-04-01 2021-07-06 中国科学技术大学 Training method for depth image/video compression network
CN113537456A (en) * 2021-06-15 2021-10-22 北京大学 Depth feature compression method
CN113747163A (en) * 2021-08-17 2021-12-03 上海交通大学 Image coding and decoding method and compression method based on context reorganization modeling
CN113949867A (en) * 2020-07-16 2022-01-18 武汉Tcl集团工业研究院有限公司 Image processing method and device
CN114079771A (en) * 2020-08-14 2022-02-22 华为技术有限公司 Image coding and decoding method and device based on wavelet transformation
CN114866782A (en) * 2022-03-21 2022-08-05 上海工程技术大学 Video image processing method based on depth dimension-variable code rate control
WO2023124148A1 (en) * 2021-12-27 2023-07-06 上海商汤智能科技有限公司 Data processing method and apparatus, electronic device and storage medium
WO2023221590A1 (en) * 2022-05-16 2023-11-23 华为技术有限公司 Encoding method, decoding method, and electronic device
CN117336494A (en) * 2023-12-01 2024-01-02 湖南大学 Dual-path remote sensing image compression method based on frequency domain characteristics

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113949867B (en) * 2020-07-16 2023-06-20 武汉Tcl集团工业研究院有限公司 Image processing method and device
CN113949867A (en) * 2020-07-16 2022-01-18 武汉Tcl集团工业研究院有限公司 Image processing method and device
CN112019865A (en) * 2020-07-26 2020-12-01 杭州皮克皮克科技有限公司 Cross-platform entropy coding method and decoding method for deep learning coding
CN114079771A (en) * 2020-08-14 2022-02-22 华为技术有限公司 Image coding and decoding method and device based on wavelet transformation
CN112702600A (en) * 2020-12-29 2021-04-23 南京大学 Image coding and decoding neural network layered fixed-point method
CN113067832A (en) * 2021-03-29 2021-07-02 郑州铁路职业技术学院 Communication data encryption method based on block chain and artificial intelligence
CN113067832B (en) * 2021-03-29 2022-01-21 郑州铁路职业技术学院 Communication data encryption method based on block chain and artificial intelligence
CN113079377A (en) * 2021-04-01 2021-07-06 中国科学技术大学 Training method for depth image/video compression network
CN113537456A (en) * 2021-06-15 2021-10-22 北京大学 Depth feature compression method
CN113537456B (en) * 2021-06-15 2023-10-17 北京大学 Depth feature compression method
CN113747163A (en) * 2021-08-17 2021-12-03 上海交通大学 Image coding and decoding method and compression method based on context reorganization modeling
CN113747163B (en) * 2021-08-17 2023-09-26 上海交通大学 Image coding and decoding method and compression method based on context recombination modeling
WO2023124148A1 (en) * 2021-12-27 2023-07-06 上海商汤智能科技有限公司 Data processing method and apparatus, electronic device and storage medium
CN114866782A (en) * 2022-03-21 2022-08-05 上海工程技术大学 Video image processing method based on depth dimension-variable code rate control
WO2023221590A1 (en) * 2022-05-16 2023-11-23 华为技术有限公司 Encoding method, decoding method, and electronic device
CN117336494A (en) * 2023-12-01 2024-01-02 湖南大学 Dual-path remote sensing image compression method based on frequency domain characteristics
CN117336494B (en) * 2023-12-01 2024-03-12 湖南大学 Dual-path remote sensing image compression method based on frequency domain characteristics

Similar Documents

Publication Publication Date Title
CN111009018A (en) Image dimensionality reduction and reconstruction method based on deep neural network
Cheng et al. Learned image compression with discretized gaussian mixture likelihoods and attention modules
Minnen et al. Joint autoregressive and hierarchical priors for learned image compression
US6704718B2 (en) System and method for trainable nonlinear prediction of transform coefficients in data compression
CN111641832B (en) Encoding method, decoding method, device, electronic device and storage medium
CN103329522B (en) For the method using dictionary encoding video
Conoscenti et al. Constant SNR, rate control, and entropy coding for predictive lossy hyperspectral image compression
CN108174218B (en) Video coding and decoding system based on learning
JP7356513B2 (en) Method and apparatus for compressing neural network parameters
CN111314709A (en) Video compression based on machine learning
CN113747163B (en) Image coding and decoding method and compression method based on context recombination modeling
CN111246206A (en) Optical flow information compression method and device based on self-encoder
Li et al. Multiple description coding based on convolutional auto-encoder
CN114449276B (en) Super prior side information compensation image compression method based on learning
CN115988215A (en) Variable bit rate image compression method, system, device, terminal and storage medium
CN113438481B (en) Training method, image encoding method, image decoding method and device
CN111343458B (en) Sparse gray image coding and decoding method and system based on reconstructed residual
US9948928B2 (en) Method and apparatus for encoding an image
CN1848960B (en) Residual coding in compliance with a video standard using non-standardized vector quantization coder
CN112437300B (en) Distributed video coding method based on self-adaptive interval overlapping factor
US9501717B1 (en) Method and system for coding signals using distributed coding and non-monotonic quantization
EP3180862B1 (en) Method for coding pulse vectors using statistical properties
EP3180863B1 (en) Method for coding pulse vectors using statistical properties
Kumar et al. An optimized block estimation based image compression and decompression algorithm
Vasuki et al. Image compression using lifting and vector quantization

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20200414

WW01 Invention patent application withdrawn after publication