CN110717948A - Image post-processing method, system and terminal equipment - Google Patents

Image post-processing method, system and terminal equipment Download PDF

Info

Publication number
CN110717948A
CN110717948A CN201911199740.1A CN201911199740A CN110717948A CN 110717948 A CN110717948 A CN 110717948A CN 201911199740 A CN201911199740 A CN 201911199740A CN 110717948 A CN110717948 A CN 110717948A
Authority
CN
China
Prior art keywords
post
network
processing
image
image data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911199740.1A
Other languages
Chinese (zh)
Inventor
李东阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei Map Duck Mdt Infotech Ltd
Original Assignee
Hefei Map Duck Mdt Infotech Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei Map Duck Mdt Infotech Ltd filed Critical Hefei Map Duck Mdt Infotech Ltd
Publication of CN110717948A publication Critical patent/CN110717948A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks

Abstract

The embodiment of the invention provides an image post-processing method, which comprises the following steps: acquiring image data decompressed by a decoder; processing the decompressed image data according to a post-processing network to obtain post-processing image data; the training method of the post-processing network comprises the following steps: compressing the training image through a coding network to obtain a compressed bit stream; inputting the compressed bit stream into a decoding network to obtain reconstructed image data; inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image; comparing the post-processing reconstructed image with the training image, and estimating according to a code rate to obtain a rate-distortion optimization result; and adjusting the parameters of the post-processing network according to the rate-distortion optimization result, so that the method can improve the quality of image decoding and has great performance improvement space.

Description

Image post-processing method, system and terminal equipment
Technical Field
The present invention relates to the field of image compression, and in particular, to an image post-processing method, system and terminal device.
Background
The current image coding network method has the problems of easy loss of details, general quality of detail reconstruction and the like.
Disclosure of Invention
In order to solve the above problem, embodiments of the present invention provide an image post-processing method, system and terminal device.
According to a first aspect of the present invention, there is provided an image post-processing method comprising:
acquiring image data decompressed by a decoder;
processing the decompressed image data according to a post-processing network to obtain post-processing image data;
the training method of the post-processing network comprises the following steps:
compressing the training image through a coding network to obtain a compressed bit stream;
inputting the compressed bit stream into a decoding network to obtain reconstructed image data;
inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
comparing the post-processing reconstructed image with the training image, and estimating according to a code rate to obtain a rate-distortion optimization result;
and adjusting parameters of the post-processing network according to the rate-distortion optimization result.
Further, the encoding network, the decoding network and the post-processing network are convolutional neural networks, and the structures of the networks are the same and/or different.
Further, the post-processing network comprises a three-layer convolutional neural network.
According to a second aspect of the present invention, there is provided an image post-processing system comprising:
first acquiring means for acquiring image data decompressed by the decoder;
the post-processing device is used for processing the decompressed image data according to a post-processing network to obtain post-processed image data;
wherein the post-processing device comprises:
the first compression unit is used for compressing the training image through a coding network to obtain a compressed bit stream;
a first reconstruction unit, configured to input the compressed bitstream into a decoding network to obtain reconstructed image data;
the first post-processing unit is used for inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
the first optimization unit is used for comparing the post-processing reconstructed image with the training image and obtaining a rate-distortion optimization result according to code rate estimation;
and the first adjusting unit is used for adjusting the parameters of the post-processing network according to the rate-distortion optimization result.
Further, the encoding network, the decoding network and the post-processing network are convolutional neural networks, and the structures of the networks are the same and/or different.
According to a third aspect of the present invention, there is provided a terminal device comprising:
a memory for storing a program;
a processor, coupled to the memory, for executing the program, which when executed performs the method provided by the present invention.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a flow chart of a method provided by an embodiment of the present invention;
FIG. 2 is a flow chart of a training method provided by an embodiment of the present invention;
FIG. 3 is a flow chart of a method provided by an embodiment of the present invention;
FIG. 4 is a flowchart of a training method provided by an embodiment of the present invention;
FIG. 5 is a schematic diagram of model training provided by an embodiment of the present invention;
FIG. 6 is a schematic diagram of model training provided by an embodiment of the present invention;
FIG. 7 is a schematic diagram of an apparatus provided by an embodiment of the present invention;
FIG. 8 is a schematic diagram of an exercise apparatus provided in accordance with an embodiment of the present invention;
FIG. 9 is a schematic diagram of an apparatus provided by an embodiment of the present invention;
FIG. 10 is a schematic view of an exercise apparatus provided in accordance with an embodiment of the present invention;
FIG. 11 is a schematic diagram of a post-processing network provided by an embodiment of the present invention;
fig. 12 is a schematic structural diagram of a terminal device according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As an embodiment of the present invention, there is provided an image post-processing method, as shown in fig. 1, including:
step 102, image data decompressed by a decoder is acquired.
Specifically, when the decoder decodes the compressed data, the decompressed image data is obtained.
And 104, processing the decompressed image data according to a post-processing network to obtain post-processing image data.
Specifically, the decompressed image data is processed according to a trained post-processing network to obtain post-processing image data.
The training method for the post-processing network is shown in fig. 2, and includes:
step 202, compressing the training image through a coding network to obtain a compressed bit stream.
In one possible embodiment, the training image may be compressed by a convolutional neural network based coding network.
And step 204, inputting the compressed bit stream into a decoding network to obtain reconstructed image data.
Specifically, the compressed bit stream is decoded according to a decoding network corresponding to the encoding network, so as to obtain reconstructed image data.
And step 206, inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image.
In one possible embodiment, the post-processing reconstructed map is obtained by processing the reconstructed map data through a post-processing network including three layers of convolutional neural networks.
FIG. 11 is a schematic diagram of the post-processing network, as shown in FIG. 11, wherein the input data is the reconstructed map data and the output data is the post-processing reconstructed map; the first layer of the three-layer convolutional neural network is Conv96x5x5/1, wherein 96 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the second layer of the three-layer convolutional neural network is Conv64x5x5/1, wherein 64 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the third layer of the three-layer convolutional neural network is Conv32x5x5/1, wherein 32 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the GDN between the convolutional neural networks is a generalized normalized network, and downsampling is not used here to avoid distortion caused by downsampling.
And 208, comparing the post-processing reconstructed image with the training image, and estimating according to the code rate to obtain a rate-distortion optimization result.
In particular, the method comprises the following steps of,
comparing the post-processing reconstruction image with the training image to obtain a distortion residual error;
and obtaining the rate-distortion optimization result according to the code rate estimation result and the distortion residual error.
In the training model, distortion D can be expressed as Mean Square Error (MSE)
Figure BDA0002295570270000051
A representation is made, wherein x represents the image (also referred to as original image or input image),
Figure BDA0002295570270000052
representing a reconstructed map or calculated using subjective distortion such as MS-SSIM. And performing end-to-end optimization on the self-coding compression algorithm according to a loss function R + lambda D for weighting the code rate and the distortion, wherein R represents the code rate, D represents the distortion, and lambda represents the weight, and in the optimization process, firstly defining the loss function and then optimizing the network parameters by using a back propagation algorithm.
And step 210, adjusting parameters of the post-processing network according to the rate-distortion optimization result.
Specifically, parameters of the post-processing network are trained according to the rate-distortion optimization result, and the parameters are optimized according to the training result.
Further, step 202 further comprises: and compressing the training image through a coding network to obtain compressed data, and quantizing the compressed data to obtain quantized compressed data.
Specifically, in the training process, an additive uniform noise is used to set a quantifier, and the representation mode is
Figure BDA0002295570270000053
Wherein
Figure BDA0002295570270000054
Is the quantized parameter, yiIs the coding feature and e is random noise. Wherein the variable
Figure BDA0002295570270000055
Entropy of (2) can be expressed by variable yiSo that it can be used in the actual use of the modelAs a quantization operation, such a mannerIn the following, the code rate can be accurately estimated.
As an embodiment of the present invention, there is provided an image compression method, as shown in fig. 3, including:
step 302, image data is acquired.
Specifically, image data to be compressed is acquired.
And step 304, compressing the image data according to a coding network to obtain a compressed bit stream.
Specifically, the image data is compressed according to the trained coding network, and compressed image data is obtained.
The method for training the coding network is shown in fig. 4, and includes:
step 402, extracting the features of the image through a feature extraction network.
Specifically, the three layers of convolutional neural networks as shown in fig. 5 are used to extract features of the image, and in an optional manner, the result obtained by each layer of convolutional neural networks is used as an input to calculate a final feature, that is, the normalized feature obtained after each layer of convolution is convolved again and used as a cascade input.
And step 404, estimating the characteristics according to a probability model to obtain a code rate estimation result.
Specifically, the method comprises the following steps:
and estimating the distribution according to a probability model, and estimating the code rate according to the entropy to obtain the code rate estimation result.
The data distribution of natural images is generally considered to be gaussian-like, so a zero mean, variance, can be taken as
Figure BDA0002295570270000061
Laplacian distribution of
Figure BDA0002295570270000062
For feature yiThe probability distribution is modeled, and the formula of the probability is as follows:
Figure BDA0002295570270000063
wherein μ represents the average distribution of the particles,representing the compression characteristics of a hyper-parametric network.
Further, a self-coding network may be employed to pair variancesLearning is carried out, the structure of the self-coding network is shown in figure 6,
Figure BDA0002295570270000066
and expressing the compression characteristics as the input of a hyper-parametric self-coding network, and learning the standard difference distribution, wherein in the hyper-parametric self-coding network, the expression formula of a variable z is as follows: z is he(y) wherein heRepresenting the encoder of the hyper-parametric learning network, and then performing quantization, the quantization formula being
Figure BDA0002295570270000067
The quantized representation may then be transmitted as an additional variable.
Code rate of features can be modeled using entropy structure
Figure BDA0002295570270000068
The prior distribution can be fitted using a parameterized approach and then the prior probability model is learned in a data-driven manner.
Step 406, quantizing the features to obtain quantized features.
Specifically, the method comprises the following steps:
in the training process, an additive uniform noise is used to set a quantifier, and the representation mode is
Figure BDA0002295570270000069
Where e is random noise. Wherein the variable
Figure BDA00022955702700000610
Entropy energy variable of
Figure BDA00022955702700000611
So that it can be used in the actual use of the model
Figure BDA0002295570270000071
As a quantization operation, in such a manner, the code rate can be estimated accurately.
And step 408, inputting the quantized features into a decoding network to obtain a reconstructed image.
In particular, the method comprises the following steps of,
and decoding the quantized features according to a decoding network to obtain a reconstructed image.
And step 410, inputting the reconstructed image into a post-processing network to obtain a post-processing reconstructed image.
In one possible embodiment, the post-processing reconstructed map is obtained by processing the reconstructed map data through a post-processing network including three layers of convolutional neural networks.
FIG. 11 is a schematic diagram of the post-processing network, as shown in FIG. 11, wherein the input data is the reconstructed map data and the output data is the post-processing reconstructed map; the first layer of the three-layer convolutional neural network is Conv96x5x5/1, wherein 96 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the second layer of the three-layer convolutional neural network is Conv64x5x5/1, wherein 64 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the third layer of the three-layer convolutional neural network is Conv32x5x5/1, wherein 32 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the GDN between the convolutional neural networks is a generalized normalized network, and downsampling is not used here to avoid distortion caused by downsampling.
And step 412, comparing the post-processing reconstructed image with the training image, and estimating according to the code rate to obtain a rate-distortion optimization result.
In particular, the method comprises the following steps of,
comparing the post-processing reconstruction image with the training image to obtain a distortion residual error;
and obtaining the rate-distortion optimization result according to the code rate estimation result and the distortion residual error.
In the training model, distortion D can be expressed as Mean Square Error (MSE)
Figure BDA0002295570270000072
A representation is made, wherein x represents the image (also referred to as original image or input image),representing a reconstructed map or calculated using subjective distortion such as MS-SSIM. And performing end-to-end optimization on the self-coding compression algorithm according to a loss function R + lambda D for weighting the code rate and the distortion, wherein R represents the code rate, D represents the distortion, and lambda represents the weight, and in the optimization process, firstly defining the loss function and then optimizing the network parameters by using a back propagation algorithm.
Step 414, adjusting parameters of the feature extraction network according to the rate-distortion optimization result.
Specifically, a gradient back propagation algorithm is adopted to update parameters of the convolutional neural network.
As an embodiment of the present invention, there is provided an image post-processing system, as shown in fig. 7, including:
first obtaining means 701 is used for obtaining the image data decompressed by the decoder.
Specifically, the first obtaining device 701 is used for obtaining the image data decompressed by the decoder.
A post-processing device 702, configured to process the decompressed image data according to a post-processing network, so as to obtain post-processed image data.
Specifically, the post-processing device 702 processes the decompressed image data according to a trained post-processing network, so as to obtain post-processed image data.
In particular, the method comprises the following steps of,
as shown in fig. 8, the post-processing apparatus 702 includes:
the first compression unit 801 is configured to compress the training image through the coding network to obtain a compressed bitstream.
In particular, the training image may be compressed by a convolutional neural network-based coding network.
The function of the self-coding network is to convert data from image space x to data coding space y, which contains an encoder fe. The role of the encoder is to convert the pixel value x of the image into a compression characteristic y-fe(x)。
A first reconstructing unit 802, configured to input the compressed bitstream into a decoding network, so as to obtain reconstructed image data.
Specifically, the compressed bit stream is decoded according to a decoding network corresponding to the encoding network, so as to obtain reconstructed image data.
A first post-processing unit 803, configured to input the reconstructed map data into a post-processing network, so as to obtain a post-processing reconstructed map.
Specifically, in one possible embodiment, the post-processing reconstructed map is obtained by processing the reconstructed map data through a post-processing network including three layers of convolutional neural networks.
FIG. 11 is a schematic diagram of the post-processing network, as shown in FIG. 11, wherein the input data is the reconstructed map data and the output data is the post-processing reconstructed map; the first layer of the three-layer convolutional neural network is Conv96x5x5/1, wherein 96 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the second layer of the three-layer convolutional neural network is Conv64x5x5/1, wherein 64 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the third layer of the three-layer convolutional neural network is Conv32x5x5/1, wherein 32 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the GDN between the convolutional neural networks is a generalized normalized network, and downsampling is not used here to avoid distortion caused by downsampling.
A first optimization unit 804, configured to compare the post-processing reconstructed image with the training image, and obtain a rate-distortion optimization result according to code rate estimation.
Specifically, the post-processing reconstructed image is compared with the training image to obtain a distortion residual error;
and obtaining the rate-distortion optimization result according to the code rate estimation result and the distortion residual error.
In the training model, distortion D can be expressed as Mean Square Error (MSE)
Figure BDA0002295570270000091
A representation is made, wherein x represents the image (also referred to as original image or input image),
Figure BDA0002295570270000092
representing a reconstructed map or calculated using subjective distortion such as MS-SSIM. And performing end-to-end optimization on the self-coding compression algorithm according to a loss function R + lambda D for weighting the code rate and the distortion, wherein R represents the code rate, D represents the distortion, and lambda represents the weight, and in the optimization process, firstly defining the loss function and then optimizing the network parameters by using a back propagation algorithm.
A first adjusting unit 805, configured to adjust a parameter of the post-processing network according to the rate-distortion optimization result.
Specifically, parameters of the post-processing network are trained according to the rate-distortion optimization result, and the parameters are optimized according to the training result.
As an embodiment of the present invention, there is provided an image compression system, as shown in fig. 9, including:
second acquiring means 901 for acquiring image data.
Specifically, the second acquiring means 901 acquires image data to be compressed.
A compressing device 902, configured to compress the image data according to a coding network, so as to obtain a compressed bitstream.
Specifically, the compression device 902 compresses the image data according to the trained coding network to obtain compressed image data.
As shown in fig. 10, the compressing apparatus 902 includes:
the second compression unit 1001 is configured to compress the training image through the coding network, so as to obtain a compressed bitstream.
Specifically, the three layers of convolutional neural networks as shown in fig. 5 are used to extract features of the image, and in an optional manner, the result obtained by each layer of convolutional neural networks is used as an input to calculate a final feature, that is, the normalized feature obtained after each layer of convolution is convolved again and used as a cascade input.
A second reconstructing unit 1002, configured to input the compressed bitstream into a decoding network, so as to obtain reconstructed image data.
Specifically, the second reconstruction unit 1002 decodes the compressed bitstream according to a decoding network to obtain a reconstructed image.
A second post-processing unit 1003, configured to input the reconstructed map data into a post-processing network, so as to obtain a post-processing reconstructed map.
Specifically, in one possible embodiment, the post-processing reconstructed map is obtained by processing the reconstructed map data through a post-processing network including three layers of convolutional neural networks.
FIG. 11 is a schematic diagram of the post-processing network, as shown in FIG. 11, wherein the input data is the reconstructed map data and the output data is the post-processing reconstructed map; the first layer of the three-layer convolutional neural network is Conv96x5x5/1, wherein 96 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the second layer of the three-layer convolutional neural network is Conv64x5x5/1, wherein 64 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the third layer of the three-layer convolutional neural network is Conv32x5x5/1, wherein 32 represents the number of channels, 5x5 represents the size of a convolution kernel, and/1 represents that downsampling is not adopted; the GDN between the convolutional neural networks is a generalized normalized network, and downsampling is not used here to avoid distortion caused by downsampling.
Modeling code rate by using entropy structure
Figure BDA0002295570270000111
The prior distribution can be fitted using a parameterized approach and then the prior probability model is learned in a data-driven manner.
And a second optimization unit 1004, configured to compare the post-processing reconstructed image with the training image, and obtain a rate-distortion optimization result according to the code rate estimation.
Specifically, the post-processing reconstructed image is compared with the training image to obtain a distortion residual error;
and obtaining the rate-distortion optimization result according to the code rate estimation result and the distortion residual error.
In the training model, distortion D can be expressed as Mean Square Error (MSE)
Figure BDA0002295570270000112
A representation is made, wherein x represents the image (also referred to as original image or input image),
Figure BDA0002295570270000113
representing a reconstructed map or calculated using subjective distortion such as MS-SSIM. And performing end-to-end optimization on the self-coding compression algorithm according to a loss function R + lambda D for weighting the code rate and the distortion, wherein R represents the code rate, D represents the distortion, and lambda represents the weight, and in the optimization process, firstly defining the loss function and then optimizing the network parameters by using a back propagation algorithm.
A second adjusting unit 1005, configured to adjust parameters of the coding network according to the rate-distortion optimization result.
Specifically, a gradient back propagation algorithm is adopted to update parameters of the convolutional neural network.
As an embodiment of the present invention, there is provided a terminal device, as shown in fig. 12, including:
a memory 1201 and a processor 1202.
A memory 1201 for storing a program.
In addition to the above-described programs, the memory 1201 may also be configured to store other various data to support operations on the terminal device. Examples of such data include instructions for any application or method operating on the terminal device, contact data, phonebook data, messages, pictures, videos, etc.
The memory 1201 may be implemented by any type or combination of volatile or non-volatile memory devices, such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), flash memory, magnetic or optical disks.
A processor 1202, coupled to the memory 1201, for executing a program in the memory 1201, the program when executed performing a method as described in any of fig. 1-4.
The above specific processing operations have been described in detail in the foregoing embodiments, and are not described again here.

Claims (10)

1. A method of image post-processing, the method comprising:
acquiring image data decompressed by a decoder;
processing the decompressed image data according to a post-processing network to obtain post-processing image data;
the training method of the post-processing network comprises the following steps:
compressing the training image through a coding network to obtain a compressed bit stream;
inputting the compressed bit stream into a decoding network to obtain reconstructed image data;
inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
comparing the post-processing reconstructed image with the training image, and estimating according to a code rate to obtain a rate-distortion optimization result;
and adjusting parameters of the post-processing network according to the rate-distortion optimization result.
2. The method of claim 1, wherein the encoding network, the decoding network, and the post-processing network are convolutional neural networks, and wherein the respective networks are identical and/or different in structure.
3. The method of claim 1, wherein the post-processing network comprises a three-layer convolutional neural network.
4. A method of image compression, the method comprising:
acquiring image data;
compressing the image data according to a coding network to obtain a compressed bit stream;
the training method of the coding network comprises the following steps:
compressing the training image through a coding network to obtain a compressed bit stream;
inputting the compressed bit stream into a decoding network to obtain reconstructed image data;
inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
comparing the post-processing reconstructed image with the training image, and estimating according to a code rate to obtain a rate-distortion optimization result;
and adjusting parameters of the coding network according to the rate-distortion optimization result.
5. The method of claim 4, wherein compressing the training image through the coding network to obtain a compressed bitstream comprises:
performing at least one downsampling compression on the training image according to the coding network to obtain a bit stream of the compressed image;
decompressing the bit stream of the compressed image according to the decoding network to obtain a reconstructed image;
calculating a residual error between the reconstructed image and the image to obtain a residual error;
carrying out at least one down-sampling compression on the residual error according to a residual error coding network to obtain a compressed residual error bit stream;
and decompressing the compressed residual bit stream according to a residual decoding network to obtain a residual reconstruction image.
6. The method according to claim 5, wherein the encoding network, the decoding network, the post-processing network and the residual encoding network are convolutional neural networks, and the structures of the respective networks are the same and/or different.
7. An image post-processing system, the system comprising:
first acquiring means for acquiring image data decompressed by the decoder;
the post-processing device is used for processing the decompressed image data according to a post-processing network to obtain post-processed image data;
wherein the post-processing device comprises:
the first compression unit is used for compressing the training image through a coding network to obtain a compressed bit stream;
a first reconstruction unit, configured to input the compressed bitstream into a decoding network to obtain reconstructed image data;
the first post-processing unit is used for inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
the first optimization unit is used for comparing the post-processing reconstructed image with the training image and obtaining a rate-distortion optimization result according to code rate estimation;
and the first adjusting unit is used for adjusting the parameters of the post-processing network according to the rate-distortion optimization result.
8. The system of claim 7, wherein the encoding network, the decoding network, and the post-processing network are convolutional neural networks, and the structures of the respective networks are the same and/or different.
9. An image compression system, the system comprising:
second acquiring means for acquiring image data;
the compression device is used for compressing the image data according to a coding network to obtain a compressed bit stream;
wherein the compression device comprises:
the second compression unit is used for compressing the training image through a coding network to obtain a compressed bit stream;
the second reconstruction unit is used for inputting the compressed bit stream into a decoding network to obtain reconstruction image data;
the second post-processing unit is used for inputting the reconstructed image data into a post-processing network to obtain a post-processing reconstructed image;
the second optimization unit is used for comparing the post-processing reconstructed image with the training image and obtaining a rate-distortion optimization result according to code rate estimation;
and the second adjusting unit is used for adjusting the parameters of the coding network according to the rate-distortion optimization result.
10. A terminal device, comprising:
a memory for storing a program;
a processor coupled to the memory for executing the program, the program when executed performing the method of any of claims 1-6.
CN201911199740.1A 2019-04-28 2019-11-29 Image post-processing method, system and terminal equipment Pending CN110717948A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910347491X 2019-04-28
CN201910347491 2019-04-28

Publications (1)

Publication Number Publication Date
CN110717948A true CN110717948A (en) 2020-01-21

Family

ID=69215744

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911199740.1A Pending CN110717948A (en) 2019-04-28 2019-11-29 Image post-processing method, system and terminal equipment

Country Status (1)

Country Link
CN (1) CN110717948A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023169319A1 (en) * 2022-03-10 2023-09-14 华为技术有限公司 Feature map coding method, feature map decoding method, and apparatus

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107018422A (en) * 2017-04-27 2017-08-04 四川大学 Still image compression method based on depth convolutional neural networks
CN109191402A (en) * 2018-09-03 2019-01-11 武汉大学 The image repair method and system of neural network are generated based on confrontation
CN109495741A (en) * 2018-11-29 2019-03-19 四川大学 Method for compressing image based on adaptive down-sampling and deep learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107018422A (en) * 2017-04-27 2017-08-04 四川大学 Still image compression method based on depth convolutional neural networks
CN109191402A (en) * 2018-09-03 2019-01-11 武汉大学 The image repair method and system of neural network are generated based on confrontation
CN109495741A (en) * 2018-11-29 2019-03-19 四川大学 Method for compressing image based on adaptive down-sampling and deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
于波等: "《基于深度卷积神经网络的图像重建算法》" *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023169319A1 (en) * 2022-03-10 2023-09-14 华为技术有限公司 Feature map coding method, feature map decoding method, and apparatus

Similar Documents

Publication Publication Date Title
CN111641832B (en) Encoding method, decoding method, device, electronic device and storage medium
CN110677651A (en) Video compression method
WO2020237646A1 (en) Image processing method and device, and computer-readable storage medium
CN110024391B (en) Method and apparatus for encoding and decoding a digital image or video stream
CN110753225A (en) Video compression method and device and terminal equipment
CN110892419B (en) Stop code tolerant image compression neural network
CN111641826B (en) Method, device and system for encoding and decoding data
CN111163314A (en) Image compression method and system
WO2020261314A1 (en) Image encoding method and image decoding method
CN107113426B (en) Method and apparatus for performing graph-based transformations using generalized graph parameters
CN113079378A (en) Image processing method and device and electronic equipment
CN116916036A (en) Video compression method, device and system
CN110730347A (en) Image compression method and device and electronic equipment
Chen et al. Neural network-based video compression artifact reduction using temporal correlation and sparsity prior predictions
CN110717948A (en) Image post-processing method, system and terminal equipment
CN115866252B (en) Image compression method, device, equipment and storage medium
CN111163320A (en) Video compression method and system
CN111161363A (en) Image coding model training method and device
CN111050170A (en) Image compression system construction method, compression system and method based on GAN
WO2022037146A1 (en) Image processing method, apparatus, device, computer storage medium, and system
CN110234011B (en) Video compression method and system
CN111565314A (en) Image compression method, coding and decoding network training method and device and electronic equipment
CN111565317A (en) Image compression method, coding and decoding network training method and device and electronic equipment
WO2019225344A1 (en) Encoding device, image interpolation system and encoding program
CN114501034B (en) Image compression method and medium based on discrete Gaussian mixture super prior and Mask

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200121

WD01 Invention patent application deemed withdrawn after publication