WO2019056898A1 - Procédé et dispositif de codage et de décodage - Google Patents

Procédé et dispositif de codage et de décodage Download PDF

Info

Publication number
WO2019056898A1
WO2019056898A1 PCT/CN2018/101040 CN2018101040W WO2019056898A1 WO 2019056898 A1 WO2019056898 A1 WO 2019056898A1 CN 2018101040 W CN2018101040 W CN 2018101040W WO 2019056898 A1 WO2019056898 A1 WO 2019056898A1
Authority
WO
WIPO (PCT)
Prior art keywords
resolution image
image
full
encoder
residual
Prior art date
Application number
PCT/CN2018/101040
Other languages
English (en)
Chinese (zh)
Inventor
徐威
葛新宇
周力
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2019056898A1 publication Critical patent/WO2019056898A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding

Definitions

  • the present application relates to image processing technologies, and in particular, to a coding and decoding method and apparatus.
  • 4K TV has gradually moved toward thousands of households, providing users with ultra-high definition visual enjoyment.
  • the UK market research organization's future consulting report pointed out that it is estimated that global 4K TV shipments will reach 100 million units in 2018, and the Chinese market accounts for 70% of the global 4K TV market demand.
  • 4K video has a large demand for transmission bandwidth.
  • the basic 4K60P video requires a code rate of 30 to 50 Mbps to ensure good performance.
  • China's current access network bandwidth is only 20Mbps, the peak rate is 18.4Mbps, and the average speed is only 3.4Mbps.
  • the traditional universal video codec technology cannot guarantee the transmission of high quality 4K video in the existing bandwidth environment.
  • the present application provides an encoding method and apparatus, which can solve the transmission of high quality video in an existing bandwidth environment.
  • an encoding apparatus including: a down sampler, an up converter, a calculator, a self encoder, a main encoder, and a secondary encoder, wherein
  • the down sampler is configured to receive an original resolution image, perform down sampling processing on the original resolution image, acquire a low resolution image, and send the acquired low resolution image to the up converter and the Main encoder
  • the primary encoder is connected to the downsampler for encoding the received low resolution image to obtain primary coding information
  • the up converter is coupled to the downsampler, configured to convert the low resolution image sent by the downsampler into a first full resolution image, and send the first full resolution image to The calculator;
  • the calculator is configured to receive the original resolution image and the first full resolution image, and calculate a residual image according to the original resolution image and the first full resolution image, and the residual image
  • the difference image is sent to the self-encoder for processing
  • the self-encoder is connected to the calculator, and is configured to encode the residual image according to a self-encoding algorithm to obtain residual image encoding information, and send the residual image encoding information to the secondary encoder. ;
  • the secondary encoder is connected to the self-encoder for entropy encoding the residual image coding information from the self-encoder to obtain secondary coding information.
  • the encoding apparatus further includes: a filter booster connected to the up converter, configured to acquire a first full resolution image from the up converter, to the first full The resolution image is filtered and enhanced to obtain a second full resolution image; the calculator is further configured to calculate a residual image according to the original resolution image and the second full resolution image.
  • the filter enhancer is specifically configured to perform filtering enhancement on the first full-resolution image by using a bilateral filtering algorithm to obtain a second full-resolution image.
  • the calculator calculates a residual image according to the original resolution image and the second full resolution image, including:
  • the original resolution image is subtracted from the pixel value at the same position as the second full resolution image to obtain a residual image.
  • the self-encoder encodes the residual image according to a self-encoding algorithm to obtain residual image coding information, including:
  • Each image block is separately encoded according to a preset self-encoder network parameter to obtain residual image coding information.
  • a decoding apparatus including: a primary decoder, an up converter, a secondary decoder, a self decoder, and a synthesizer, wherein
  • the primary decoder is configured to obtain primary coding information, and decode the primary coding information to obtain a low resolution image
  • the up converter is connected to the main decoder for receiving the low resolution image from the main decoder, and converting the low resolution image to obtain a first full resolution image;
  • the secondary decoder is configured to obtain secondary coding information, and perform entropy decoding on the secondary coding information to obtain residual image coding information.
  • the self-decoder is connected to the auxiliary decoder, and is configured to self-decode the residual image coding information from the secondary decoder to acquire a residual image;
  • a synthesizer coupled to the upconverter and the self decoder for transmitting the first full resolution image from the upconverter and the residual image from the self decoder The synthesis is performed to obtain a first original resolution image.
  • the decoding apparatus further includes a filter enhancer coupled to the up converter for performing filter enhancement processing on the first full resolution image to obtain a second full resolution image.
  • the synthesizer is further configured to synthesize the residual image and the second full resolution image to obtain a second original resolution image.
  • the synthesizer is specifically configured to:
  • an encoding method including:
  • the original resolution image is input to a downsampler for downsampling processing to obtain a low resolution image
  • the residual image coding information is input to a secondary encoder for entropy coding to obtain secondary coding information.
  • a decoding method including:
  • the residual image is synthesized with the first full resolution image to obtain a first original resolution image.
  • an encoding apparatus having the function of implementing the encoding apparatus in the method described in the first aspect or the method described in the second aspect.
  • This function can be implemented in hardware or in hardware by executing the corresponding software.
  • the hardware or software includes one or more modules (or units) corresponding to the functions described above.
  • a decoding apparatus having the function of implementing the decoding apparatus in the method of the second aspect described above.
  • This function can be implemented in hardware or in hardware by executing the corresponding software.
  • the hardware or software includes one or more modules (or units) corresponding to the functions described above.
  • a computer program product comprising executable program code, wherein the program code comprises instructions that, when executed by the processor, cause the encoding device to perform the method as described in the above aspect Coding method.
  • a computer program product comprising executable program code, wherein the program code comprises instructions that, when executed by the processor, cause the decoding device to perform the method as described in the above aspect Decoding method.
  • embodiments of the present application provide a computer storage medium for storing computer software instructions for use in an encoding apparatus as described above, including a program designed to perform the above aspects.
  • embodiments of the present application provide a computer storage medium for storing computer software instructions for use in a decoding apparatus as described above, including a program designed to perform the above aspects.
  • a chip system comprising a processor for supporting an encoding device or a decoding device as described above to implement an encoding method or a decoding method as referred to in the above aspect.
  • the chip system further includes a memory for holding program instructions and data necessary for the communication device.
  • the chip system can be composed of chips, and can also include chips and other discrete devices.
  • FIG. 1 is a schematic structural diagram of an encoding apparatus according to an embodiment of the present application.
  • FIG. 2 is a schematic structural diagram of a decoding apparatus according to an embodiment of the present application.
  • FIG. 3 is a schematic flowchart of an encoding method according to an embodiment of the present application.
  • FIG. 4 is a schematic flowchart of a decoding method according to an embodiment of the present application.
  • FIG. 5 is a schematic structural diagram of a self-encoder according to an embodiment of the present application.
  • FIG. 6 is a schematic flowchart of a self-encoding method according to an embodiment of the present application.
  • the encoding apparatus includes: a down sampler 101, an up converter 102, a calculator 104, a self-encoder 105, a secondary encoder 106, and a main Encoder 107.
  • the down sampler 101 is configured to receive an original resolution image, perform down sampling processing on the original resolution image, acquire a low resolution image, and send the acquired low resolution image to the up converter 102 and the main
  • the encoder 107 performs processing; the main encoder 107 is connected to the downsampler 101 for encoding the received low resolution image to output main coding information; the up converter 102 and the down sampler 101 Connected to convert the low resolution image sent by the downsampler 101 into a first full resolution image, send the first full resolution image to the calculator 104; the calculator 104 is configured to receive And calculating the residual image according to the original resolution image and the first full resolution image, and transmitting the residual image to the encoder 105.
  • the encoder 105 is connected to the calculator 104, and configured to encode the residual image according to a self-encoding algorithm to obtain residual image encoding information, and send the residual image encoding information.
  • the secondary encoder 106; a secondary encoder 106, and from the encoder 105 is connected, for future entropy encoding the residual image encoding information to the encoder 105 is obtained from the secondary encoded information.
  • the encoding device further includes a filter booster 103 connected to the up converter 102, configured to acquire a first full resolution image from the up converter 102, and perform filtering enhancement on the first full resolution image. , obtaining a second full resolution image.
  • a filter booster 103 connected to the up converter 102, configured to acquire a first full resolution image from the up converter 102, and perform filtering enhancement on the first full resolution image. , obtaining a second full resolution image.
  • FIG. 2 is a schematic structural diagram of a decoding apparatus according to an embodiment of the present application, as shown in FIG. 2.
  • the decoding apparatus includes a main decoder 201, an up converter 202, a synthesizer 204, a self decoder 205, and a sub decoder 206.
  • the main decoder 201 is configured to obtain main coding information, and decode the main coding information to obtain a low resolution image.
  • the up converter 202 is connected to the main decoder 201 for receiving from the main decoder 201.
  • the low-resolution image is subjected to conversion processing to obtain a first full-resolution image; the secondary decoder 206 is configured to acquire secondary encoding information, and perform entropy decoding on the secondary encoding information to obtain a residual image.
  • the difference image coding information is output to the self-decoder 205 for processing; the self-decoder 205 is coupled to the secondary decoder 206 for the residual image from the secondary decoder 206.
  • the encoded information is self-decoded to obtain a residual image; a synthesizer 204 is coupled to the upconverter 202 and the self decoder 205, respectively, for the first full resolution image and the source from the upconverter 202
  • the residual image described from the decoder is synthesized to obtain a first original resolution image.
  • the decoding device further includes a filter enhancer 203, connected to the up converter 202, configured to perform filter enhancement processing on the first full resolution image to obtain a second full resolution image.
  • a filter enhancer 203 connected to the up converter 202, configured to perform filter enhancement processing on the first full resolution image to obtain a second full resolution image.
  • the self-encoder and the self-decoder may be a module or device, such as self-encoding.
  • Auto Encoder which itself contains both an encoding part and a decoding part, the self-encoder can correspond to the encoding part of the self-encoder, and the self-decoder corresponds to the decoding part of the self-encoder.
  • the encoder encodes the primary encoding information and the secondary encoding information, or the decoder pairs the primary encoding information and the secondary encoding.
  • the decoding process of information can be implemented by a software program executing on a programmable device and/or other hardware device, such as a graphics processing unit (GPU), a field programmable gate array (FPGA), a central processing unit (CPU), and The implementation of the software program is executed on the computing device.
  • the encoding process of the primary encoding information and the secondary encoding information by the encoding device, or the decoding process of the primary encoding information and the secondary encoding information by the decoding device may be at least partially hardware and/or embedded in a specific application integrated circuit ( The code in ASIC) is implemented together.
  • the encoding method of an embodiment of the present application is described below with reference to FIG. 3. As shown in FIG. 3, the encoding method can be implemented by the encoding apparatus shown in FIG. 1, and includes the following steps:
  • the downsampler functions to reduce the image and reduce the resolution of the image.
  • the original resolution image of the input size of M ⁇ N it is downsampled by s times to obtain an image of (M/s) ⁇ (N/s) size.
  • s should be the common divisor of M and N. Row.
  • An alternative method is to convert the image in the original resolution image s ⁇ s window into a pixel whose value is the mean of all pixels in the window.
  • Another method is to take a pixel of the original resolution image every (s-1) line and every (s-1) column to form a new image.
  • the image of 4K resolution (3840 ⁇ 2160) is reduced to 1080P resolution (1920 ⁇ 1080), which is actually 2 times downsampling of 4K objects, and the image in 2 ⁇ 2 window in 4K image can be turned into one.
  • Pixels which represent the pixels by the mean of all the pixels in the window; or one pixel at every other row and every other column of the original 4K image to form a reduced 1080P image.
  • the primary encoder can be encoded using any coding standard, such as the H.264, H.265, and VP9 standards.
  • the low-resolution images are sequentially subjected to predictive coding, transform coding, quantization loop post-processing, and entropy coding to obtain a binary code stream file.
  • the role of the upconverter is to convert a low resolution image into a high resolution image, also known as upsampling, such as upsampling a 1080P image into a 4K image.
  • Interpolation methods are generally used to insert new elements between pixels using appropriate interpolation algorithms based on the original image pixels. For example, a classical bicubic interpolation algorithm can be employed.
  • the full resolution image obtained at this time is of low quality, so it is called a low quality full resolution image.
  • the higher quality full resolution image is improved in image quality after filtering enhancement processing with respect to the low quality full resolution image as described above.
  • the filter booster uses a Bilateral filter algorithm to enhance the low-quality full-resolution image obtained in step 303, so as to remove the image particle noise and maintain the image edge and detail texture, and the human eye visual attention.
  • the mechanism is sensitive to information such as edge texture, and the subjective visual quality of the image can be improved by using the bilateral filtering algorithm.
  • the bilateral filtering algorithm is a classical filtering enhancement algorithm. It is a nonlinear filtering method. It is a kind of compromise processing that combines the spatial proximity of the image and the similarity of the pixel values. At the same time, the spatial domain information and the grayscale similarity are considered to achieve the edge preservation.
  • the purpose of denoising is simple, non-iterative, and local.
  • step 304 is an optional step.
  • step 305 Perform difference processing on the image of the original resolution and the high-quality full-resolution image obtained in step 304 to obtain a residual image.
  • the image of the original resolution is A
  • the image of the higher quality full resolution is B
  • the pixel values at all the same positions in A and B are subtracted, and a residual image is obtained.
  • the network structure of the constructed self-encoder encodes the plurality of image blocks of the residual image separately. For the specific implementation process, reference may be made to the self-encoding method as shown in FIG. 6.
  • the so-called entropy coding refers to a lossless coding method according to the principle of information entropy. Encode the input information into a binary stream file.
  • the entropy coding method used here may be Context-based Adaptive Binary Arithmetic Coding (CABAC), which is also widely used by the H.265 coding standard.
  • CABAC Context-based Adaptive Binary Arithmetic Coding
  • the self-encoded code output corresponding to each image block in step 306 is input to an entropy encoder, and CABAC is used as a binary code stream file to obtain auxiliary coded information of each image block.
  • a decoding method according to an embodiment of the present application is described below based on FIG. 4. As shown in FIG. 4, the decoding method includes the following steps:
  • the main coding information is input to the main decoder for decoding, and the main decoder can adopt any decoding standard, for example, adopting standards such as H.264, H.265, and VP9.
  • the main decoder can adopt any decoding standard, for example, adopting standards such as H.264, H.265, and VP9.
  • H.265 adopting standards such as H.264, H.265, and VP9.
  • the binary code stream file obtained after encoding is sequentially subjected to entropy decoding, inverse quantization, inverse transform and the like to decode a low resolution image.
  • this step may adopt the same method as step 303 of FIG. 3, such as using a classical bicubic interpolation algorithm to convert a low resolution image into a low quality full resolution image.
  • step 304 of FIG. 3 the same method as step 304 of FIG. 3 may be used, such as using a Bilateral filter algorithm to enhance the low-quality full-resolution image, so as to remove image particle noise and maintain image edge and detail texture. purpose.
  • the secondary coding information is input to the secondary decoder, and the residual image coding information is obtained after entropy decoding.
  • This step is opposite to step 307 of FIG. 3, and the binary stream file is decoded by the CABAC entropy decoding algorithm, and each image block is decoded in turn, and the self-encoded L 2 layer output corresponding to each image block is obtained.
  • the resolution of the residual image obtained in step 405 is the same as the higher quality full resolution image obtained in step 403.
  • the residual image recovered in step 405 is A
  • the high-quality full-resolution image obtained in step 403 is B
  • the pixel values at all the same positions in A and B are added to obtain reconstruction. After the original resolution image.
  • the decoding method of this embodiment is the inverse of the encoding method of the embodiment of FIG. 3, and the related details may be referred to and applied to each other, and details are not described herein again.
  • the so-called self-encoder is a fully connected neural network, setting the target output value of the network equal to the input value.
  • a self-encoder with 6 nodes is shown in Figure 5:
  • the self-encoder includes three layers.
  • the input layer L1 has six common nodes and one offset node.
  • the circle labeled "+1" is called a bias node;
  • the L2 layer contains three common nodes and one offset.
  • Node; output layer L3 contains 6 ordinary nodes.
  • Each common node of L2 is connected to all nodes of the L1 layer by one edge, and each connected node has a weight parameter w.
  • Each common node of L3 is connected with one edge of all nodes of the L2 layer, each There will also be a weight parameter w between the connected nodes, taking the output of the L2 layer as an example, and the output of the first node of the L2 layer. for:
  • the weight parameter between the first node of the L1 layer and the first node of the L2 layer The weight parameter between the second node of the L1 layer and the first node of the L2 layer, The weight parameter between the third node of the L1 layer and the first node of the L2 layer, and so on.
  • f(.) represents the activation function, here the sigmoid function is used.
  • W l-1 represents the set of weight parameters between the l-1 layer and the l layer
  • b l-1 represents the set of the bias term parameters of the l-1 layer and the first layer
  • the key is the calculation of the weight w of the edge connected between each node, which is generally divided into the following steps:
  • the training samples are generally selected to be more than 10,000.
  • loss function J(w,b) that is, the degree of the original input loss compared to the output of the self-encoder, is generally expressed by the square of the difference between the two.
  • the back propagation algorithm (BP) is used to calculate the weight w layer by layer.
  • the output of the L3 layer is calculated according to the equation (3), that is, the recovery of the original actual input 6-dimensional vector is obtained, which is the decoding process of the self-encoder.
  • the self-encoding method in the embodiment of the present application is further introduced in the following with reference to FIG. 6.
  • the self-encoding method includes the following steps:
  • each image block includes N pixels, the corresponding input layer L1 has N corresponding to the pixel points.
  • the output layer L3 contains N ordinary nodes.
  • Each common node of L2 is connected to all nodes of the L1 layer by one edge, and each connected node has a weight parameter w; similarly, each common node of L3 uses one edge with all nodes of the L2 layer. Connected, there will be a weight parameter w between each connected node.
  • the output of the first node of the L2 layer for:
  • a weight parameter is the weight between the second node N and L L. 1 layer 2 layer of the first node.
  • Weight is the weight parameter bias node between L and L. 1 layer 2 layer of the first node.
  • f(.) represents the activation function, here the sigmoid function is used.
  • the L 3 layer has a total of N nodes. Similarly, the output of the first, second, and Nth nodes of the L 3 layer is:
  • the output L N layer 3 comprises the original input image block of N pixels equal.
  • Another possible loss function construction method is to add sparse constraints to the intermediate layer L 2 .
  • the so-called sparse constraint means that when the output of a node of the L 2 layer is close to 1, we think that it is activated, and when the output of the node is close to 0, it is considered to be suppressed, and as many as possible, the number of suppression nodes of the L 2 layer is large. Then the sparse constraint is reached.
  • the jth output of the L 2 layer The average activity relative to an image block containing N pixels (x 1 to x N ) is:
  • join restrictions ⁇ is the sparsity coefficient, usually a small value close to zero, such as 0.05.
  • an additional penalty factor needs to be added to the constructed loss function, which will punish those that are significantly different from ⁇ . Therefore, the average activity of each node in the L 2 layer is kept in a small range, and the penalty factor is:
  • our goal is to determine the optimal parameters W and b, taking J(W,b) to a minimum, which needs to be achieved by the gradient descent method.
  • the network parameters of the self-encoder are initialized with random numbers, and the training samples are sequentially input into the self-encoder; then the loss function J(W, b) is obtained according to the difference between the output of the self-encoder and the original input; (a total of three layers) parameters are expressed as W l , the first layer bias term is expressed as b l , and the partial derivative is calculated:
  • is called the learning speed, and the range of values is generally [0.01, 0.1], which is a smaller real number.
  • the key to solving the gradient descent method is to calculate the partial derivative of the cost function J(W,b) for each layer of parameters, which needs to be implemented by the Back Propagation (BP) algorithm (this is a classic algorithm, no longer here. Brief description).
  • BP Back Propagation
  • the final optimal weight parameter W l and the offset term parameter b l of each layer of the self-encoder are obtained.
  • each image block of the residual image is input to an encoding portion of the self-encoder to be encoded.
  • the coding portion of the self-encoder refers to the input layer Layer L 1 and the intermediate layer Layer L 2 and the network structure therebetween.
  • the decoding process of the self-encoder corresponds to the encoding process inversely. Specifically, as shown in step 405 of FIG. 4, the entropy decoded result is input to the decoding part of the self-encoder, and the residual image is restored.
  • the specific process is as follows:
  • the decoding portion of the self-encoder refers to the intermediate layer L 2 and the output layer L 3 and the network structure therebetween. After constructing the network structure of the self-encoder and training the network parameters, what you need to do is to decode the entropy and get the output of the L 2 layer of the self-encoder. to Connected to the L 3 layer and calculated according to the formula (6), the output of the L 3 layer is obtained, which is the decoding result of the self-encoder, that is, the approximate recovery of the original image block containing N pixel points.
  • the entropy decoding results of all k image blocks m 1 , . . . m k in the original residual image are sequentially input to the decoding portion of the self-encoder, and the approximate restorations m ' 1 , . . . m ' k of the image blocks are obtained after decoding.
  • These approximate restored image blocks are combined in accordance with the positions of m 1 , . . . m k in the original residual image to obtain a restored residual image.
  • the codec method based on the self-encoder is nonlinear, and the distortion of the original residual image is less in the encoding and decoding process, so that the reconstructed video image has better subjective quality and objective quality.
  • both the pre-processing module and the post-processing module need to store a reference image matrix with a large amount of data
  • the self-described embodiment is based on The encoding and decoding method of the encoder only needs to store the parameters of the encoding part in the encoder (the parameters between the input layer L1 and the intermediate layer L2) and the decoding parameters (the parameters between the intermediate layer L2 and the output layer L3) stored from the encoder. ), the amount of data is small, saving storage space.
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, or an electrical, mechanical or other form of connection.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the embodiments of the present application.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • the technical solution of the present application may be in essence or part of the contribution to the prior art, or all or part of the technical solution may be embodied in the form of a software product stored in a storage medium.
  • a number of instructions are included to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present application.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

La présente invention porte, selon un mode de réalisation, sur un procédé et un dispositif de codage et de décodage. Le procédé de codage consiste : premièrement, à réaliser un sous-échantillonnage sur une image à résolution d'origine afin d'obtenir une image basse résolution, puis à entrer l'image basse résolution dans un codeur principal en vue d'un codage afin d'obtenir des informations de codage principal ; puis, à entrer l'image basse résolution dans un convertisseur élévateur en vue d'une interpolation afin d'obtenir une image à pleine résolution ; à calculer une image résiduelle en fonction de l'image à résolution d'origine et de l'image à pleine résolution ; à entrer l'image résiduelle dans un auto-codeur en vue d'un codage, et à acquérir des informations de codage d'image résiduelle ; et à entrer les informations de codage d'image résiduelle dans un codeur secondaire en vue d'un codage entropique afin d'obtenir des informations de codage secondaire. De cette manière, des images de qualité supérieure ou égale à 4K peuvent être divisées en des informations codées principales et en des informations codées auxiliaires pouvant être transmises séparément. Ce procédé permet de réaliser une transmission vidéo de qualité supérieure ou égale à 4K dans l'environnement de réseau actuel, et de garantir la qualité d'image vidéo reconstruite.
PCT/CN2018/101040 2017-09-21 2018-08-17 Procédé et dispositif de codage et de décodage WO2019056898A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710861871.6 2017-09-21
CN201710861871.6A CN109547784A (zh) 2017-09-21 2017-09-21 一种编码、解码方法及装置

Publications (1)

Publication Number Publication Date
WO2019056898A1 true WO2019056898A1 (fr) 2019-03-28

Family

ID=65811031

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/101040 WO2019056898A1 (fr) 2017-09-21 2018-08-17 Procédé et dispositif de codage et de décodage

Country Status (2)

Country Link
CN (1) CN109547784A (fr)
WO (1) WO2019056898A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111343458A (zh) * 2020-01-15 2020-06-26 浙江工业大学 一种基于重建残差的稀疏灰度图像编解码方法及系统

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110971903A (zh) * 2019-10-17 2020-04-07 西安万像电子科技有限公司 编码方法、设备及系统
CN110677681A (zh) * 2019-11-01 2020-01-10 合肥图鸭信息科技有限公司 一种视频编解码方法、装置及终端设备
KR20210067783A (ko) * 2019-11-29 2021-06-08 삼성전자주식회사 전자 장치, 그 제어 방법 및 시스템
CN111246206B (zh) * 2020-01-14 2021-09-21 山东浪潮科学研究院有限公司 一种基于自编码器的光流信息压缩方法及装置
EP4090024A4 (fr) * 2020-02-28 2023-01-18 Huawei Technologies Co., Ltd. Procédé et appareil de transmission d'image
CN112055223A (zh) * 2020-08-21 2020-12-08 浙江大华技术股份有限公司 一种图像编解码方法及编解码器

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003036980A1 (fr) * 2001-10-26 2003-05-01 Koninklijke Philips Electronics N.V. Schema de compression evolutif d'un point de vue spatial employant des techniques d'amelioration de nettete spatiale
CN1939067A (zh) * 2004-04-02 2007-03-28 汤姆森许可贸易公司 用于复杂度可伸缩视频编码器的方法和设备
CN103607591A (zh) * 2013-10-28 2014-02-26 四川大学 结合超分辨率重建的图像压缩方法
CN105959705A (zh) * 2016-05-10 2016-09-21 武汉大学 一种面向可穿戴设备的视频直播方法
CN107018422A (zh) * 2017-04-27 2017-08-04 四川大学 基于深度卷积神经网络的静止图像压缩方法
CN107181949A (zh) * 2017-06-23 2017-09-19 四川大学 一种结合超分辨率和残差编码技术的图像压缩框架

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100703740B1 (ko) * 2004-10-21 2007-04-05 삼성전자주식회사 다 계층 기반의 모션 벡터를 효율적으로 부호화하는 방법및 장치
CN105163121B (zh) * 2015-08-24 2018-04-17 西安电子科技大学 基于深度自编码网络的大压缩比卫星遥感图像压缩方法
CN106254879B (zh) * 2016-08-31 2019-11-08 广东精点数据科技股份有限公司 一种应用自编码神经网络的有损图像压缩方法

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003036980A1 (fr) * 2001-10-26 2003-05-01 Koninklijke Philips Electronics N.V. Schema de compression evolutif d'un point de vue spatial employant des techniques d'amelioration de nettete spatiale
CN1939067A (zh) * 2004-04-02 2007-03-28 汤姆森许可贸易公司 用于复杂度可伸缩视频编码器的方法和设备
CN103607591A (zh) * 2013-10-28 2014-02-26 四川大学 结合超分辨率重建的图像压缩方法
CN105959705A (zh) * 2016-05-10 2016-09-21 武汉大学 一种面向可穿戴设备的视频直播方法
CN107018422A (zh) * 2017-04-27 2017-08-04 四川大学 基于深度卷积神经网络的静止图像压缩方法
CN107181949A (zh) * 2017-06-23 2017-09-19 四川大学 一种结合超分辨率和残差编码技术的图像压缩框架

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111343458A (zh) * 2020-01-15 2020-06-26 浙江工业大学 一种基于重建残差的稀疏灰度图像编解码方法及系统
CN111343458B (zh) * 2020-01-15 2022-03-01 浙江工业大学 一种基于重建残差的稀疏灰度图像编解码方法及系统

Also Published As

Publication number Publication date
CN109547784A (zh) 2019-03-29

Similar Documents

Publication Publication Date Title
WO2019056898A1 (fr) Procédé et dispositif de codage et de décodage
US20200145692A1 (en) Video processing method and apparatus
WO2020237646A1 (fr) Procédé et dispositif de traitement d'image et support de stockage lisible par ordinateur
CN108737823B (zh) 基于超分辨技术的图像编码方法和装置、解码方法和装置
WO2023000179A1 (fr) Réseau de super-résolution vidéo, et procédé et dispositif de traitement de codage, décodage et super-résolution vidéo
CN111510739B (zh) 一种视频传输方法及装置
CN111800629A (zh) 视频解码方法、编码方法以及视频解码器和编码器
WO2023279961A1 (fr) Procédé et appareil de codage d'image vidéo, et procédé et appareil de décodage d'image vidéo
WO2018190199A1 (fr) Dispositif de traitement d'image, système de traitement d'image, système de traitement d'informations, et procédé de traitement d'image
CN115552905A (zh) 用于图像和视频编码的基于全局跳过连接的cnn滤波器
Öztürk et al. Performance evaluation of jpeg standards, webp and png in terms of compression ratio and time for lossless encoding
WO2022266955A1 (fr) Procédé et appareil de décodage d'images, procédé et appareil de traitement d'images, et dispositif
WO2023193629A1 (fr) Procédé et appareil de codage pour couche d'amélioration de région, et procédé et appareil de décodage pour couche d'amélioration de zone
JP7125559B2 (ja) ビットレート削減のためのビデオストリーム適応フィルタリング
Zhang et al. Dual-layer image compression via adaptive downsampling and spatially varying upconversion
WO2021168827A1 (fr) Procédé et appareil de transmission d'image
CN116762338A (zh) 使用预处理的视频编码
JP2024511587A (ja) ニューラルネットワークベースのピクチャ処理における補助情報の独立した配置
JP2024513693A (ja) ピクチャデータ処理ニューラルネットワークに入力される補助情報の構成可能な位置
WO2023010981A1 (fr) Procédés et appareil de codage et de décodage
WO2023279968A1 (fr) Appareil et procédé de codage et de décodage d'une image vidéo
CN111246205B (zh) 基于方向双四元数滤波器组的图像压缩方法
KR102604657B1 (ko) 영상 압축 성능 개선 방법 및 장치
WO2023092388A1 (fr) Procédé de décodage, procédé de codage, décodeur, codeur et système de codage et de décodage
WO2022246809A1 (fr) Procédé de codage, procédé de décodage, flux de code, codeur, décodeur et support de stockage

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18859839

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18859839

Country of ref document: EP

Kind code of ref document: A1