WO2020019279A1 - Procédé et appareil de compression vidéo, système informatique, et dispositif mobile - Google Patents

Procédé et appareil de compression vidéo, système informatique, et dispositif mobile Download PDF

Info

Publication number
WO2020019279A1
WO2020019279A1 PCT/CN2018/097343 CN2018097343W WO2020019279A1 WO 2020019279 A1 WO2020019279 A1 WO 2020019279A1 CN 2018097343 W CN2018097343 W CN 2018097343W WO 2020019279 A1 WO2020019279 A1 WO 2020019279A1
Authority
WO
WIPO (PCT)
Prior art keywords
quantization
video
specific
content analysis
quantization matrix
Prior art date
Application number
PCT/CN2018/097343
Other languages
English (en)
Chinese (zh)
Inventor
朱磊
高修峰
林茂疆
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to PCT/CN2018/097343 priority Critical patent/WO2020019279A1/fr
Priority to CN201880038972.6A priority patent/CN110771167A/zh
Publication of WO2020019279A1 publication Critical patent/WO2020019279A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • H04N19/126Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation

Definitions

  • the present application relates to the field of video processing, and more particularly, to a method, an apparatus, a computer system, and a mobile device for video compression.
  • Video content analysis is the use of visual algorithms to analyze the saved video in order to use the video content analysis results for corresponding applications.
  • the video content analysis of the video stored in the black box of the flight system can be used to analyze the cause of the accident.
  • the embodiments of the present application provide a method, a device, a computer system, and a mobile device for video compression, which can improve the efficiency of video compression.
  • a video compression method including: obtaining a first quantization matrix, wherein the quantization parameter of a high frequency component in the first quantization matrix is smaller than the quantization parameter of a low frequency component, and the quantization parameter of the high frequency component The corresponding frequency is higher than the frequency corresponding to the quantization parameter of the low-frequency component; the first video is compressed according to the first quantization matrix.
  • an apparatus for video compression including: an acquisition module for acquiring a first quantization matrix, wherein a quantization parameter of a high frequency component in the first quantization matrix is smaller than a quantization parameter of a low frequency component, and the high frequency component The frequency corresponding to the quantization parameter of is higher than the frequency corresponding to the quantization parameter of the low-frequency component; a compression module is configured to compress the first video according to the first quantization matrix.
  • a computer system including: a memory for storing computer-executable instructions; a processor for accessing the memory and executing the computer-executable instructions to perform the method of the first aspect Operation.
  • a mobile device including: the video compression device of the second aspect; or the computer system of the third aspect.
  • a computer storage medium stores program code, where the program code may be used to instruct execution of the method of the first aspect.
  • the technical solution of the embodiment of the present application compresses a video according to a quantization matrix with a quantization parameter of a high-frequency component smaller than a quantization parameter of a low-frequency component, and can compress a video while satisfying a video content analysis requirement, thereby improving the efficiency of video compression.
  • FIG. 1 is an architecture diagram of a technical solution to which an embodiment of the present application is applied.
  • FIG. 2 is a schematic diagram of data to be processed according to an embodiment of the present application.
  • FIG. 3 is a schematic diagram of a coding framework according to an embodiment of the present application.
  • FIG. 4 is a schematic architecture diagram of a mobile device according to an embodiment of the present application.
  • FIG. 5 is a schematic flowchart of a video compression method according to an embodiment of the present application.
  • FIG. 6 is a schematic block diagram of a video compression apparatus according to an embodiment of the present application.
  • FIG. 7 is a schematic block diagram of a video compression apparatus according to another embodiment of the present application.
  • FIG. 8 is a schematic block diagram of a computer system according to an embodiment of the present application.
  • the size of the sequence number of each process does not mean the order of execution.
  • the execution order of each process should be determined by its function and internal logic, and should not deal with the embodiments of the present application.
  • the implementation process constitutes any limitation.
  • FIG. 1 is an architecture diagram of a technical solution to which an embodiment of the present application is applied.
  • the system 100 may receive data to be processed 102, process the data to be processed 102, and generate processed data 108.
  • the system 100 may receive a video to be compressed, and perform processing such as encoding on the video to be compressed to compress the video.
  • the components in the system 100 may be implemented by one or more processors, which may be processors in a computing device or processors in a removable device (such as a drone).
  • the processor may be any kind of processor, which is not limited in the embodiment of the present application.
  • the processor may include an encoder, a decoder, or a codec.
  • the system 100 may also include one or more memories.
  • the memory may be used to store instructions and data, for example, computer-executable instructions that implement the technical solutions of the embodiments of the present application, data to be processed 102, data 108 to be processed, and the like.
  • the memory may be any kind of memory, which is not limited in the embodiment of the present application.
  • FIG. 2 is a schematic diagram of data to be processed according to an embodiment of the present application.
  • the data to be processed 202 may include a plurality of frames 204.
  • multiple frames 204 may represent consecutive image frames in a video stream.
  • Each frame 204 may include one or more tiles or tiles 206.
  • Each slice or tile 206 may include one or more macroblocks or coding units 208.
  • Each macroblock or coding unit 208 may include one or more blocks 210.
  • Each block 210 may include one or more pixels 212.
  • Each pixel 212 may include one or more data sets corresponding to one or more data portions, such as a luminance data portion and a chrominance data portion.
  • the data unit may be a frame, a slice, a tile, a coding unit, a macro block, a block, a pixel, or a group of any of the above.
  • the size of the data unit may vary.
  • a frame 204 may include 100 slices 206, each slice 206 may include 10 macro blocks 208, each macro block 208 may include 4 (for example, 2x2) blocks 210, and each block 210 may include 64 (e.g., 8x8) pixels 212.
  • the video data In order to reduce the bandwidth occupied by video storage and transmission, the video data needs to be encoded and compressed. Any suitable encoding technique can be used to encode the data to be encoded. The type of encoding depends on the data being encoded and the specific encoding requirements.
  • the encoder may implement one or more different codecs.
  • Each codec can include code, instructions, or computer programs that implement different encoding algorithms. Based on various factors, including the type and / or source of the data to be encoded, the receiving entity of the encoded data, available computing resources, network environment, business environment, rules and standards, etc., a suitable encoding algorithm can be selected to encode a given Data to be encoded.
  • the encoder may be configured to encode a series of video frames.
  • a series of steps can be used to encode the data in each frame.
  • the encoding step may include processing steps such as prediction, transform, quantization, and entropy encoding.
  • Prediction includes two types of intra prediction and inter prediction, the purpose of which is to remove redundant information of the current image block to be encoded by using prediction block information.
  • Intra prediction uses the information of the image in this frame to obtain prediction block data.
  • Inter prediction uses the information of the reference frame to obtain prediction block data, and the process includes dividing the image block to be encoded into several sub-image blocks; then, for each sub-image block, searching in the reference image for an image that best matches the current sub-image block The block is used as a prediction block; thereafter, the sub-pixel block is subtracted from the corresponding pixel value of the prediction block to obtain a residual, and the residuals corresponding to the obtained sub-image blocks are combined to obtain the residual of the image block.
  • Using a transformation matrix to transform the residual block of the image can remove the correlation of the residual of the image block, that is, remove the redundant information of the image block in order to improve the coding efficiency.
  • the transformation of the data block in the image block usually uses two-dimensional transformation. That is, at the encoding end, the residual information of the data block is respectively multiplied with an NxM transform matrix and its transpose matrix, and the transform coefficients are obtained after the multiplication. Transform coefficients can be quantized to obtain quantized coefficients. Finally, the quantized coefficients are entropy encoded. Finally, the bit stream obtained by entropy encoding and the encoding mode information after encoding are performed, such as intra prediction mode and motion vector information. Store or send to the decoder.
  • the entropy-coded bitstream is first obtained and then the entropy decoding is performed to obtain the corresponding residuals.
  • the predicted image block corresponding to the information image block such as the motion vector or intra prediction obtained by the decoding according to the predicted image block and the image block Residual to get the value of each pixel in the current sub-image block.
  • FIG. 3 is a schematic diagram of a coding framework according to an embodiment of the present application.
  • the encoding process can be as follows:
  • a current frame image is acquired.
  • a reference frame image is acquired.
  • a reference frame image is used to perform motion estimation to obtain a motion vector (MV) of each image block of the current frame image.
  • the motion vector obtained by the motion estimation is used to perform motion compensation to obtain an estimated value / predicted value of the current image block.
  • the estimated value / predicted value of the current image block is subtracted from the current image block to obtain a residual error.
  • the residuals are transformed to obtain transformation coefficients.
  • the transform coefficient is quantized to obtain a quantized coefficient.
  • the quantized coefficients are subjected to entropy encoding, and finally the bit stream obtained by entropy encoding and the encoding mode information after encoding are stored or sent to the decoding end.
  • the quantized result is inversely quantized.
  • the inverse quantization result is inversely transformed.
  • the inverse transform result and the motion compensation result are used to obtain a reconstructed pixel.
  • the reconstructed pixels are filtered (ie, video compressed).
  • the filtered reconstructed pixels are output. Subsequently, the reconstructed image can be used as a reference frame image for other frame images for inter-frame prediction.
  • the encoding process can be as follows:
  • a current frame image is acquired.
  • intra prediction selection is performed on the current frame image.
  • the current image block in the current frame is intra-predicted.
  • the estimated value of the current image block is subtracted from the current image block to obtain a residual.
  • the residuals of the image blocks are transformed to obtain transformation coefficients.
  • the transform coefficient is quantized to obtain a quantized coefficient.
  • the quantized coefficients are subjected to entropy encoding, and finally the bit stream obtained by the entropy encoding and the encoding mode information after encoding are stored or sent to the decoding end.
  • the quantization result is inversely quantized.
  • the inverse quantization result is inversely transformed.
  • the inverse transform result and the intra prediction result are used to obtain a reconstructed pixel.
  • the reconstructed image block can be used for intra prediction of the next image block.
  • the residual information is obtained by using entropy decoding, inverse quantization and inverse transformation, and it is determined whether the current image block uses intra prediction or inter prediction according to the decoded code stream. If it is intra prediction, use the reconstructed image blocks in the current frame to construct prediction information according to the intra prediction method; if it is inter prediction, you need to parse out the motion information, and use the parsed motion information to reconstruct the image
  • the reference block is determined in order to obtain the prediction information.
  • the prediction information and the residual information are superimposed, and the reconstruction information can be obtained after the filtering operation.
  • the technical solution of the embodiment of the present application mainly relates to a quantization step in the encoding process, that is, to improve the compression efficiency of a video through improvement in the quantization step, and other steps may refer to related steps in the encoding process.
  • the mobile device may use the technical solution of the embodiment of the present application to compress the captured video.
  • the movable device may be an unmanned aerial vehicle, an unmanned boat, an autonomous vehicle, a robot, or an aerial photography vehicle, but the embodiment of the present application is not limited thereto.
  • FIG. 4 is a schematic architecture diagram of a mobile device 400 according to an embodiment of the present application.
  • the mobile device 400 may include a power system 410, a control system 420, a sensing system 430, and a processing system 440.
  • the power system 410 is used to power the mobile device 400.
  • the power system of the drone may include an electronic governor (referred to as an ESC), a propeller, and a motor corresponding to the propeller.
  • the motor is connected between the electronic governor and the propeller, and the motor and the propeller are arranged on the corresponding arm; the electronic governor is used to receive the driving signal generated by the control system and provide the driving current to the motor according to the driving signal to control the Rotating speed.
  • the motor is used to drive the propellers to rotate, thereby powering the drone's flight.
  • the sensing system 430 may be used to measure the posture information of the mobile device 400, that is, the position information and status information of the mobile device 400 in space, such as three-dimensional position, three-dimensional angle, three-dimensional velocity, three-dimensional acceleration, and three-dimensional angular velocity.
  • the sensing system 430 may include, for example, at least one of a gyroscope, an electronic compass, an Inertial Measurement Unit (IMU), a vision sensor, a Global Positioning System (GPS), a barometer, and an airspeed meter. Species.
  • the sensing system 430 may also be used to acquire an image, that is, the sensing system 430 includes a sensor for acquiring an image, such as a camera.
  • the control system 420 is used to control the movement of the mobile device 400.
  • the control system 420 may control the mobile device 400 according to a preset program instruction.
  • the control system 420 may control the movement of the mobile device 400 according to the posture information of the mobile device 400 measured by the sensing system 430.
  • the control system 420 may also control the mobile device 400 according to a control signal from a remote controller.
  • the control system 420 may be a flight control system (flight control), or a control circuit in the flight control.
  • the processing system 440 may process images acquired by the sensing system 430.
  • the processing system 440 may be an image signal processing (Image Signal Processing, ISP) type chip.
  • ISP Image Signal Processing
  • the processing system 440 may be the system 100 in FIG. 1, or the processing system 440 may include the system 100 in FIG. 1.
  • the mobile device 400 may further include other components not shown in FIG. 4, which is not limited in the embodiment of the present application.
  • FIG. 5 shows a schematic flowchart of a video compression method 500 according to an embodiment of the present application.
  • the method 500 may be performed by the system 100 shown in FIG. 1; or may be performed by the mobile device 400 shown in FIG. 4. Specifically, when executed by the mobile device 400, it may be executed by the processing system 440 in FIG. 4.
  • 510 Obtain a first quantization matrix, wherein the quantization parameter of the high-frequency component in the first quantization matrix is smaller than the quantization parameter of the low-frequency component, and the frequency corresponding to the quantization parameter of the high-frequency component is higher than the quantization parameter of the low-frequency component. Frequency of.
  • Quantization implements video compression by dividing the transformed transform coefficients by the corresponding quantization step size.
  • the quantization step size is indicated by a quantization parameter in the quantization matrix.
  • the quantization matrix includes quantization parameters corresponding to each frequency, and the quantization parameters of different frequencies may be different to achieve selective energy region loss. The larger the quantization parameter, the larger the quantization step size, and the larger the compression ratio.
  • the quantization parameter of the high frequency component in the first quantization matrix used is smaller than the quantization parameter of the low frequency component. That is, in the technical solution of the embodiment of the present application, low-frequency information is more lost, and high-frequency information is retained.
  • the first video is quantized by using the first quantization matrix in the embodiment of the present application. Specifically, in the quantization process, each frequency component (transformation coefficient) is divided by a quantization step size indicated by a corresponding quantization parameter. Because the quantization parameter of the high frequency component in the first quantization matrix is smaller than the quantization parameter of the low frequency component, that is, the quantization step size used for the high frequency component is small, and the quantization step size used for the low frequency component is large. Therefore, after quantization, the compression ratio of the high frequency component Small, compression loss is small, low frequency component compression ratio is large, compression loss is large. Using this asymmetric quantization strategy can not only achieve a higher compression rate as a whole, increase the compression time, but also meet the quality requirements of video content analysis. Therefore, the video compression method in the embodiments of the present application can improve the efficiency of video compression.
  • different quantization matrices can be configured for different scenarios. That is, based on the basic setting that the quantization parameter of the high frequency component is smaller than the quantization parameter of the low frequency component, the quantization matrix of different scenes can be set with different quantization parameters.
  • multiple quantization matrices can be pre-configured, and each scene corresponds to a quantization matrix.
  • the first quantization matrix corresponding to the scene of the first video may be selected from multiple quantization matrices according to the scene of the first video.
  • a quantization matrix can be configured for each scene of video content analysis.
  • a corresponding quantization matrix is found according to the scene of the video, and the quantization matrix is used for video compression.
  • the correspondence between the scene and the quantization matrix may be one-to-one or one-to-many, which is not limited in this application.
  • the multiple quantization matrices may be determined according to video samples of multiple scenes.
  • a quantization matrix corresponding to the scene may be trained in advance based on video samples of the scene.
  • the training process may be to gradually adjust the quantization parameters in the quantization matrix, and finally obtain a quantization matrix that meets the requirements.
  • the quantization parameters in the initial quantization matrix are adjusted according to the specific video samples of the specific scenes and the difference threshold of the video content analysis result to obtain A specific quantization matrix corresponding to the specific scene, wherein a difference between a video content analysis result corresponding to the specific video sample compressed by using the specific quantization matrix and a video content analysis result corresponding to the uncompressed specific video sample is not different Greater than the difference threshold of the video content analysis result.
  • the difference threshold of the video content analysis result is the difference threshold of the video content analysis result that can be tolerated when the video content analysis is performed. That is to say, the difference between the video content analysis result obtained from the video content analysis based on the compressed video and the video content analysis result obtained from the video content analysis based on the uncompressed video is within the threshold of the video content analysis result difference.
  • the video content analysis results obtained from the compressed video meet the requirements. In this way, when determining a specific quantization matrix corresponding to a specific scene, based on the initial quantization matrix, the quantization parameters in the initial quantization matrix can be adjusted, the specific video samples are compressed with the adjusted quantization matrix, and then video content analysis is performed. Compare the obtained video content analysis result with the video content analysis result corresponding to the uncompressed specific video sample.
  • the difference between the two is greater than the difference threshold of the video content analysis result, continue to adjust, if the difference is not greater than the difference of the video content analysis result
  • the threshold is adjusted, and the adjusted quantization matrix is a specific quantization matrix corresponding to the specific scene.
  • the above initial quantization matrix may be a standard quantization matrix, or may be another quantization matrix, for example, a predefined quantization matrix, a quantization matrix used before, and the like, which are not limited in this embodiment of the present application.
  • the quantization parameters in the initial quantization matrix may be adjusted in the following manner:
  • the next quantization parameter is adjusted; wherein the first set includes M quantization parameters in the initial quantization matrix, and the M quantization parameters correspond to The lowest M frequencies, M is a predetermined value;
  • the adjustment is stopped to obtain the specific quantization matrix; wherein the second set Including other quantization parameters in the initial quantization matrix except the M quantization parameters.
  • the quantization parameters in the initial quantization matrix may be different adjustment directions.
  • the quantization parameters of the low frequency part (quantization parameters in the first set)
  • Adjust in the direction of increasing the quantization parameters is adopted
  • the quantization parameters of the high frequency part (quantization parameters in the second set)
  • Adjust in the direction of decreasing the quantization parameter For example, the above-mentioned M may be 4, that is, for the DC component and the first three AC main components, it is adjusted in the direction of increasing the quantization parameter, and the other AC components are adjusted in the direction of decreasing the quantization parameter.
  • the quantization parameters in the first set are adjusted first. Because the adjustment is in the direction of increasing the quantization parameters, the adjusted video content analysis results will be worse, that is, the above differences will become larger. When the differences converge (that is, the differences are basically When it does not change), adjust the next quantization parameter.
  • the quantization parameters in the first set may be adjusted in order from low to high frequency.
  • the quantization parameters in the first set After adjusting the quantization parameters in the first set, adjust the quantization parameters in the second set. Because the adjustment is to reduce the quantization parameters, the adjusted video content analysis results will be better, that is, the above differences will become smaller. Before the difference is not greater than the difference threshold of the video content analysis result, after adjusting to the convergence of the difference, the next quantization parameter is adjusted until the difference is not greater than the difference threshold of the video content analysis result.
  • the adjusted quantization matrix is a specific quantization matrix corresponding to the specific scene.
  • the quantization parameters in the second set may be adjusted in order of frequency from low to high, or adjusted in order of frequency from high to low, or may be adjusted in predetermined order.
  • the predetermined order may be related to the degree of dependence of the video content analysis on each frequency component, for example, it may be adjusted in the order of the degree of dependence from high to low, but this embodiment of the present application is not limited thereto.
  • the adjustment of the initial quantization matrix may be adjusted by a scaling matrix.
  • Qr is an initial quantization matrix
  • Scl is a scaling matrix
  • Q ' is a quantization matrix finally used for quantization.
  • Scl includes a scaling factor corresponding to each quantization parameter.
  • Each quantization parameter is scaled by Scl to adjust its quantization intensity.
  • the process of adjusting the quantization parameters is the process of adjusting the scaling factor corresponding to each quantization parameter.
  • the scaling factor corresponding to each quantization parameter is determined, that is, the scaling matrix Scl is determined, and thus the quantization matrix Q 'is determined.
  • the scaling matrix Scl can be obtained in the following manner.
  • Adjust the scaling factors one by one in Scl (the initial value of each scaling factor can be 1). Taking the adjustment from the lowest frequency as an example, adjust to g convergence, or g ⁇ T; if g ⁇ T, stop adjusting; if g convergence , That is, the difference in the analysis results of the video content between the two adjustments is small, for example, ⁇ 0.01T, the adjustment of the current frequency is stopped, and the adjustment of the next frequency is started.
  • the first component the DC component, which is adjusted in the direction of the amplification and quantization parameter, that is, the zoom factor is greater than 1;
  • the second, third, and fourth components the AC main component, which is adjusted in the direction of the enlarged quantization parameter, that is, the scaling factor is greater than 1; if the g distortion is large during the first adjustment, the adjustment of the quantization parameter of the corresponding component can be abandoned;
  • the remaining components are adjusted in the direction of reducing the quantization parameter, that is, the scaling factor is ⁇ 1.
  • subsequent video content analysis may be performed based on the compressed first video.
  • the above quantization matrix is used for video compression, which retains high-frequency information such as edges, structures, and textures that are meaningful for video content analysis, and simultaneously reduces low-frequency energy, achieving the purpose of compression rate and the quality of compression for offline analysis effectiveness.
  • the technical solution of the embodiment of the present application compresses a video according to a quantization matrix with a quantization parameter of a high-frequency component smaller than a quantization parameter of a low-frequency component, and can compress the video under a condition that satisfies video content analysis requirements, thereby improving effectiveness.
  • the video compression method in the embodiment of the present application is described in detail above.
  • the video compression device, computer system, and mobile device in the embodiment of the present application will be described below.
  • FIG. 6 shows a schematic block diagram of a video compression apparatus 600 according to an embodiment of the present application.
  • the apparatus 600 may execute the video compression method in the embodiment of the present application.
  • the apparatus 600 may include:
  • the obtaining module 610 is configured to obtain a first quantization matrix, wherein a quantization parameter of a high frequency component in the first quantization matrix is smaller than a quantization parameter of a low frequency component, and a frequency corresponding to the quantization parameter of the high frequency component is higher than the low frequency component.
  • a compression module 620 is configured to compress a first video according to the first quantization matrix.
  • the compressed first video is used for video content analysis.
  • the obtaining module 610 is configured to:
  • the first quantization matrix corresponding to the scene of the first video is selected from a plurality of quantization matrices.
  • the apparatus 600 further includes:
  • a configuration module 630 is configured to pre-configure the multiple quantization matrices.
  • the configuration module 630 is configured to:
  • the multiple quantization matrices are determined according to video samples of multiple scenes.
  • the configuration module 630 is configured to:
  • a difference between a video content analysis result corresponding to the specific video sample compressed by using the specific quantization matrix and a video content analysis result corresponding to the uncompressed specific video sample is not greater than a difference threshold of the video content analysis result.
  • the configuration module 630 is configured to:
  • the next quantization parameter is adjusted; wherein the first set includes M quantization parameters in the initial quantization matrix, and the M quantization parameters correspond to The lowest M frequencies, M is a predetermined value;
  • the adjustment is stopped to obtain the specific quantization matrix; wherein the second set Including other quantization parameters in the initial quantization matrix except the M quantization parameters.
  • the configuration module 630 is configured to adjust the quantization parameters in the first set in a descending order of frequency.
  • the configuration module 630 is configured to adjust the quantization parameters in the second set according to a frequency from low to high or from high to low.
  • the configuration module 630 is configured to adjust the quantization parameters in the second set in a predetermined order.
  • the M is 4.
  • the initial quantization matrix is a standard quantization matrix
  • the foregoing video compression device in the embodiment of the present application may be a chip, which may be specifically implemented by a circuit, but the embodiment of the present application does not limit a specific implementation form.
  • An embodiment of the present invention further provides an encoder, and the encoder includes the foregoing video compression apparatus according to various embodiments of the present invention.
  • FIG. 8 shows a schematic block diagram of a computer system 800 according to an embodiment of the present application.
  • the computer system 800 may include a processor 810 and a memory 820.
  • computer system 800 may also include components generally included in other computer systems, such as input-output devices, communication interfaces, and the like, which is not limited in the embodiments of the present application.
  • the memory 820 is configured to store computer-executable instructions.
  • the memory 820 may be various types of memory, for example, it may include high-speed random access memory (Random Access Memory, RAM), and may also include non-volatile memory (non-volatile memory), such as at least one magnetic disk memory. Examples are not limited to this.
  • RAM Random Access Memory
  • non-volatile memory such as at least one magnetic disk memory. Examples are not limited to this.
  • the processor 810 is configured to access the memory 820 and execute the computer-executable instructions to perform operations in the video compression method in the embodiment of the present application.
  • the processor 810 may include a microprocessor, a Field-Programmable Gate Array (FPGA), a Central Processing Unit (CPU), and a Graphics Processing Unit (GPU). Examples are not limited to this.
  • FPGA Field-Programmable Gate Array
  • CPU Central Processing Unit
  • GPU Graphics Processing Unit
  • the embodiment of the present application further provides a movable device, and the movable device may include the video compression device or the computer system of the various embodiments of the present application described above.
  • the video compression device, computer system, and mobile device may correspond to the execution subject of the video compression method according to the embodiments of the present application. And other operations and / or functions are respectively used to implement the corresponding processes of the foregoing methods, and for the sake of brevity, they are not repeated here.
  • An embodiment of the present application further provides a computer storage medium, and the computer storage medium stores a program code, where the program code may be used to instruct to perform the foregoing video compression method in the embodiment of the present application.
  • the term “and / or” is merely an association relationship describing an associated object, and indicates that there may be three relationships.
  • a and / or B can indicate that there are three cases in which A exists alone, A and B exist, and B exists alone.
  • the character "/" in this article generally indicates that the related objects are an "or" relationship.
  • the disclosed systems, devices, and methods may be implemented in other ways.
  • the device embodiments described above are only schematic.
  • the division of the unit is only a logical function division.
  • multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may also be electrical, mechanical or other forms of connection.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, may be located in one place, or may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions in the embodiments of the present application.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above integrated unit may be implemented in the form of hardware or in the form of software functional unit.
  • the integrated unit When the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it may be stored in a computer-readable storage medium.
  • the technical solution of this application is essentially a part that contributes to the existing technology, or all or part of the technical solution may be embodied in the form of a software product, which is stored in a storage medium
  • Included are instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the method described in the embodiments of the present application.
  • the foregoing storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks and other media that can store program codes .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé et un appareil de compression vidéo, un système informatique, et un dispositif mobile. Le procédé consiste à : obtenir une première matrice de quantification, dans laquelle un paramètre de quantification d'une composante haute fréquence est inférieur à un paramètre de quantification d'une composante basse fréquence, et la fréquence correspondant au paramètre de quantification de la composante haute fréquence est supérieure à la fréquence correspondant au paramètre de quantification de la composante basse fréquence ; et compresser une première vidéo selon la première matrice de quantification. La solution technique des modes de réalisation de la présente invention peut améliorer l'efficacité d'une compression vidéo.
PCT/CN2018/097343 2018-07-27 2018-07-27 Procédé et appareil de compression vidéo, système informatique, et dispositif mobile WO2020019279A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2018/097343 WO2020019279A1 (fr) 2018-07-27 2018-07-27 Procédé et appareil de compression vidéo, système informatique, et dispositif mobile
CN201880038972.6A CN110771167A (zh) 2018-07-27 2018-07-27 视频压缩的方法、装置、计算机系统和可移动设备

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/097343 WO2020019279A1 (fr) 2018-07-27 2018-07-27 Procédé et appareil de compression vidéo, système informatique, et dispositif mobile

Publications (1)

Publication Number Publication Date
WO2020019279A1 true WO2020019279A1 (fr) 2020-01-30

Family

ID=69180715

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/097343 WO2020019279A1 (fr) 2018-07-27 2018-07-27 Procédé et appareil de compression vidéo, système informatique, et dispositif mobile

Country Status (2)

Country Link
CN (1) CN110771167A (fr)
WO (1) WO2020019279A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112437301B (zh) * 2020-10-13 2021-11-02 北京大学 一种面向视觉分析的码率控制方法、装置、存储介质及终端

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101695132A (zh) * 2004-01-20 2010-04-14 松下电器产业株式会社 图像编码方法和设备、图像解码方法和设备及其程序
US20130188692A1 (en) * 2012-01-25 2013-07-25 Yi-Jen Chiu Systems, methods, and computer program products for transform coefficient sub-sampling
CN103444180A (zh) * 2011-03-09 2013-12-11 日本电气株式会社 视频编码设备、视频解码设备、视频编码方法以及视频解码方法
CN106063265A (zh) * 2014-02-26 2016-10-26 杜比实验室特许公司 用于视频压缩的基于亮度的编码工具

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101695132A (zh) * 2004-01-20 2010-04-14 松下电器产业株式会社 图像编码方法和设备、图像解码方法和设备及其程序
CN103444180A (zh) * 2011-03-09 2013-12-11 日本电气株式会社 视频编码设备、视频解码设备、视频编码方法以及视频解码方法
US20130188692A1 (en) * 2012-01-25 2013-07-25 Yi-Jen Chiu Systems, methods, and computer program products for transform coefficient sub-sampling
CN106063265A (zh) * 2014-02-26 2016-10-26 杜比实验室特许公司 用于视频压缩的基于亮度的编码工具

Also Published As

Publication number Publication date
CN110771167A (zh) 2020-02-07

Similar Documents

Publication Publication Date Title
US11783512B2 (en) Attribute value of reconstructed position associated with plural original points
US10666979B2 (en) Image processing device and image processing method for encoding/decoding omnidirectional image divided vertically
US10911781B2 (en) Image processing device and image processing method for encoding/decoding omnidirectional images at different resolutions
US20210211698A1 (en) Parallel video encoding
US20230239516A1 (en) Entropy encoding/decoding method and apparatus
KR102599314B1 (ko) 포인트 클라우드 압축을 위한 양자화 단차 파라미터
US10735724B2 (en) Method and device for compressing image on basis of photography information
CN112771859A (zh) 基于感兴趣区域的视频数据编码方法、装置和存储介质
JP2023504408A (ja) クロマサンプリングのための方法及び装置
KR20130136170A (ko) 3차원 영상을 위한 영상 처리 장치 및 방법
US20210014486A1 (en) Image transmission
US11113866B2 (en) Method and apparatus for point cloud compression
WO2021098030A1 (fr) Procédé et appareil de codage vidéo
US20240205435A1 (en) Encoding and decoding method and apparatus
KR20220143859A (ko) 크로마 신호를 처리하기 위한 방법
JP2023507259A (ja) ラップアラウンド動き補償を実行する方法
WO2020019279A1 (fr) Procédé et appareil de compression vidéo, système informatique, et dispositif mobile
WO2012118569A1 (fr) Quantification optimisée visuellement
EP4264946A1 (fr) Compression de données temporelles au moyen d'une compression de nuage de points basée sur une géométrie
JP2022548203A (ja) サブブロック変換情報を信号化するための方法及び装置
WO2020073882A1 (fr) Décodeur vidéo et procédés correspondants
CN116325732A (zh) 点云的解码、编码方法、解码器、编码器和编解码系统
US20240283952A1 (en) Adaptive coding tool selection with content classification
RU2778377C1 (ru) Способ и устройство для кодирования облака точек
US20240129546A1 (en) Artificial intelligence-based image encoding and decoding apparatus, and image encoding and decoding method thereby

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18927971

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18927971

Country of ref document: EP

Kind code of ref document: A1