WO2022021519A1 - Video decoding method, system and device and computer-readable storage medium - Google Patents

Video decoding method, system and device and computer-readable storage medium Download PDF

Info

Publication number
WO2022021519A1
WO2022021519A1 PCT/CN2020/111448 CN2020111448W WO2022021519A1 WO 2022021519 A1 WO2022021519 A1 WO 2022021519A1 CN 2020111448 W CN2020111448 W CN 2020111448W WO 2022021519 A1 WO2022021519 A1 WO 2022021519A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
viewpoint
target
image
playback
Prior art date
Application number
PCT/CN2020/111448
Other languages
French (fr)
Chinese (zh)
Inventor
王荣刚
王振宇
高文
Original Assignee
北京大学深圳研究生院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京大学深圳研究生院 filed Critical 北京大学深圳研究生院
Publication of WO2022021519A1 publication Critical patent/WO2022021519A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47202End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/643Communication protocols

Definitions

  • the present application relates to the technical field of digital signal processing, and in particular, to a video decoding method, system, device, and computer-readable storage medium.
  • Free viewpoint applications allow viewers to watch videos in a range of continuous viewpoints.
  • the viewer can set the position and angle of the viewpoint, instead of being limited to a fixed camera angle.
  • This application often requires multiple cameras to shoot at the same time to generate videos from multiple viewpoints at the same time.
  • Virtual reality technology allows users to watch videos in a 180- or 360-degree view. Therefore, whether it is a multi-channel video in a free viewpoint application or a panoramic video in a virtual reality application, there is a very large amount of data, which brings great challenges to video transmission.
  • the commonly used methods of encoding and decoding such videos are usually divided into two categories. One is the method with inter-frame dependence. Although this method can achieve higher compression efficiency, it needs to rely on other decoded images during decoding.
  • the main purpose of this application is to provide a video decoding method, system, device and computer-readable storage medium to solve the technology that it is difficult to take into account the compression efficiency and the smoothness of viewpoint switching through the existing encoding and decoding methods of multi-channel video or panoramic video. problem.
  • the present application provides a video decoding method, the video decoding method includes:
  • a playback viewpoint selection instruction When a playback viewpoint selection instruction is received, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, multiple image streams corresponding to different viewpoints of the target video are received, and based on a plurality of the reconstructed background frames, the different The multi-channel image code stream of the viewpoint is decoded to obtain the multi-channel target video.
  • the present application also provides a video decoding system
  • the video decoding system includes:
  • the background code stream decoding module is used for receiving multiple background frame code streams corresponding to different viewpoints sent by the encoder, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
  • the image code stream decoding module is used for receiving the multi-channel image code streams corresponding to different viewpoints of the target video according to the viewpoint selection sequence determined by the playback viewpoint selection instruction when receiving the playback viewpoint selection instruction, and based on a plurality of The reconstructed background frame decodes the multi-channel image code streams of different viewpoints to obtain multi-channel target videos.
  • the image code stream decoding module includes:
  • the viewpoint-by-view determination unit is configured to, when receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoints currently selected by the user one by one based on the playback viewpoint selection instruction, and obtain the target playback viewpoints generated by the encoding end and the target playback viewpoints one by one.
  • a code stream decoding unit one by one, configured to select a target reconstructed background frame corresponding to the target playback viewpoint from a plurality of the reconstructed background frames, and based on the target reconstructed background frame, pair the target reconstruction background frame corresponding to the target playback viewpoint
  • One channel of the image stream is decoded to obtain and play a channel of target video corresponding to the target playback viewpoint.
  • the playback viewpoint selection instruction includes a first viewpoint selection instruction
  • the viewpoint-by-view determination unit is further configured to, when receiving a first-time viewpoint selection instruction sent by the user, acquire the first-time viewpoint selection instruction in the first-time viewpoint selection instruction, and determine the first viewpoint corresponding to the first-time selection viewpoint.
  • the code stream one by one decoding unit is further configured to, according to the first video coding, determine a first reconstructed background frame corresponding to the first target video from a plurality of the reconstructed background frames, and based on the first reconstructed background frame The background frame is reconstructed, and the first image code stream is decoded.
  • the playback viewpoint selection instruction includes a viewpoint switching instruction
  • the viewpoint-by-view determining unit is further configured to: when receiving a viewpoint switching instruction sent by a user, acquire a target switching viewpoint in the viewpoint switching instruction, and determine the second target video corresponding to the target switching viewpoint. video number;
  • the viewpoint-by-view determination unit is further configured to, according to the second video coding, determine a second reconstructed background frame corresponding to the second target video from a plurality of the reconstructed background frames, and based on the second reconstructed background frame For background frames, the second image code stream is decoded.
  • the image code stream decoding module further includes:
  • the inter-frame prediction judgment unit is configured to judge whether the image block corresponding to the image code stream uses the inter-frame prediction mode when decoding each of the image code streams, wherein the image code stream is composed of multiple the code stream composition of the image block;
  • a motion vector setting unit configured to set the motion vector in the image block to 0 if the image block uses an inter prediction mode
  • a residual skip decoding unit configured to use the reconstructed background frame corresponding to the image block using the inter-frame prediction mode as a reference frame, and skip the motion vector residual information corresponding to the image block using the inter-frame prediction mode. , and decode the image code stream.
  • the background frame code stream is obtained by encoding a background frame obtained by background modeling according to a video image in the target video.
  • the method for performing background modeling on the video image includes a single-frame generation method, a median filtering method, and an average filtering method.
  • the present application also provides a video decoding device
  • the video decoding device includes: a memory, a processor, and a computer-readable instruction stored on the memory and executable on the processor, The computer-readable instructions, when executed by the processor, implement the steps of the video decoding method as described above.
  • the present application also provides a computer-readable storage medium, where computer-readable instructions are stored on the computer-readable storage medium, and when the computer-readable instructions are executed by a processor, the above video can be realized The steps of the decoding method.
  • the present application provides a video decoding method, system, device, and computer-readable storage medium.
  • This embodiment of the present application obtains multiple reconstructed background frames by receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams;
  • receive the multi-channel image code streams corresponding to different viewpoints of the target video In the viewpoint selection order determined by the playback viewpoint selection instruction, receive the multi-channel image code streams corresponding to different viewpoints of the target video, and decode the multi-channel image code streams of the different viewpoints based on a plurality of the reconstructed background frames, to get multi-channel target video.
  • the present application decodes the background frame code streams corresponding to different viewpoints of the video to be played before decoding the image code stream to obtain the reconstructed background frame, so that all the decoding process of the image code stream is completed in advance.
  • the necessary part of the operation reduces the processing burden of the device when decoding the image code stream and improves the overall video decoding efficiency; by only referring to the independent reference basis of the reconstructed background frame during decoding, the corresponding target video of each channel is determined.
  • the image code stream is decoded. Since the decoding of each image code stream only depends on the reconstructed background frame, and does not depend on other decoded images, it achieves the purpose of real-time free switching and smooth playback between target videos corresponding to different viewpoints.
  • the method of inter-frame dependence has higher video compression efficiency, thereby solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching.
  • FIG. 1 is a schematic structural diagram of a video decoding device of a hardware operating environment involved in a solution of an embodiment of the present application
  • FIG. 2 is a schematic flowchart of a first embodiment of a video decoding method of the present application
  • FIG. 3 is a system structure diagram in a specific embodiment of the video decoding method of the present application.
  • FIG. 4 is a schematic diagram of panorama image division in another specific embodiment of the video decoding method of the present application.
  • FIG. 5 is a schematic diagram of encoding/decoding and code stream storage in another specific embodiment of the video decoding method of the present application.
  • FIG. 1 is a schematic structural diagram of a video decoding device of a hardware operating environment involved in an embodiment of the present application.
  • the video decoding device in this embodiment of the present application may be a server, a PC, or a terminal device such as a smart phone or a tablet computer.
  • the video decoding device may include: a processor 1001 , such as a CPU, a communication bus 1002 , a user interface 1003 , a network interface 1004 , and a memory 1005 .
  • the communication bus 1002 is used to realize the connection and communication between these components.
  • the user interface 1003 may include a display screen (Display), an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface.
  • the network interface 1004 may include a standard wired interface and a wireless interface (eg, a WI-FI interface).
  • the memory 1005 may be a high-speed RAM memory or a non-volatile memory (non-volatile memory). memory), such as disk storage.
  • the memory 1005 may also be a storage device independent of the aforementioned processor 1001 .
  • the terminal structure shown in FIG. 1 does not constitute a limitation on the video decoding device, and may include more or less components than the one shown, or combine some components, or arrange different components.
  • the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module, and computer-readable instructions.
  • the network interface 1004 is mainly used to connect to the background server and perform data communication with the background server;
  • the user interface 1003 is mainly used to connect to the client (client) and perform data communication with the client;
  • the processor 1001 can be used to call the computer-readable instructions stored in the memory 1005, and execute and execute the video decoding method provided by the embodiment of the present invention.
  • Free viewpoint applications allow viewers to watch videos in a range of continuous viewpoints.
  • the viewer can set the position and angle of the viewpoint, instead of being limited to a fixed camera angle.
  • This application often requires multiple cameras to shoot at the same time to generate videos from multiple viewpoints at the same time.
  • Virtual reality technology allows users to watch videos in a 180- or 360-degree view. Therefore, whether it is a multi-channel video in a free viewpoint application or a panoramic video in a virtual reality application, there is a very large amount of data, which brings great challenges to video transmission.
  • the commonly used methods of encoding and decoding such videos are usually divided into two categories. One is the method with inter-frame dependence. Although this method can achieve higher compression efficiency, it needs to rely on other decoded images during decoding.
  • the present application provides a video decoding method, that is, before decoding the image code stream, the background frame code stream corresponding to different viewpoints of the video to be played is decoded to obtain the reconstructed background frame, so that the completion of the process is completed in advance. It eliminates some necessary operations when decoding the image code stream, reduces the processing burden of the device when decoding the image code stream, and improves the overall video decoding efficiency; by only referring to the independent reference basis for reconstructing the background frame during decoding Decode the image stream corresponding to each target video. Since the decoding of each image stream only depends on the reconstructed background frame, not on other decoded images, it can achieve real-time free switching between target videos corresponding to different viewpoints and smooth playback. At the same time, it has higher video compression efficiency than the method without inter-frame dependence, thus solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching. .
  • FIG. 2 is a schematic flowchart of a first embodiment of a video decoding method.
  • the first embodiment of the present application provides a video decoding method, the video decoding method is applied to a decoding end, and the video decoding method includes the following steps:
  • Step S10 receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
  • Step S20 when receiving the playback viewpoint selection instruction, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, receive multiple image code streams corresponding to different viewpoints of the target video, and based on a plurality of the reconstructed background frames.
  • the multi-channel image code streams of different viewpoints are decoded to obtain multi-channel target videos.
  • the decoding end can acquire, in real time, the multi-channel image code streams output from the encoding end after compressing and encoding the multi-channel target videos, and each background frame code stream corresponding to each channel of the target video.
  • the decoding end decodes each background frame code stream according to the coding standard consistent with the encoding end to obtain the corresponding reconstructed background frame.
  • the decoding end uses the reconstructed background frame as a reference frame to decode the image code stream and output it in real time, so as to obtain the video image frames composing the multi-channel target video and play the multi-channel target video. It should be noted that the operation of decoding the background frame code stream by the decoding end is completed before decoding and playing the image code stream.
  • the decoding end When the decoding end decodes multiple image streams and background frame streams, it can start a decoder for each target video.
  • the background frame streams and image streams of a certain target video use the corresponding video of this channel.
  • the decoder can be used for decoding; or only one decoder can be started, and the background frame code stream and image code stream of each video are decoded using this decoder.
  • 16 cameras need to be deployed to shoot video, each camera is a shooting viewpoint, and 16 background generators are used to capture video. And 16 encoders to generate 16 real-time code streams (including background frame code stream and image code stream), and store the real-time code stream in the http server for the client to download.
  • the client receives the background frame streams in the 16 real-time streams, and creates 16 decoders to decode the 16 background frame streams respectively; during the playback process, the video number and video frame to be played at the next moment are selected in real time The serial number in the video stream is obtained, and the image code stream in the real-time code stream is obtained; finally, the decoder corresponding to the video number is called to decode, and after the decoded video image is obtained, it is transmitted to the playback module of the client for playback.
  • each background generator collects 100 frames of images from the corresponding cameras; in the second step, the background generator generates a background frame through median filtering, that is, each pixel
  • the pixel values of the corresponding pixels on the background frame are obtained by the following operations: Obtain 100 pixel values of the corresponding pixels in 100 frames, sort the pixel values by numerical value, and select the 50th pixel value as the pixel value of the corresponding pixel on the background frame. .
  • each encoder uses the AVS2 coding standard to perform I-frame (intra-frame prediction image) encoding on the background frame of the corresponding channel and output the code stream, and obtain the reconstructed background frame at the same time.
  • each encoder continues to obtain the to-be-encoded image from the corresponding camera.
  • each encoder encodes the acquired image to be encoded and outputs a code stream.
  • the encoding process uses the reconstructed background frame of the corresponding road as the reference frame, and encodes the to-be-encoded image based on the S-frame (single-forward inter-frame decoded image that should refer to the scene image) encoding of the AVS2 encoding standard.
  • the fourth step and the fifth step above are repeatedly performed, and the code stream of the video image is continuously generated and output.
  • the process of video stream transmission and decoding is as follows: the first step is to transmit the background frame code stream corresponding to each channel of video, allocate a decoder for each channel of video, and analyze the background frame code stream of the corresponding channel. Decoding is performed to obtain the reconstructed background frame corresponding to each channel of video. The decoding process uses the I-frame decoding method of the AVS2 coding standard for decoding. The second step is to cache the reconstructed background frames of each video. The reconstructed background frame of each channel is buffered inside the corresponding decoder. In the third step, the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained.
  • the specific acquisition method is: during playback, determine the currently playing video number K according to the currently playing viewpoint selected by the user, and calculate the frame number L in the corresponding video of the image that should be displayed at the current moment according to the current moment.
  • the fourth step is to request and transmit the L-th frame code stream of the K-th video from the http service.
  • the fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image and provide it to the playback module for display.
  • the decoding process is carried out in the S-frame mode of the AVS2 coding standard.
  • the third to fifth steps above are repeatedly performed until the playback ends.
  • the original panoramic video is divided into regions as shown in FIG. 4, and the image of the original panoramic video is divided into 9 regions from 0 to 8 in total , the picture of each area constitutes one video, a total of 9 video.
  • the encoding and decoding process and the code stream storage method are shown in Figure 5.
  • the corresponding background frame is first generated for each channel of video, and the encoder first encodes the background frames of each channel of video in sequence to generate a code stream, and obtains the reconstructed background frame.
  • the decoder firstly decodes each background frame in the code stream one by one and buffers it.
  • the playback process at any time, according to the playback focus selected by the user, determine the number of channels that currently need to be displayed, read the code stream corresponding to the frame, and then use the reconstructed background frame of the corresponding channel as the reference frame to decode the code stream of the frame. Get the decoded image.
  • the specific encoding process at the encoding end in this embodiment is as follows: the first step is to acquire the first 100 frames of images of each channel of video; the second step is to use the first 100 frames of images of each channel of video to generate background frames of each channel of video through mean filtering. That is, perform the following operations on each pixel to obtain the pixel value of the corresponding pixel on the background frame: obtain 100 pixel values of the corresponding pixel in 100 frames, calculate the average value of the 100 pixel values, and use it as the pixel value of the corresponding pixel on the background frame. .
  • the third step is to use the H.265 coding standard to perform I-frame coding on each background frame in sequence and output the code stream, and obtain the reconstructed background frame at the same time.
  • the current to-be-encoded image is acquired according to the picture sequence and the time sequence.
  • the specific sequence is shown in Figure 5.
  • the time sequence is prioritized, and the first frame of all the ways is encoded first, then the second frame of all the ways is encoded, and so on.
  • the fifth step is to encode the acquired image to be encoded and output the code stream.
  • the encoding process uses the reconstructed background frame of the corresponding road as a reference frame, and encodes the to-be-encoded image based on the P-frame encoding method of the H.265 encoding standard.
  • the fourth step and the fifth step above are repeatedly performed, and the code stream of the video image is continuously generated and output.
  • the specific process of reading and decoding is as follows: the first step is to read and decode the background frame code stream corresponding to each channel of video to obtain the reconstructed background frame corresponding to each channel of video.
  • the decoding process uses the I-frame decoding method of the H.265 coding standard for decoding.
  • the second step is to cache the reconstructed background frames of each video.
  • the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained.
  • the specific acquisition method is: during playback, determine the current corresponding video number K according to the current playback focus selected by the user, and calculate the frame number L in the corresponding video of the image that should be displayed at the current moment according to the current moment.
  • the fourth step is to read the L-th frame code stream of the K-th video.
  • the fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image.
  • the decoding process is decoded in the way of P frames of the h.265 coding standard.
  • the third to fifth steps above are repeatedly performed until the playback ends.
  • the background frame code stream corresponding to different viewpoints of the video to be played is decoded to obtain the reconstructed background frame, so that all the decoding process of the image code stream is completed in advance.
  • the necessary part of the operation reduces the processing burden of the device when decoding the image code stream and improves the overall video decoding efficiency; by only referring to the independent reference basis of the reconstructed background frame during decoding, the corresponding target video of each channel is determined.
  • the image code stream is decoded. Since the decoding of each image code stream only depends on the reconstructed background frame, and does not depend on other decoded images, it achieves the purpose of real-time free switching and smooth playback between target videos corresponding to different viewpoints.
  • the method of inter-frame dependence has higher video compression efficiency, thereby solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching.
  • step S20 includes:
  • Step a when receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoint currently selected by the user one by one based on the playback viewpoint selection instruction, and obtain a path corresponding to the target playback viewpoint generated by the encoding end. image stream;
  • Step b selecting a target reconstructed background frame corresponding to the target playback viewpoint from a plurality of the reconstructed background frames, and based on the target reconstructed background frame to the one image code stream corresponding to the target playback viewpoint.
  • Decoding is performed to obtain and play a target video corresponding to the target playback viewpoint.
  • the decoding end can receive and decode the viewpoints according to the real-time selection of viewpoints by the user.
  • the image code stream of the corresponding viewpoint If there are currently eight image streams corresponding to eight different viewpoints and one reconstructed background frame, each image stream can be played as a target video after decoding. If the user selects the first view point first, the decoding end can directly use the reconstructed background frame of the first view point obtained by decoding the background frame code stream before to decode and output the first image code stream for playback.
  • the target video corresponding to the first viewpoint is the decoding end.
  • the decoding end can receive the second image stream, and directly select the reconstructed background frame corresponding to the second viewpoint for its corresponding position for subsequent decoding and playback. Before the entire video is played, the user can switch the current playback viewpoint at any time, and the decoding end will receive the corresponding number of image streams at any time, and decode it through the existing reconstructed background frames, so as to realize the video switching of different viewpoints. play.
  • the image code stream is generated by the encoder and sent directly to the decoder; for the application scenarios of on-demand and local playback, the image code stream is generated by the encoder in advance and then uploaded to the server or locally. Then, when the user needs to watch, the decoding end obtains it from the server or from a local file.
  • the playback viewpoint selection instruction includes a first viewpoint selection instruction, and step a includes:
  • Step c when receiving the first viewpoint selection instruction sent by the user, obtain the first selected viewpoint in the first viewpoint selection instruction, and determine the first video number of the first target video corresponding to the first selected viewpoint ;
  • Step d determine the first image number of the video image at the initial position in the first target video
  • Step e Receive a first image code stream corresponding to the first video number and the first image number from the encoding end.
  • Step b includes:
  • Step f according to the first video encoding, determine a first reconstructed background frame corresponding to the first target video from a plurality of the reconstructed background frames, and based on the first reconstructed background frame, perform an analysis on the first reconstructed background frame.
  • An image code stream is decoded.
  • the playback viewpoint selection instruction includes a viewpoint switching instruction
  • Step a also includes:
  • Step g when receiving the viewpoint switching instruction sent by the user, acquiring the target switching viewpoint in the viewpoint switching instruction, and determining the second video number of the second target video corresponding to the target switching viewpoint;
  • Step h acquiring the playback progress of the first target video, and determining the second image number of the second video image corresponding to the playback progress in the second target video;
  • Step i receive the second image code stream corresponding to the second video number and the second image number from the encoding end;
  • Step b further includes: the first viewpoint selection instruction is an instruction initiated by the user to perform viewpoint selection for the first time.
  • the first target video is one target video corresponding to the viewpoint selected for the first time.
  • Step j according to the second video encoding, determine a second reconstructed background frame corresponding to the second target video from a plurality of the reconstructed background frames, and based on the second reconstructed background frame, perform an analysis on the first reconstructed background frame.
  • the two-image code stream is decoded.
  • the first-time viewpoint selection instruction is an instruction issued by the user when the user selects a viewpoint for the first time.
  • the viewpoint selected for the first time is the viewpoint currently selected for playback for the first time.
  • the first target video is a target video corresponding to the viewpoint selected for the first time.
  • the first video number is the belonging number of the first target video.
  • the first image number is a video image number corresponding to the multi-frame video images composing the first target video when the current video playback progress is zero.
  • the first image code stream is the compressed image code stream of the video image corresponding to the first video number.
  • the first reconstructed background frame is a reconstructed background frame corresponding to the viewpoint selected for the first time.
  • the viewpoint switching instruction is an instruction issued when the user intends to switch the currently selected viewpoint.
  • the target switching viewpoint is the new playback viewpoint currently selected by the user.
  • the second target video is a target video corresponding to the target switching viewpoint.
  • the second video number is the belonging number of the second target video.
  • the second image number is a video image number corresponding to the multiple frames of video images composing the second target video according to the current video playback progress.
  • the second image code stream is the compressed image code stream of the video image corresponding to the second video number.
  • the second reconstructed background frame is a reconstructed background frame corresponding to the target switching viewpoint.
  • the decoding end transmits or reads the background frame code stream corresponding to each channel of video, and decodes it to obtain the reconstructed background frame corresponding to each channel of video.
  • the decoding end uses the corresponding encoding standard to decode the background frame code stream.
  • the second step is to cache the reconstructed background frames of each video.
  • the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained.
  • the specific acquisition method is: when playing, select the corresponding video number K according to the viewpoint or focus currently selected by the user, and at the same time, according to the current moment, obtain the frame number L of the corresponding image in the corresponding video; the fourth step is to transmit or read The L-th frame code stream of the K-th video.
  • the fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image.
  • the decoding end uses the corresponding encoding standard to decode the background frame.
  • the third to fifth steps above are repeatedly performed until the playback ends.
  • the step of decoding the multi-channel image code streams of different viewpoints based on a plurality of the reconstructed background frames in this embodiment includes:
  • Step k when decoding each of the image code streams, determine that the image block corresponding to the image code stream uses the inter prediction mode, and set the motion vector in the image block to 0, wherein the The image code stream is composed of a plurality of code streams of the image blocks;
  • Step 1 using the reconstructed background frame corresponding to the image block in the inter-frame prediction mode as a reference frame, and by skipping the motion vector residual information corresponding to the image block using the inter-frame prediction mode, to the image code. stream to decode.
  • the encoding process is to cut the image into image blocks one by one as encoding units, and each image block can be selected from inter-frame prediction or intra-frame prediction.
  • the inter prediction mode is used, the motion vectors in the x and y directions are both 0 by default, so the encoder may not encode the motion vector residual information, thereby saving encoding overhead.
  • the encoding end determines whether the image blocks in each video image use the inter-frame prediction mode one by one, and when detecting that the current video image uses the inter-frame prediction mode, the motion vector in the current video image is set to 0 to skip the
  • the process of encoding the motion vector residual information corresponding to the current video image, and at the decoding end, the image code stream is composed of code streams of a plurality of the image blocks. If the encoding end uses the default 0 motion vector for encoding, when the decoding end detects that the previous image block is encoded in the inter-frame prediction mode when decoding the image code stream, the motion vector corresponding to the image block can be used for encoding. Defaults to 0 to skip decoding motion vector residual information.
  • the background frame code stream is obtained by encoding a background frame obtained by background modeling according to a video image in the target video.
  • the method for performing background modeling on the video image includes a single-frame generation method, a median filtering method, and an average filtering method.
  • the single-frame generation method is to directly select a frame in the video of this channel, for example, select the first frame or the last frame or a frame in the center as the background frame;
  • the middle-finger filtering method is to select H frames from the video of this channel. , sort the H pixel values of each pixel point, and select the pixel at the position of H/2 or H/2-1 or H/2+1 as the filter output value of the pixel point, and then obtain the background frame;
  • the filtering method is to select H frames from the video, and calculate the average value of H pixel values at each pixel position as the filtering output value of the pixel, and then obtain the background frame.
  • the encoding end acquires multiple target videos collected based on different viewpoints, and generates a background frame corresponding to each target video one by one based on a single frame generation method, a median filtering method or a preset background modeling method of mean filtering;
  • the encoding end traverses all the target videos to obtain each background frame corresponding to each target video, and then encodes each background frame to obtain the background frame code stream and the reconstructed background frame generated during the encoding process.
  • the application also provides a video decoding system, which implements the following steps:
  • the background code stream decoding module is used for receiving multiple background frame code streams corresponding to different viewpoints sent by the encoder, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
  • the image code stream decoding module is used for receiving the multi-channel image code streams corresponding to different viewpoints of the target video according to the viewpoint selection sequence determined by the playback viewpoint selection instruction when receiving the playback viewpoint selection instruction, and based on a plurality of The reconstructed background frame decodes the multi-channel image code streams of different viewpoints to obtain multi-channel target videos.
  • the present application also provides a video decoding device.
  • the video decoding apparatus includes a processor, a memory, and computer-readable instructions stored on the memory and executable on the processor, wherein when the computer-readable instructions are executed by the processor, the implementation is as above. The steps of the video decoding method.
  • the present application also provides a computer-readable storage medium.
  • Computer-readable instructions are stored on the computer-readable storage medium of the present application, and when the computer-readable instructions are executed by a processor, implement the steps of the video decoding method described above.
  • the method of the above embodiment can be implemented by means of software plus a necessary general hardware platform, and of course can also be implemented by hardware, but in many cases the former is better implementation.
  • the technical solutions of the present application can be embodied in the form of software products in essence or the parts that make contributions to the prior art.
  • the computer software products are stored in a storage medium (such as ROM/RAM) as described above. , magnetic disk, optical disk), including several instructions to make a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.

Abstract

Disclosed in the present application are a video decoding method, system and device and a computer-readable storage medium. In the video decoding method, before an image code stream is decoded, background frame code streams corresponding to different viewpoints of a video to be playbacked are firstly decoded to obtain reconstructed background frames, so that part of necessary operations during image code stream decoding are completed in advance, thus reducing the processing burden of a device to perform image code stream decoding and improving the overall video decoding efficiency; the image code stream corresponding to each target video is decoded by only referring to the independent reference basis of the reconstructed background frames during the decoding, and decoding of each image code stream only relies on the reconstructed background frames and does not rely on other decoded images, so that the purposes of real-time free switch and smooth playback of target videos corresponding to different viewpoints are achieved; and meanwhile, compared with a mode completely without inter-frame dependence, the method is higher in video compression efficiency and simple in system implementation.

Description

视频解码方法、系统、设备及计算机可读存储介质Video decoding method, system, device, and computer-readable storage medium
本申请要求于2020年7月29日提交中国专利局、申请号为202010748734.3、发明名称为“视频解码方法、系统、设备及计算机可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202010748734.3 and the invention title "video decoding method, system, device and computer-readable storage medium" filed with the China Patent Office on July 29, 2020, the entire contents of which are passed Reference is incorporated in this application.
技术领域technical field
本申请涉及数字信号处理技术领域,尤其涉及一种视频解码方法、系统、设备及计算机可读存储介质。The present application relates to the technical field of digital signal processing, and in particular, to a video decoding method, system, device, and computer-readable storage medium.
背景技术Background technique
自由视点应用允许观看者在一定范围内以连续视点的形式观看视频。观看者可以设定视点的位置、角度,而不再局限于一个固定的摄像机视角。该应用往往需要多个摄像机同时拍摄,同时生成多个视点的视频。虚拟现实技术则是允许用户以180度或360度的视角观看视频。因此,不管是自由视点应用中的多路视频还是虚拟现实应用中的全景视频,都具有非常大的数据量,对视频的传输带来了很大的挑战。目前常用的编解码此类视频的方式通常分为两类,一类是有帧间依赖的方式,这一方式虽能实现较高的压缩效率,但由于在解码时需依赖已解码的其他图像,因此在视点切换时流畅度不佳;另一类是完全无帧间依赖的方式,这一方式虽能实现较好的视点切换流畅度,但是由于压缩时无帧间依赖,压缩效率并不理想。因此,上述种种情况反映出现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的问题。Free viewpoint applications allow viewers to watch videos in a range of continuous viewpoints. The viewer can set the position and angle of the viewpoint, instead of being limited to a fixed camera angle. This application often requires multiple cameras to shoot at the same time to generate videos from multiple viewpoints at the same time. Virtual reality technology allows users to watch videos in a 180- or 360-degree view. Therefore, whether it is a multi-channel video in a free viewpoint application or a panoramic video in a virtual reality application, there is a very large amount of data, which brings great challenges to video transmission. At present, the commonly used methods of encoding and decoding such videos are usually divided into two categories. One is the method with inter-frame dependence. Although this method can achieve higher compression efficiency, it needs to rely on other decoded images during decoding. , so the fluency is not good when switching viewpoints; the other is the way that there is no inter-frame dependence. Although this way can achieve better fluency of viewpoint switching, but because there is no inter-frame dependence during compression, the compression efficiency is not high. ideal. Therefore, the above situations reflect the problem that the existing encoding and decoding methods for multi-channel video or panoramic video are difficult to take into account both the compression efficiency and the smoothness of viewpoint switching.
技术解决方案technical solutions
本申请的主要目的在于提供一种视频解码方法、系统、设备及计算机可读存储介质旨在解决通过现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的技术问题。The main purpose of this application is to provide a video decoding method, system, device and computer-readable storage medium to solve the technology that it is difficult to take into account the compression efficiency and the smoothness of viewpoint switching through the existing encoding and decoding methods of multi-channel video or panoramic video. problem.
为实现上述目的,本申请提供一种视频解码方法,所述视频解码方法包括:In order to achieve the above object, the present application provides a video decoding method, the video decoding method includes:
接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。When a playback viewpoint selection instruction is received, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, multiple image streams corresponding to different viewpoints of the target video are received, and based on a plurality of the reconstructed background frames, the different The multi-channel image code stream of the viewpoint is decoded to obtain the multi-channel target video.
此外,为实现上述目的,本申请还提供一种视频解码系统,所述视频解码系统包括:In addition, in order to achieve the above purpose, the present application also provides a video decoding system, the video decoding system includes:
背景码流解码模块,用于接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;The background code stream decoding module is used for receiving multiple background frame code streams corresponding to different viewpoints sent by the encoder, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
图像码流解码模块,用于在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。The image code stream decoding module is used for receiving the multi-channel image code streams corresponding to different viewpoints of the target video according to the viewpoint selection sequence determined by the playback viewpoint selection instruction when receiving the playback viewpoint selection instruction, and based on a plurality of The reconstructed background frame decodes the multi-channel image code streams of different viewpoints to obtain multi-channel target videos.
可选地,所述图像码流解码模块包括:Optionally, the image code stream decoding module includes:
视点逐一确定单元,用于在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流;The viewpoint-by-view determination unit is configured to, when receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoints currently selected by the user one by one based on the playback viewpoint selection instruction, and obtain the target playback viewpoints generated by the encoding end and the target playback viewpoints one by one. One image stream corresponding to the viewpoint;
码流逐一解码单元,用于从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码,以得到并播放与所述目标播放视点对应的一路目标视频。A code stream decoding unit one by one, configured to select a target reconstructed background frame corresponding to the target playback viewpoint from a plurality of the reconstructed background frames, and based on the target reconstructed background frame, pair the target reconstruction background frame corresponding to the target playback viewpoint One channel of the image stream is decoded to obtain and play a channel of target video corresponding to the target playback viewpoint.
可选地,所述播放视点选择指令包括首次视点选择指令,Optionally, the playback viewpoint selection instruction includes a first viewpoint selection instruction,
所述视点逐一确定单元还用于,在接收到由用户发送的首次视点选择指令时,获取所述首次视点选择指令中的首次所选视点,并确定与所述首次所选视点对应的第一目标视频的第一视频编号;The viewpoint-by-view determination unit is further configured to, when receiving a first-time viewpoint selection instruction sent by the user, acquire the first-time viewpoint selection instruction in the first-time viewpoint selection instruction, and determine the first viewpoint corresponding to the first-time selection viewpoint. The first video number of the target video;
确定所述第一目标视频中位于初始位置的视频图像的第一图像编号;determining the first image number of the video image at the initial position in the first target video;
从编码端接收与所述第一视频编号以及所述第一图像编号对应的第一图像码流;Receive a first image code stream corresponding to the first video number and the first image number from the encoding end;
所述码流逐一解码单元还用于,根据所述第一视频编码,从多个所述重建背景帧中确定与所述第一目标视频对应的第一重建背景帧,并基于所述第一重建背景帧,对所述第一图像码流进行解码。The code stream one by one decoding unit is further configured to, according to the first video coding, determine a first reconstructed background frame corresponding to the first target video from a plurality of the reconstructed background frames, and based on the first reconstructed background frame The background frame is reconstructed, and the first image code stream is decoded.
可选地,所述播放视点选择指令包括视点切换指令,Optionally, the playback viewpoint selection instruction includes a viewpoint switching instruction,
所述视点逐一确定单元还用于:在接收到用户发送的视点切换指令时,获取所述视点切换指令中的目标切换视点,并确定与所述目标切换视点对应的第二目标视频的第二视频编号;The viewpoint-by-view determining unit is further configured to: when receiving a viewpoint switching instruction sent by a user, acquire a target switching viewpoint in the viewpoint switching instruction, and determine the second target video corresponding to the target switching viewpoint. video number;
获取所述第一目标视频的播放进度,并确定所述第二目标视频中与所述播放进度对应的第二视频图像的第二图像编号;Acquire the playback progress of the first target video, and determine the second image number of the second video image corresponding to the playback progress in the second target video;
从编码端接收与所述第二视频编号以及所述第二图像编号对应的第二图像码流;Receive a second image code stream corresponding to the second video number and the second image number from the encoding end;
所述视点逐一确定单元还用于,根据所述第二视频编码,从多个所述重建背景帧中确定与所述第二目标视频对应的第二重建背景帧,并基于所述第二重建背景帧,对所述第二图像码流进行解码。The viewpoint-by-view determination unit is further configured to, according to the second video coding, determine a second reconstructed background frame corresponding to the second target video from a plurality of the reconstructed background frames, and based on the second reconstructed background frame For background frames, the second image code stream is decoded.
可选地,所述图像码流解码模块还包括:Optionally, the image code stream decoding module further includes:
帧间预测判断单元,用于在对每一所述图像码流进行解码时,判断与所述图像码流对应的图像块是否使用帧间预测模式,其中,所述图像码流是由多个所述图像块的码流组成;The inter-frame prediction judgment unit is configured to judge whether the image block corresponding to the image code stream uses the inter-frame prediction mode when decoding each of the image code streams, wherein the image code stream is composed of multiple the code stream composition of the image block;
运动矢量设置单元,用于若所述图像块使用帧间预测模式,则将所述图像块中的运动矢量设置为0;a motion vector setting unit, configured to set the motion vector in the image block to 0 if the image block uses an inter prediction mode;
残差跳过解码单元,用于将使用帧间预测模式的图像块对应的重建背景帧作为参考帧,并按照跳过所述使用帧间预测模式的图像块对应的运动矢量残差信息的方式,对所述图像码流进行解码。A residual skip decoding unit, configured to use the reconstructed background frame corresponding to the image block using the inter-frame prediction mode as a reference frame, and skip the motion vector residual information corresponding to the image block using the inter-frame prediction mode. , and decode the image code stream.
可选地,所述背景帧码流为根据所述目标视频中的视频图像经过背景建模获取的背景帧进行编码所得。Optionally, the background frame code stream is obtained by encoding a background frame obtained by background modeling according to a video image in the target video.
可选地,所述对所述视频图像进行背景建模的方式包括单帧生成方式、中值滤波方式与均值滤波方式。Optionally, the method for performing background modeling on the video image includes a single-frame generation method, a median filtering method, and an average filtering method.
此外,为实现上述目的,本申请还提供一种视频解码设备,所述视频解码设备包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机可读指令,所述计算机可读指令被所述处理器执行时实现如上述的视频解码方法的步骤。In addition, in order to achieve the above object, the present application also provides a video decoding device, the video decoding device includes: a memory, a processor, and a computer-readable instruction stored on the memory and executable on the processor, The computer-readable instructions, when executed by the processor, implement the steps of the video decoding method as described above.
此外,为实现上述目的,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机可读指令,所述计算机可读指令被处理器执行时实现如上述的视频解码方法的步骤。In addition, in order to achieve the above purpose, the present application also provides a computer-readable storage medium, where computer-readable instructions are stored on the computer-readable storage medium, and when the computer-readable instructions are executed by a processor, the above video can be realized The steps of the decoding method.
本申请提供一种视频解码方法、系统、设备及计算机可读存储介质。本申请实施例通过接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。通过上述方式,本申请通过在对图像码流进行解码之前,先将待播放视频的不同视点所对应的背景帧码流进行解码,得到重建背景帧,使得提前完成了解码图像码流时的所必须的部分操作,减轻了设备在进行图像码流解码时的处理负担,提升了整体的视频解码效率;通过在解码时仅参考重建背景帧的这一独立参照依据对各路目标视频所对应的图像码流进行解码,由于各图像码流解码仅依赖重建背景帧,而不依赖已解码的其他图像,因此达到不同视点对应目标视频间的实时自由切换并流畅播放的目的,同时较之完全无帧间依赖的方式,具有更高的视频压缩效率,从而解决了现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的技术问题。The present application provides a video decoding method, system, device, and computer-readable storage medium. This embodiment of the present application obtains multiple reconstructed background frames by receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams; In the viewpoint selection order determined by the playback viewpoint selection instruction, receive the multi-channel image code streams corresponding to different viewpoints of the target video, and decode the multi-channel image code streams of the different viewpoints based on a plurality of the reconstructed background frames, to get multi-channel target video. In the above manner, the present application decodes the background frame code streams corresponding to different viewpoints of the video to be played before decoding the image code stream to obtain the reconstructed background frame, so that all the decoding process of the image code stream is completed in advance. The necessary part of the operation reduces the processing burden of the device when decoding the image code stream and improves the overall video decoding efficiency; by only referring to the independent reference basis of the reconstructed background frame during decoding, the corresponding target video of each channel is determined. The image code stream is decoded. Since the decoding of each image code stream only depends on the reconstructed background frame, and does not depend on other decoded images, it achieves the purpose of real-time free switching and smooth playback between target videos corresponding to different viewpoints. The method of inter-frame dependence has higher video compression efficiency, thereby solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching.
附图说明Description of drawings
图1是本申请实施例方案涉及的硬件运行环境的视频解码设备结构示意图;1 is a schematic structural diagram of a video decoding device of a hardware operating environment involved in a solution of an embodiment of the present application;
图2为本申请视频解码方法第一实施例的流程示意图;2 is a schematic flowchart of a first embodiment of a video decoding method of the present application;
图3为本申请视频解码方法一具体实施例中的系统结构图;FIG. 3 is a system structure diagram in a specific embodiment of the video decoding method of the present application;
图4为本申请视频解码方法另一具体实施例中全景图像划分示意图;4 is a schematic diagram of panorama image division in another specific embodiment of the video decoding method of the present application;
图5为本申请视频解码方法另一具体实施例中编解码和码流存储示意图。FIG. 5 is a schematic diagram of encoding/decoding and code stream storage in another specific embodiment of the video decoding method of the present application.
本发明的实施方式Embodiments of the present invention
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application.
如图1所示,图1是本申请实施例方案涉及的硬件运行环境的视频解码设备结构示意图。As shown in FIG. 1 , FIG. 1 is a schematic structural diagram of a video decoding device of a hardware operating environment involved in an embodiment of the present application.
本申请实施例视频解码设备可以是服务器、PC,也可以是智能手机、平板电脑等的终端设备。The video decoding device in this embodiment of the present application may be a server, a PC, or a terminal device such as a smart phone or a tablet computer.
如图1所示,该视频解码设备可以包括:处理器1001,例如CPU,通信总线1002,用户接口1003,网络接口1004,存储器1005。其中,通信总线1002用于实现这些组件之间的连接通信。用户接口1003可以包括显示屏(Display)、输入单元比如键盘(Keyboard),可选的用户接口1003还可以包括标准的有线接口、无线接口。网络接口1004可选的可以包括标准的有线接口、无线接口(如WI-FI接口)。存储器1005可以是高速RAM存储器,也可以是稳定的存储器(non-volatile memory),例如磁盘存储器。存储器1005可选的还可以是独立于前述处理器1001的存储设备。As shown in FIG. 1 , the video decoding device may include: a processor 1001 , such as a CPU, a communication bus 1002 , a user interface 1003 , a network interface 1004 , and a memory 1005 . Among them, the communication bus 1002 is used to realize the connection and communication between these components. The user interface 1003 may include a display screen (Display), an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface. Optionally, the network interface 1004 may include a standard wired interface and a wireless interface (eg, a WI-FI interface). The memory 1005 may be a high-speed RAM memory or a non-volatile memory (non-volatile memory). memory), such as disk storage. Optionally, the memory 1005 may also be a storage device independent of the aforementioned processor 1001 .
本领域技术人员可以理解,图1中示出的终端结构并不构成对视频解码设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。Those skilled in the art can understand that the terminal structure shown in FIG. 1 does not constitute a limitation on the video decoding device, and may include more or less components than the one shown, or combine some components, or arrange different components.
如图1所示,作为一种计算机存储介质的存储器1005中可以包括操作系统、网络通信模块、用户接口模块以及计算机可读指令。As shown in FIG. 1 , the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module, and computer-readable instructions.
在图1所示的终端中,网络接口1004主要用于连接后台服务器,与后台服务器进行数据通信;用户接口1003主要用于连接客户端(用户端),与客户端进行数据通信;而处理器1001可以用于调用存储器1005中存储的计算机可读指令,并执行并执行本发明实施例提供的视频解码方法。In the terminal shown in FIG. 1 , the network interface 1004 is mainly used to connect to the background server and perform data communication with the background server; the user interface 1003 is mainly used to connect to the client (client) and perform data communication with the client; and the processor 1001 can be used to call the computer-readable instructions stored in the memory 1005, and execute and execute the video decoding method provided by the embodiment of the present invention.
基于上述硬件结构,提出本申请视频解码方法的各个实施例。Based on the above hardware structure, various embodiments of the video decoding method of the present application are proposed.
自由视点应用允许观看者在一定范围内以连续视点的形式观看视频。观看者可以设定视点的位置、角度,而不再局限于一个固定的摄像机视角。该应用往往需要多个摄像机同时拍摄,同时生成多个视点的视频。虚拟现实技术则是允许用户以180度或360度的视角观看视频。因此,不管是自由视点应用中的多路视频还是虚拟现实应用中的全景视频,都具有非常大的数据量,对视频的传输带来了很大的挑战。目前常用的编解码此类视频的方式通常分为两类,一类是有帧间依赖的方式,这一方式虽能实现较高的压缩效率,但由于在解码时需依赖已解码的其他图像,因此在视点切换时流畅度不佳;另一类是完全无帧间依赖的方式,这一方式虽能实现较好的视点切换流畅度,但是由于压缩时无帧间依赖,压缩效率并不理想。因此,上述种种情况反映出现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的问题。Free viewpoint applications allow viewers to watch videos in a range of continuous viewpoints. The viewer can set the position and angle of the viewpoint, instead of being limited to a fixed camera angle. This application often requires multiple cameras to shoot at the same time to generate videos from multiple viewpoints at the same time. Virtual reality technology allows users to watch videos in a 180- or 360-degree view. Therefore, whether it is a multi-channel video in a free viewpoint application or a panoramic video in a virtual reality application, there is a very large amount of data, which brings great challenges to video transmission. At present, the commonly used methods of encoding and decoding such videos are usually divided into two categories. One is the method with inter-frame dependence. Although this method can achieve higher compression efficiency, it needs to rely on other decoded images during decoding. , so the fluency is not good when switching viewpoints; the other is the way that there is no inter-frame dependence. Although this way can achieve better fluency of viewpoint switching, but because there is no inter-frame dependence during compression, the compression efficiency is not high. ideal. Therefore, the above situations reflect the problem that the existing encoding and decoding methods for multi-channel video or panoramic video are difficult to take into account both the compression efficiency and the smoothness of viewpoint switching.
为解决上述问题,本申请提供一种视频解码方法,即在对图像码流进行解码之前,先将待播放视频的不同视点所对应的背景帧码流进行解码,得到重建背景帧,使得提前完成了解码图像码流时的所必须的部分操作,减轻了设备在进行图像码流解码时的处理负担,提升了整体的视频解码效率;通过在解码时仅参考重建背景帧的这一独立参照依据对各路目标视频所对应的图像码流进行解码,由于各图像码流解码仅依赖重建背景帧,而不依赖已解码的其他图像,因此达到不同视点对应目标视频间的实时自由切换并流畅播放的目的,同时较之完全无帧间依赖的方式,具有更高的视频压缩效率,从而解决了现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的技术问题。In order to solve the above problems, the present application provides a video decoding method, that is, before decoding the image code stream, the background frame code stream corresponding to different viewpoints of the video to be played is decoded to obtain the reconstructed background frame, so that the completion of the process is completed in advance. It eliminates some necessary operations when decoding the image code stream, reduces the processing burden of the device when decoding the image code stream, and improves the overall video decoding efficiency; by only referring to the independent reference basis for reconstructing the background frame during decoding Decode the image stream corresponding to each target video. Since the decoding of each image stream only depends on the reconstructed background frame, not on other decoded images, it can achieve real-time free switching between target videos corresponding to different viewpoints and smooth playback. At the same time, it has higher video compression efficiency than the method without inter-frame dependence, thus solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching. .
参照图2,图2为视频解码方法第一实施例的流程示意图。本申请第一实施例提供一种视频解码方法,所述视频解码方法应用于解码端,所述视频解码方法包括以下步骤:Referring to FIG. 2, FIG. 2 is a schematic flowchart of a first embodiment of a video decoding method. The first embodiment of the present application provides a video decoding method, the video decoding method is applied to a decoding end, and the video decoding method includes the following steps:
步骤S10,接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;Step S10, receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
步骤S20,在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。Step S20, when receiving the playback viewpoint selection instruction, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, receive multiple image code streams corresponding to different viewpoints of the target video, and based on a plurality of the reconstructed background frames. The multi-channel image code streams of different viewpoints are decoded to obtain multi-channel target videos.
在本实施例中,解码端可实时获取从编码端输出的对多路目标视频压缩编码后的多路图像码流,以及各路目标视频分别对应的各个背景帧码流。解码端依据与编码端一致的编码标准,对各个背景帧码流进行解码后得到对应的重建背景帧。解码端以重建背景帧作为参考帧对图像码流进行解码并实时输出,以得到组成多路目标视频的视频图像帧并播放多路目标视频。需要说明的是,解码端对背景帧码流进行解码的操作是在对图像码流进行解码播放之前完成的。解码端在对多路图像码流与背景帧码流进行解码时,可以为每一路目标视频分别启动一个解码器,某一路目标视频的背景帧码流和图像码流,均使用该路视频对应的解码器进行解码;也可以只启动一个解码器,各路视频的背景帧码流和图像码流,均使用该解码器进行解码。In this embodiment, the decoding end can acquire, in real time, the multi-channel image code streams output from the encoding end after compressing and encoding the multi-channel target videos, and each background frame code stream corresponding to each channel of the target video. The decoding end decodes each background frame code stream according to the coding standard consistent with the encoding end to obtain the corresponding reconstructed background frame. The decoding end uses the reconstructed background frame as a reference frame to decode the image code stream and output it in real time, so as to obtain the video image frames composing the multi-channel target video and play the multi-channel target video. It should be noted that the operation of decoding the background frame code stream by the decoding end is completed before decoding and playing the image code stream. When the decoding end decodes multiple image streams and background frame streams, it can start a decoder for each target video. The background frame streams and image streams of a certain target video use the corresponding video of this channel. The decoder can be used for decoding; or only one decoder can be started, and the background frame code stream and image code stream of each video are decoded using this decoder.
作为一具体实施例,针对自由视点直播这一实际应用场景,如图3所示,在该具体实施例中需要部署16台摄像机拍摄视频,每台摄像机为一个拍摄视点,通过16个背景生成器和16个编码器生成16路实时码流(包括背景帧码流与图像码流),并将实时码流存储于http服务器中,供客户端下载。客户端接收16路实时码流中的背景帧码流,并创建16个解码器分别对16个背景帧码流进行解码;在播放过程中,实时选取下一时刻需要播放的视频编号以及视频帧在视频流中的序号,并获取实时码流中的图像码流;最后调用对应视频编号的解码器进行解码,得到解码后的视频图像后,再将其传输至客户端的播放模块进行播放。As a specific embodiment, for the practical application scenario of free-view live broadcast, as shown in FIG. 3 , in this specific embodiment, 16 cameras need to be deployed to shoot video, each camera is a shooting viewpoint, and 16 background generators are used to capture video. And 16 encoders to generate 16 real-time code streams (including background frame code stream and image code stream), and store the real-time code stream in the http server for the client to download. The client receives the background frame streams in the 16 real-time streams, and creates 16 decoders to decode the 16 background frame streams respectively; during the playback process, the video number and video frame to be played at the next moment are selected in real time The serial number in the video stream is obtained, and the image code stream in the real-time code stream is obtained; finally, the decoder corresponding to the video number is called to decode, and after the decoded video image is obtained, it is transmitted to the playback module of the client for playback.
该实施例在编码端的编码过程具体为:第一步,各路背景生成器从对应摄像机采集100帧图像;第二步,背景生成器通过中值滤波生成背景帧,即对每一个像素点进行以下操作得到背景帧上对应像素的像素值:获取100帧中对应像素的100个像素值,对像素值按数值大小进行排序,选取排第50位的像素值作为背景帧上对应像素的像素值。第三步,各路编码器使用AVS2编码标准对对应路的背景帧进行I帧(帧内预测图像)编码并输出码流,同时得到重建背景帧。第四步,各路编码器继续从对应摄像机获取待编码图像。第五步,各路编码器对获取的待编码图像进行编码并输出码流。编码过程使用对应路的重建的背景帧作为参考帧,基于AVS2编码标准的S帧(应参考场景图像的单前向帧间解码图像)编码对待编码图像进行编码。上述第四步和第五步反复执行,不断生成视频图像的码流并输出。The encoding process at the encoding end in this embodiment is as follows: in the first step, each background generator collects 100 frames of images from the corresponding cameras; in the second step, the background generator generates a background frame through median filtering, that is, each pixel The pixel values of the corresponding pixels on the background frame are obtained by the following operations: Obtain 100 pixel values of the corresponding pixels in 100 frames, sort the pixel values by numerical value, and select the 50th pixel value as the pixel value of the corresponding pixel on the background frame. . In the third step, each encoder uses the AVS2 coding standard to perform I-frame (intra-frame prediction image) encoding on the background frame of the corresponding channel and output the code stream, and obtain the reconstructed background frame at the same time. In the fourth step, each encoder continues to obtain the to-be-encoded image from the corresponding camera. In the fifth step, each encoder encodes the acquired image to be encoded and outputs a code stream. The encoding process uses the reconstructed background frame of the corresponding road as the reference frame, and encodes the to-be-encoded image based on the S-frame (single-forward inter-frame decoded image that should refer to the scene image) encoding of the AVS2 encoding standard. The fourth step and the fifth step above are repeatedly performed, and the code stream of the video image is continuously generated and output.
该实施例在解码端,视频流传输和解码的过程具体为:第一步,传输每一路视频对应的背景帧码流,为每一路视频分配一个解码器,并对对应路的背景帧码流进行解码,得到每一路视频对应的重建的背景帧。解码过程使用AVS2编码标准的I帧解码方式进行解码。第二步,缓存每一路视频的重建背景帧。每一路的重建背景帧缓存于对应解码器内部。第三步,获取待解码图像码流所在的视频编号K以及其在视频中的帧编号L。具体的获取方法为:播放时,根据用户选择的当前播放的视点确定当前播放的视频编号K,同时根据当前时刻,计算当前时刻应该显示的图像在对应视频中的帧序号L。第四步,向http服务请求并传输第K路视频的第L帧码流。第五步,以第K路视频的重建背景帧为参考帧解码第K路视频的第L帧码流,得到解码图像并提供给播放模块进行显示。解码过程以AVS2编码标准的S帧方式进行解码。上述第三步至第五步反复执行,直至播放结束。In this embodiment, at the decoding end, the process of video stream transmission and decoding is as follows: the first step is to transmit the background frame code stream corresponding to each channel of video, allocate a decoder for each channel of video, and analyze the background frame code stream of the corresponding channel. Decoding is performed to obtain the reconstructed background frame corresponding to each channel of video. The decoding process uses the I-frame decoding method of the AVS2 coding standard for decoding. The second step is to cache the reconstructed background frames of each video. The reconstructed background frame of each channel is buffered inside the corresponding decoder. In the third step, the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained. The specific acquisition method is: during playback, determine the currently playing video number K according to the currently playing viewpoint selected by the user, and calculate the frame number L in the corresponding video of the image that should be displayed at the current moment according to the current moment. The fourth step is to request and transmit the L-th frame code stream of the K-th video from the http service. The fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image and provide it to the playback module for display. The decoding process is carried out in the S-frame mode of the AVS2 coding standard. The third to fifth steps above are repeatedly performed until the playback ends.
作为另一具体实施例,针对虚拟现实视频编解码及播放这一实际应用场景,按图4的方式对原始的全景视频进行区域划分,将原始全景视频的图像划分成0~8共9个区域,每个区域的画面构成一路视频,一共9路视频。编解码过程以及码流存储方式图如5所示。编码过程中,首先对每路视频生成对应的背景帧,编码器首先按顺序将各路视频的背景帧编码生成码流,并得到重建的背景帧。然后按顺序逐一编码各路视频的第一帧,生成对应码流;然后按顺序逐一编码各路视频的第二帧,生成对应码流,以此类推。其中,编码第N路视频的任意帧时,使用第N路视频的重建背景帧作为参考帧。解码过程中解码器首先逐一解码码流中的各路背景帧,并进行缓存。播放过程中,随时根据用户选择的播放焦点确定当前需要显示的画面所在的路数,读取该帧对应的码流,然后以对应路的重建背景帧作为参考帧进行该帧码流的解码,得到解码图像。As another specific embodiment, for the practical application scenario of virtual reality video encoding, decoding and playback, the original panoramic video is divided into regions as shown in FIG. 4, and the image of the original panoramic video is divided into 9 regions from 0 to 8 in total , the picture of each area constitutes one video, a total of 9 video. The encoding and decoding process and the code stream storage method are shown in Figure 5. During the encoding process, the corresponding background frame is first generated for each channel of video, and the encoder first encodes the background frames of each channel of video in sequence to generate a code stream, and obtains the reconstructed background frame. Then encode the first frame of each channel of video one by one in order to generate a corresponding code stream; then encode the second frame of each channel of video one by one in order to generate a corresponding code stream, and so on. Wherein, when encoding any frame of the Nth video, the reconstructed background frame of the Nth video is used as a reference frame. During the decoding process, the decoder firstly decodes each background frame in the code stream one by one and buffers it. During the playback process, at any time, according to the playback focus selected by the user, determine the number of channels that currently need to be displayed, read the code stream corresponding to the frame, and then use the reconstructed background frame of the corresponding channel as the reference frame to decode the code stream of the frame. Get the decoded image.
该实施例在编码端的具体编码过程为:第一步,获取各路视频前100帧图像;第二步,利用各路视频前100帧图像,通过均值滤波生成各路视频的背景帧。即对每一个像素点进行以下操作得到背景帧上对应像素的像素值:获取100帧中对应像素的100个像素值,计算该100个像素值的平均值,作为背景帧上对应像素的像素值。第三步,使用H.265编码标准,按顺序对各路背景帧进行I帧编码并输出码流,同时得到重建的背景帧。第四步,按画面顺序和时间顺序获取当前待编码图像。具体顺序如图5所示,时间顺序优先,先编码所有路的第一帧,然后编码所有路的第二帧,以此类推。在编码所有路同一时刻的图像帧时,按画面顺序,从第0路到第8路依次编码。第五步,对获取的待编码图像进行编码并输出码流。编码过程使用对应路的重建的背景帧作为参考帧,基于H.265编码标准的P帧编码方式对待编码图像进行编码。上述第四步和第五步反复执行,不断生成视频图像的码流并输出。The specific encoding process at the encoding end in this embodiment is as follows: the first step is to acquire the first 100 frames of images of each channel of video; the second step is to use the first 100 frames of images of each channel of video to generate background frames of each channel of video through mean filtering. That is, perform the following operations on each pixel to obtain the pixel value of the corresponding pixel on the background frame: obtain 100 pixel values of the corresponding pixel in 100 frames, calculate the average value of the 100 pixel values, and use it as the pixel value of the corresponding pixel on the background frame. . The third step is to use the H.265 coding standard to perform I-frame coding on each background frame in sequence and output the code stream, and obtain the reconstructed background frame at the same time. In the fourth step, the current to-be-encoded image is acquired according to the picture sequence and the time sequence. The specific sequence is shown in Figure 5. The time sequence is prioritized, and the first frame of all the ways is encoded first, then the second frame of all the ways is encoded, and so on. When encoding the image frames of all the channels at the same time, they are encoded sequentially from the 0th channel to the 8th channel according to the picture sequence. The fifth step is to encode the acquired image to be encoded and output the code stream. The encoding process uses the reconstructed background frame of the corresponding road as a reference frame, and encodes the to-be-encoded image based on the P-frame encoding method of the H.265 encoding standard. The fourth step and the fifth step above are repeatedly performed, and the code stream of the video image is continuously generated and output.
该实施例在解码端,读取和解码的具体过程为:第一步,读取并解码每一路视频对应的背景帧码流,得到每一路视频对应的重建的背景帧。解码过程使用H.265编码标准的I帧解码方式进行解码。第二步,缓存每一路视频的重建背景帧。第三步,获取待解码图像码流所在的视频编号K以及其在视频中的帧编号L。具体的获取方法为:播放时,根据用户选择的当前播放的焦点确定当前对应的视频编号K,同时根据当前时刻,计算当前时刻应该显示的图像在对应视频中的帧序号L。第四步,读取第K路视频的第L帧码流。第五步,以第K路视频的重建背景帧为参考帧解码第K路视频的第L帧码流,得到解码图像。解码过程以h.265编码标准的P帧方式进行解码。上述第三步至第五步反复执行,直至播放结束。In this embodiment, at the decoding end, the specific process of reading and decoding is as follows: the first step is to read and decode the background frame code stream corresponding to each channel of video to obtain the reconstructed background frame corresponding to each channel of video. The decoding process uses the I-frame decoding method of the H.265 coding standard for decoding. The second step is to cache the reconstructed background frames of each video. In the third step, the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained. The specific acquisition method is: during playback, determine the current corresponding video number K according to the current playback focus selected by the user, and calculate the frame number L in the corresponding video of the image that should be displayed at the current moment according to the current moment. The fourth step is to read the L-th frame code stream of the K-th video. The fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image. The decoding process is decoded in the way of P frames of the h.265 coding standard. The third to fifth steps above are repeatedly performed until the playback ends.
在本实施例中,通过在对图像码流进行解码之前,先将待播放视频的不同视点所对应的背景帧码流进行解码,得到重建背景帧,使得提前完成了解码图像码流时的所必须的部分操作,减轻了设备在进行图像码流解码时的处理负担,提升了整体的视频解码效率;通过在解码时仅参考重建背景帧的这一独立参照依据对各路目标视频所对应的图像码流进行解码,由于各图像码流解码仅依赖重建背景帧,而不依赖已解码的其他图像,因此达到不同视点对应目标视频间的实时自由切换并流畅播放的目的,同时较之完全无帧间依赖的方式,具有更高的视频压缩效率,从而解决了现有的多路视频或全景视频的编解码方式难以兼顾压缩效率与视点切换流畅度的技术问题。In this embodiment, before decoding the image code stream, the background frame code stream corresponding to different viewpoints of the video to be played is decoded to obtain the reconstructed background frame, so that all the decoding process of the image code stream is completed in advance. The necessary part of the operation reduces the processing burden of the device when decoding the image code stream and improves the overall video decoding efficiency; by only referring to the independent reference basis of the reconstructed background frame during decoding, the corresponding target video of each channel is determined. The image code stream is decoded. Since the decoding of each image code stream only depends on the reconstructed background frame, and does not depend on other decoded images, it achieves the purpose of real-time free switching and smooth playback between target videos corresponding to different viewpoints. The method of inter-frame dependence has higher video compression efficiency, thereby solving the technical problem that the existing encoding and decoding methods of multi-channel video or panoramic video cannot take into account the compression efficiency and the smoothness of viewpoint switching.
进一步地,基于上述图2所示的第一实施例,提出本申请视频解码方法的第二实施例。在本实施例中,步骤S20包括:Further, based on the first embodiment shown in FIG. 2 above, a second embodiment of the video decoding method of the present application is proposed. In this embodiment, step S20 includes:
步骤a,在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流;Step a, when receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoint currently selected by the user one by one based on the playback viewpoint selection instruction, and obtain a path corresponding to the target playback viewpoint generated by the encoding end. image stream;
步骤b,从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码,以得到并播放与所述目标播放视点对应的一路目标视频。Step b, selecting a target reconstructed background frame corresponding to the target playback viewpoint from a plurality of the reconstructed background frames, and based on the target reconstructed background frame to the one image code stream corresponding to the target playback viewpoint. Decoding is performed to obtain and play a target video corresponding to the target playback viewpoint.
在本实施例中,解码端在对多个不同视点所对应的背景帧码流进行解码,得到每一不同视点所对应的重建背景帧后,即可根据用户对视点的实时选择来接收并解码对应视点的图像码流。若当前共有对应八路不同视点的八路图像码流以及一重建背景帧,每路图像码流在解码后即可作为一路目标视频进行播放。若用户首先选择第一路视点,则解码端即可直接使用先前已对背景帧码流解码得到的第一路视点的重建背景帧,来对第一路图像码流进行解码并输出,以播放第一视点对应的目标视频。若用户在播放过程中将第一路视点切换到八路视点中的第二路视点,解码端则可接收第二路图像码流,并直接选用第二路视点对应的重建背景帧对其在对应位置进行接续解码并播放。在整个视频播放完成之前,用户可随时对当前播放视点进行切换,解码端即随时接收对应路数的图像码流,并通过已存的重建背景帧对其进行解码,以实现不同视点的视频切换播放。In this embodiment, after decoding the background frame code streams corresponding to multiple different viewpoints to obtain the reconstructed background frame corresponding to each different viewpoint, the decoding end can receive and decode the viewpoints according to the real-time selection of viewpoints by the user. The image code stream of the corresponding viewpoint. If there are currently eight image streams corresponding to eight different viewpoints and one reconstructed background frame, each image stream can be played as a target video after decoding. If the user selects the first view point first, the decoding end can directly use the reconstructed background frame of the first view point obtained by decoding the background frame code stream before to decode and output the first image code stream for playback. The target video corresponding to the first viewpoint. If the user switches the first viewpoint to the second viewpoint among the eight viewpoints during playback, the decoding end can receive the second image stream, and directly select the reconstructed background frame corresponding to the second viewpoint for its corresponding position for subsequent decoding and playback. Before the entire video is played, the user can switch the current playback viewpoint at any time, and the decoding end will receive the corresponding number of image streams at any time, and decode it through the existing reconstructed background frames, so as to realize the video switching of different viewpoints. play.
需要说明的是,在实际应用中,有直播,有点播和本地播放等多种不同的情况。若为直播这一应用场景,则图像码流是由编码端生成后直接发送至解码端;而对于点播和本地播放的应用场景,图像码流由编码端提前生成然后上传至服务器或是本地,然后在用户需要观看的时候,解码端再从服务器上或从本地文件中进行获取。It should be noted that, in practical applications, there are many different situations such as live broadcast, VOD and local playback. For the application scenario of live broadcast, the image code stream is generated by the encoder and sent directly to the decoder; for the application scenarios of on-demand and local playback, the image code stream is generated by the encoder in advance and then uploaded to the server or locally. Then, when the user needs to watch, the decoding end obtains it from the server or from a local file.
进一步地,在本实施例中,所述播放视点选择指令包括首次视点选择指令,步骤a包括:Further, in this embodiment, the playback viewpoint selection instruction includes a first viewpoint selection instruction, and step a includes:
步骤c,在接收到由用户发送的首次视点选择指令时,获取所述首次视点选择指令中的首次所选视点,并确定与所述首次所选视点对应的第一目标视频的第一视频编号;Step c, when receiving the first viewpoint selection instruction sent by the user, obtain the first selected viewpoint in the first viewpoint selection instruction, and determine the first video number of the first target video corresponding to the first selected viewpoint ;
步骤d,确定所述第一目标视频中位于初始位置的视频图像的第一图像编号;Step d, determine the first image number of the video image at the initial position in the first target video;
步骤e,从编码端接收与所述第一视频编号以及所述第一图像编号对应的第一图像码流。Step e: Receive a first image code stream corresponding to the first video number and the first image number from the encoding end.
步骤b包括:Step b includes:
步骤f,根据所述第一视频编码,从多个所述重建背景帧中确定与所述第一目标视频对应的第一重建背景帧,并基于所述第一重建背景帧,对所述第一图像码流进行解码。Step f, according to the first video encoding, determine a first reconstructed background frame corresponding to the first target video from a plurality of the reconstructed background frames, and based on the first reconstructed background frame, perform an analysis on the first reconstructed background frame. An image code stream is decoded.
进一步地,所述播放视点选择指令包括视点切换指令,Further, the playback viewpoint selection instruction includes a viewpoint switching instruction,
步骤a还包括:Step a also includes:
步骤g,在接收到用户发送的视点切换指令时,获取所述视点切换指令中的目标切换视点,并确定与所述目标切换视点对应的第二目标视频的第二视频编号;Step g, when receiving the viewpoint switching instruction sent by the user, acquiring the target switching viewpoint in the viewpoint switching instruction, and determining the second video number of the second target video corresponding to the target switching viewpoint;
步骤h,获取所述第一目标视频的播放进度,并确定所述第二目标视频中与所述播放进度对应的第二视频图像的第二图像编号;Step h, acquiring the playback progress of the first target video, and determining the second image number of the second video image corresponding to the playback progress in the second target video;
步骤i,从编码端接收与所述第二视频编号以及所述第二图像编号对应的第二图像码流;Step i, receive the second image code stream corresponding to the second video number and the second image number from the encoding end;
步骤b还包括:首次视点选择指令为用户所发起的首次进行视点选择的指令。第一目标视频是与首次所选视点对应的一路目标视频。Step b further includes: the first viewpoint selection instruction is an instruction initiated by the user to perform viewpoint selection for the first time. The first target video is one target video corresponding to the viewpoint selected for the first time.
步骤j,根据所述第二视频编码,从多个所述重建背景帧中确定与所述第二目标视频对应的第二重建背景帧,并基于所述第二重建背景帧,对所述第二图像码流进行解码。Step j, according to the second video encoding, determine a second reconstructed background frame corresponding to the second target video from a plurality of the reconstructed background frames, and based on the second reconstructed background frame, perform an analysis on the first reconstructed background frame. The two-image code stream is decoded.
在本实施例中,首次视点选择指令为用户在首次进行视点选择时所发出的指令。首次所选视点为当前首次选择播放的视点。第一目标视频为与首次所选视点对应的一路目标视频。第一视频编号为第一目标视频的所属编号。第一图像编号为当前视频播放进度为零时,对应在组成第一目标视频的多帧视频图像中的视频图像编号。第一图像码流为第一视频编号对应视频图像压缩后的图像码流。第一重建背景帧为,与首次所选视点对应的一重建背景帧。视点切换指令为用户意图对当前所选视点进行切换时所发出的指令。目标切换视点为用户当前选定的新的播放视点。第二目标视频为与目标切换视点对应的一路目标视频。第二视频编号为第二目标视频的所属编号。第二图像编号为依照当前视频播放进度,对应在组成第二目标视频的多帧视频图像中的视频图像编号。第二图像码流为第二视频编号对应视频图像压缩后的图像码流。第二重建背景帧为,与目标切换视点对应的一重建背景帧。In this embodiment, the first-time viewpoint selection instruction is an instruction issued by the user when the user selects a viewpoint for the first time. The viewpoint selected for the first time is the viewpoint currently selected for playback for the first time. The first target video is a target video corresponding to the viewpoint selected for the first time. The first video number is the belonging number of the first target video. The first image number is a video image number corresponding to the multi-frame video images composing the first target video when the current video playback progress is zero. The first image code stream is the compressed image code stream of the video image corresponding to the first video number. The first reconstructed background frame is a reconstructed background frame corresponding to the viewpoint selected for the first time. The viewpoint switching instruction is an instruction issued when the user intends to switch the currently selected viewpoint. The target switching viewpoint is the new playback viewpoint currently selected by the user. The second target video is a target video corresponding to the target switching viewpoint. The second video number is the belonging number of the second target video. The second image number is a video image number corresponding to the multiple frames of video images composing the second target video according to the current video playback progress. The second image code stream is the compressed image code stream of the video image corresponding to the second video number. The second reconstructed background frame is a reconstructed background frame corresponding to the target switching viewpoint.
具体地,解码端传输或读取每一路视频对应的背景帧码流,并对其进行解码,得到每一路视频对应的重建背景帧。解码过程根据编码端对背景帧编码的实际方式,解码端采用对应的编码标准对背景帧码流进行解码。第二步,缓存每一路视频的重建背景帧。第三步,获取待解码图像码流所在的视频编号K以及其在视频中的帧编号L。具体的获取方法为:播放时,根据用户当前选择的视点或焦点选取对应的视频编号K,同时根据当前时刻,得到对应的图像在对应视频中的帧序号L;第四步,传输或读取第K路视频的第L帧码流。第五步,以第K路视频的重建背景帧为参考帧解码第K路视频的第L帧码流,得到解码图像。解码过程根据编码端对该帧图像编码的具体方法,解码端采用对应的编码标准对背景帧进行解码。上述第三步至第五步反复执行,直至播放结束。需要说明的是,在实际情况下,可能在对每一帧图像码流进行解码前,都需要去确定视点,确定使用的重建背景帧,并根据当前播放进度确定图像编号,以此来接收对应的图像码流帧并进行解码播放。Specifically, the decoding end transmits or reads the background frame code stream corresponding to each channel of video, and decodes it to obtain the reconstructed background frame corresponding to each channel of video. In the decoding process, according to the actual way of encoding the background frame by the encoding end, the decoding end uses the corresponding encoding standard to decode the background frame code stream. The second step is to cache the reconstructed background frames of each video. In the third step, the video number K where the image code stream to be decoded is located and the frame number L in the video are obtained. The specific acquisition method is: when playing, select the corresponding video number K according to the viewpoint or focus currently selected by the user, and at the same time, according to the current moment, obtain the frame number L of the corresponding image in the corresponding video; the fourth step is to transmit or read The L-th frame code stream of the K-th video. The fifth step is to decode the L-th frame code stream of the K-th video by taking the reconstructed background frame of the K-th video as a reference frame to obtain a decoded image. In the decoding process, according to the specific method of encoding the frame image at the encoding end, the decoding end uses the corresponding encoding standard to decode the background frame. The third to fifth steps above are repeatedly performed until the playback ends. It should be noted that, in actual situations, it may be necessary to determine the viewpoint, determine the reconstructed background frame to be used, and determine the image number according to the current playback progress before decoding each frame of the image stream, so as to receive the corresponding the image stream frame and decode it for playback.
进一步地,基于上述图2所示的第一实施例,提出本申请视频解码方法的第三实施例。在本实施例中,在本实施例中所述基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码的步骤包括:Further, based on the first embodiment shown in FIG. 2 above, a third embodiment of the video decoding method of the present application is proposed. In this embodiment, the step of decoding the multi-channel image code streams of different viewpoints based on a plurality of the reconstructed background frames in this embodiment includes:
步骤k,在对每一所述图像码流进行解码时,确定与所述图像码流对应的图像块使用帧间预测模式,将所述图像块中的运动矢量设置为0,其中,所述图像码流是由多个所述图像块的码流组成;Step k, when decoding each of the image code streams, determine that the image block corresponding to the image code stream uses the inter prediction mode, and set the motion vector in the image block to 0, wherein the The image code stream is composed of a plurality of code streams of the image blocks;
步骤l,将使用帧间预测模式的图像块对应的重建背景帧作为参考帧,并按照跳过所述使用帧间预测模式的图像块对应的运动矢量残差信息的方式,对所述图像码流进行解码。Step 1, using the reconstructed background frame corresponding to the image block in the inter-frame prediction mode as a reference frame, and by skipping the motion vector residual information corresponding to the image block using the inter-frame prediction mode, to the image code. stream to decode.
在本实施例中,需要说明的是,在编码端,编码过程是把图像切割成一个一个的图像块以作为编码单元,每个图像块可以选择帧间预测或帧内预测。在使用帧间预测模式时,x和y方向的运动矢量均默认为0,因此编码端可不编码运动矢量残差信息,从而节省编码开销。编码端逐一判断每一视频图像中的图像块是否使用帧间预测模式,并在检测到当前视频图像使用帧间预测模式时,将当前视频图像中的运动矢量设置为0,以跳过对所述当前视频图像对应的运动矢量残差信息进行编码的过程而在解码端,图像码流则是由多个所述图像块的码流组成。如果编码端采用了默认0运动矢量的方式进行编码,则解码端在解码图像码流时,在检测到先前的图像块是采用帧间预测模式进行编码时,则可将图像块对应的运动矢量默认为0,以跳过对于运动矢量残差信息进行解码的过程。In this embodiment, it should be noted that, at the encoding end, the encoding process is to cut the image into image blocks one by one as encoding units, and each image block can be selected from inter-frame prediction or intra-frame prediction. When the inter prediction mode is used, the motion vectors in the x and y directions are both 0 by default, so the encoder may not encode the motion vector residual information, thereby saving encoding overhead. The encoding end determines whether the image blocks in each video image use the inter-frame prediction mode one by one, and when detecting that the current video image uses the inter-frame prediction mode, the motion vector in the current video image is set to 0 to skip the The process of encoding the motion vector residual information corresponding to the current video image, and at the decoding end, the image code stream is composed of code streams of a plurality of the image blocks. If the encoding end uses the default 0 motion vector for encoding, when the decoding end detects that the previous image block is encoded in the inter-frame prediction mode when decoding the image code stream, the motion vector corresponding to the image block can be used for encoding. Defaults to 0 to skip decoding motion vector residual information.
进一步地,在本实施例中,所述背景帧码流为根据所述目标视频中的视频图像经过背景建模获取的背景帧进行编码所得。Further, in this embodiment, the background frame code stream is obtained by encoding a background frame obtained by background modeling according to a video image in the target video.
进一步地,在本实施例中,所述对所述视频图像进行背景建模的方式包括单帧生成方式、中值滤波方式与均值滤波方式。Further, in this embodiment, the method for performing background modeling on the video image includes a single-frame generation method, a median filtering method, and an average filtering method.
在本实施例中,单帧生成方法是直接在该路视频中选取一帧,例如选取第一帧或最后一帧或居中一帧作为背景帧;中指滤波方法是从该路视频中选取H帧,对每一个像素点位置的H个像素值排序,选取排第H/2或H/2-1或H/2+1位置的像素作为该像素点的滤波输出值,进而得到背景帧;均值滤波方法是从该路视频中选取H帧,对每一个像素点位置的H个像素值计算平均值,作为该像素点的滤波输出值,进而得到背景帧。首先,编码端获取基于不同视点所采集的多路目标视频,基于单帧生成方式、中值滤波方式或均值滤波的预设背景建模方式逐一生成每路目标视频的所对应的一背景帧;编码端直至遍历全部目标视频,得到各路目标视频对应的各个背景帧,然后再对各个背景帧进行编码,以得到背景帧码流,以及在编码过程中生成的重建背景帧。In this embodiment, the single-frame generation method is to directly select a frame in the video of this channel, for example, select the first frame or the last frame or a frame in the center as the background frame; the middle-finger filtering method is to select H frames from the video of this channel. , sort the H pixel values of each pixel point, and select the pixel at the position of H/2 or H/2-1 or H/2+1 as the filter output value of the pixel point, and then obtain the background frame; The filtering method is to select H frames from the video, and calculate the average value of H pixel values at each pixel position as the filtering output value of the pixel, and then obtain the background frame. First, the encoding end acquires multiple target videos collected based on different viewpoints, and generates a background frame corresponding to each target video one by one based on a single frame generation method, a median filtering method or a preset background modeling method of mean filtering; The encoding end traverses all the target videos to obtain each background frame corresponding to each target video, and then encodes each background frame to obtain the background frame code stream and the reconstructed background frame generated during the encoding process.
本申请还提供一种视频解码系统,所述视频解码系统实现以下步骤:The application also provides a video decoding system, which implements the following steps:
背景码流解码模块,用于接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;The background code stream decoding module is used for receiving multiple background frame code streams corresponding to different viewpoints sent by the encoder, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
图像码流解码模块,用于在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。The image code stream decoding module is used for receiving the multi-channel image code streams corresponding to different viewpoints of the target video according to the viewpoint selection sequence determined by the playback viewpoint selection instruction when receiving the playback viewpoint selection instruction, and based on a plurality of The reconstructed background frame decodes the multi-channel image code streams of different viewpoints to obtain multi-channel target videos.
本申请还提供一种视频解码设备。The present application also provides a video decoding device.
所述种视频解码设备包括处理器、存储器及存储在所述存储器上并可在所述处理器上运行的计算机可读指令,其中所述计算机可读指令被所述处理器执行时,实现如上所述的视频解码方法的步骤。The video decoding apparatus includes a processor, a memory, and computer-readable instructions stored on the memory and executable on the processor, wherein when the computer-readable instructions are executed by the processor, the implementation is as above. The steps of the video decoding method.
其中,所述计算机可读指令被执行时所实现的方法可参照本申请视频解码方法的各个实施例,此处不再赘述。For the method implemented when the computer-readable instruction is executed, reference may be made to the various embodiments of the video decoding method of the present application, which will not be repeated here.
本申请还提供一种计算机可读存储介质。The present application also provides a computer-readable storage medium.
本申请计算机可读存储介质上存储有计算机可读指令,所述计算机可读指令被处理器执行时实现如上所述的视频解码方法的步骤。Computer-readable instructions are stored on the computer-readable storage medium of the present application, and when the computer-readable instructions are executed by a processor, implement the steps of the video decoding method described above.
其中,所述计算机可读指令被执行时所实现的方法可参照本申请视频解码方法各个实施例,此处不再赘述。For the method implemented when the computer-readable instruction is executed, reference may be made to the various embodiments of the video decoding method of the present application, which will not be repeated here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者系统不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者系统所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者系统中还存在另外的相同要素。It should be noted that, herein, the terms "comprising", "comprising" or any other variation thereof are intended to encompass non-exclusive inclusion, such that a process, method, article or system comprising a series of elements includes not only those elements, It also includes other elements not expressly listed or inherent to such a process, method, article or system. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in the process, method, article or system that includes the element.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。The above-mentioned serial numbers of the embodiments of the present application are only for description, and do not represent the advantages or disadvantages of the embodiments.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法。From the description of the above embodiments, those skilled in the art can clearly understand that the method of the above embodiment can be implemented by means of software plus a necessary general hardware platform, and of course can also be implemented by hardware, but in many cases the former is better implementation. Based on this understanding, the technical solutions of the present application can be embodied in the form of software products in essence or the parts that make contributions to the prior art. The computer software products are stored in a storage medium (such as ROM/RAM) as described above. , magnetic disk, optical disk), including several instructions to make a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only the preferred embodiments of the present application, and are not intended to limit the patent scope of the present application. Any equivalent structure or equivalent process transformation made by using the contents of the description and drawings of the present application, or directly or indirectly applied in other related technical fields , are similarly included within the scope of patent protection of this application.

Claims (20)

  1. 一种视频解码方法,其中,所述视频解码方法包括:A video decoding method, wherein the video decoding method comprises:
    接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
    在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。When a playback viewpoint selection instruction is received, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, multiple image streams corresponding to different viewpoints of the target video are received, and based on a plurality of the reconstructed background frames, the different The multi-channel image code stream of the viewpoint is decoded to obtain the multi-channel target video.
  2. 如权利要求1所述的视频解码方法,其中,所述在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频的步骤包括:The video decoding method according to claim 1, wherein when receiving a playback viewpoint selection instruction, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, receive the multi-channel image codes corresponding to different viewpoints of the target video The steps of decoding the multi-channel image code streams of the different viewpoints based on a plurality of the reconstructed background frames to obtain the multi-channel target video include:
    在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流;When receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoints currently selected by the user one by one based on the playback viewpoint selection instruction, and acquire an image code stream corresponding to the target playback viewpoint generated by the encoder ;
    从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码,以得到并播放与所述目标播放视点对应的一路目标视频。A target reconstructed background frame corresponding to the target playback viewpoint is selected from a plurality of the reconstructed background frames, and one image code stream corresponding to the target playback viewpoint is decoded based on the target reconstructed background frame, to obtain and play a target video corresponding to the target playback viewpoint.
  3. 如权利要求2所述的视频解码方法,其中,所述播放视点选择指令包括首次视点选择指令,The video decoding method according to claim 2, wherein the playback viewpoint selection instruction comprises a first viewpoint selection instruction,
    所述在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流的步骤包括:When receiving the playback viewpoint selection instruction sent by the user, the target playback viewpoint currently selected by the user is determined one by one based on the playback viewpoint selection instruction, and an image generated by the encoding end corresponding to the target playback viewpoint is obtained. The steps of the code stream include:
    在接收到由用户发送的首次视点选择指令时,获取所述首次视点选择指令中的首次所选视点,并确定与所述首次所选视点对应的第一目标视频的第一视频编号;When receiving the first viewpoint selection instruction sent by the user, acquiring the first selected viewpoint in the first viewpoint selection instruction, and determining the first video number of the first target video corresponding to the first selected viewpoint;
    确定所述第一目标视频中位于初始位置的视频图像的第一图像编号;determining the first image number of the video image at the initial position in the first target video;
    从编码端接收与所述第一视频编号以及所述第一图像编号对应的第一图像码流;Receive a first image code stream corresponding to the first video number and the first image number from the encoding end;
    所述从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码的步骤包括:The target reconstructed background frame corresponding to the target playback viewpoint is selected from the plurality of reconstructed background frames, and the image code stream corresponding to the target playback viewpoint is performed based on the target reconstructed background frame. The steps of decoding include:
    根据所述第一视频编码,从多个所述重建背景帧中确定与所述第一目标视频对应的第一重建背景帧,并基于所述第一重建背景帧,对所述第一图像码流进行解码。According to the first video coding, a first reconstructed background frame corresponding to the first target video is determined from a plurality of the reconstructed background frames, and based on the first reconstructed background frame, the first image code is stream to decode.
  4. 如权利要求3所述的视频解码方法,其中,所述播放视点选择指令包括视点切换指令,The video decoding method according to claim 3, wherein the playback viewpoint selection instruction comprises a viewpoint switching instruction,
    所述在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流的步骤包括:When receiving the playback viewpoint selection instruction sent by the user, the target playback viewpoint currently selected by the user is determined one by one based on the playback viewpoint selection instruction, and an image generated by the encoding end corresponding to the target playback viewpoint is obtained. The steps of the code stream include:
    在接收到用户发送的视点切换指令时,获取所述视点切换指令中的目标切换视点,并确定与所述目标切换视点对应的第二目标视频的第二视频编号;When receiving the viewpoint switching instruction sent by the user, acquiring the target switching viewpoint in the viewpoint switching instruction, and determining the second video number of the second target video corresponding to the target switching viewpoint;
    获取所述第一目标视频的播放进度,并确定所述第二目标视频中与所述播放进度对应的第二视频图像的第二图像编号;Acquire the playback progress of the first target video, and determine the second image number of the second video image corresponding to the playback progress in the second target video;
    从编码端接收与所述第二视频编号以及所述第二图像编号对应的第二图像码流;Receive a second image code stream corresponding to the second video number and the second image number from the encoding end;
    所述从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码的步骤包括:The target reconstructed background frame corresponding to the target playback viewpoint is selected from the plurality of reconstructed background frames, and the image code stream corresponding to the target playback viewpoint is performed based on the target reconstructed background frame. The steps of decoding include:
    根据所述第二视频编码,从多个所述重建背景帧中确定与所述第二目标视频对应的第二重建背景帧,并基于所述第二重建背景帧,对所述第二图像码流进行解码。According to the second video encoding, a second reconstructed background frame corresponding to the second target video is determined from a plurality of the reconstructed background frames, and based on the second reconstructed background frame, the second image code is stream to decode.
  5. 如权利要求1所述的视频解码方法,其中,所述基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码的步骤包括:The video decoding method according to claim 1, wherein the step of decoding the multi-channel image code streams of the different viewpoints based on a plurality of the reconstructed background frames comprises:
    在对每一所述图像码流进行解码时,确定与所述图像码流对应的图像块使用帧间预测模式,将所述图像块中的运动矢量设置为0,其中,所述图像码流是由多个所述图像块的码流组成;When decoding each of the image code streams, it is determined that the image block corresponding to the image code stream uses the inter prediction mode, and the motion vector in the image block is set to 0, wherein the image code stream is composed of a plurality of code streams of the image blocks;
    将使用帧间预测模式的图像块对应的重建背景帧作为参考帧,并按照跳过所述使用帧间预测模式的图像块对应的运动矢量残差信息的方式,对所述图像码流进行解码。Taking the reconstructed background frame corresponding to the image block using the inter-frame prediction mode as a reference frame, and decoding the image code stream in a manner of skipping the motion vector residual information corresponding to the image block using the inter-frame prediction mode .
  6. 如权利要求1所述的视频解码方法,其中,所述背景帧码流为根据所述目标视频中的视频图像经过背景建模获取的背景帧进行编码所得。The video decoding method according to claim 1, wherein the background frame code stream is obtained by encoding a background frame obtained by background modeling according to a video image in the target video.
  7. 如权利要求6所述的视频解码方法,其中,所述对所述视频图像进行背景建模的方式包括单帧生成方式、中值滤波方式与均值滤波方式。The video decoding method according to claim 6, wherein the method for performing background modeling on the video image includes a single-frame generation method, a median filtering method, and an average filtering method.
  8. 一种视频解码系统,其中,所述视频解码系统包括:A video decoding system, wherein the video decoding system comprises:
    背景码流解码模块,用于接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;The background code stream decoding module is used for receiving multiple background frame code streams corresponding to different viewpoints sent by the encoder, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
    图像码流解码模块,用于在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。The image code stream decoding module is used for receiving the multi-channel image code streams corresponding to different viewpoints of the target video according to the viewpoint selection sequence determined by the playback viewpoint selection instruction when receiving the playback viewpoint selection instruction, and based on a plurality of The reconstructed background frame decodes the multi-channel image code streams of different viewpoints to obtain multi-channel target videos.
  9. 如权利要求8所述的视频解码系统,其中,所述图像码流解码模块包括:The video decoding system according to claim 8, wherein the image code stream decoding module comprises:
    视点逐一确定单元,用于在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流;The viewpoint-by-view determination unit is configured to, when receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoints currently selected by the user one by one based on the playback viewpoint selection instruction, and obtain the target playback viewpoints generated by the encoding end and the target playback viewpoints one by one. One image stream corresponding to the viewpoint;
    码流逐一解码单元,用于从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码,以得到并播放与所述目标播放视点对应的一路目标视频。A code stream decoding unit one by one, configured to select a target reconstructed background frame corresponding to the target playback viewpoint from a plurality of the reconstructed background frames, and based on the target reconstructed background frame, pair the target reconstruction background frame corresponding to the target playback viewpoint One channel of the image stream is decoded to obtain and play a channel of target video corresponding to the target playback viewpoint.
  10. 如权利要求8所述的视频解码系统,其中,所述播放视点选择指令包括首次视点选择指令,The video decoding system of claim 8, wherein the playback viewpoint selection instruction comprises a first viewpoint selection instruction,
    所述视点逐一确定单元还用于,在接收到由用户发送的首次视点选择指令时,获取所述首次视点选择指令中的首次所选视点,并确定与所述首次所选视点对应的第一目标视频的第一视频编号;The viewpoint-by-view determination unit is further configured to, when receiving a first-time viewpoint selection instruction sent by the user, acquire the first-time viewpoint selection instruction in the first-time viewpoint selection instruction, and determine the first viewpoint corresponding to the first-time selection viewpoint. The first video number of the target video;
    确定所述第一目标视频中位于初始位置的视频图像的第一图像编号;determining the first image number of the video image at the initial position in the first target video;
    从编码端接收与所述第一视频编号以及所述第一图像编号对应的第一图像码流;Receive a first image code stream corresponding to the first video number and the first image number from the encoding end;
    所述码流逐一解码单元还用于,根据所述第一视频编码,从多个所述重建背景帧中确定与所述第一目标视频对应的第一重建背景帧,并基于所述第一重建背景帧,对所述第一图像码流进行解码。The code stream one by one decoding unit is further configured to, according to the first video coding, determine a first reconstructed background frame corresponding to the first target video from a plurality of the reconstructed background frames, and based on the first reconstructed background frame The background frame is reconstructed, and the first image code stream is decoded.
  11. 如权利要求9所述的视频解码系统,其中,所述播放视点选择指令包括视点切换指令,The video decoding system according to claim 9, wherein the playback viewpoint selection instruction comprises a viewpoint switching instruction,
    所述视点逐一确定单元还用于:在接收到用户发送的视点切换指令时,获取所述视点切换指令中的目标切换视点,并确定与所述目标切换视点对应的第二目标视频的第二视频编号;The viewpoint-by-view determining unit is further configured to: when receiving a viewpoint switching instruction sent by a user, acquire a target switching viewpoint in the viewpoint switching instruction, and determine the second target video corresponding to the target switching viewpoint. video number;
    获取所述第一目标视频的播放进度,并确定所述第二目标视频中与所述播放进度对应的第二视频图像的第二图像编号;Acquire the playback progress of the first target video, and determine the second image number of the second video image corresponding to the playback progress in the second target video;
    从编码端接收与所述第二视频编号以及所述第二图像编号对应的第二图像码流;Receive a second image code stream corresponding to the second video number and the second image number from the encoding end;
    所述视点逐一确定单元还用于,根据所述第二视频编码,从多个所述重建背景帧中确定与所述第二目标视频对应的第二重建背景帧,并基于所述第二重建背景帧,对所述第二图像码流进行解码。The viewpoint-by-view determination unit is further configured to, according to the second video coding, determine a second reconstructed background frame corresponding to the second target video from a plurality of the reconstructed background frames, and based on the second reconstructed background frame For background frames, the second image code stream is decoded.
  12. 如权利要求8所述的视频解码系统,其中,所述图像码流解码模块还包括:The video decoding system of claim 8, wherein the image code stream decoding module further comprises:
    运动矢量设置单元,用于在对每一所述图像码流进行解码时,确定与所述图像码流对应的图像块使用帧间预测模式,将所述图像块中的运动矢量设置为0,其中,所述图像码流是由多个所述图像块的码流组成;a motion vector setting unit, configured to determine that an image block corresponding to the image code stream uses an inter-frame prediction mode when decoding each of the image code streams, and set the motion vector in the image block to 0, Wherein, the image code stream is composed of a plurality of code streams of the image blocks;
    残差跳过解码单元,用于将使用帧间预测模式的图像块对应的重建背景帧作为参考帧,并按照跳过所述使用帧间预测模式的图像块对应的运动矢量残差信息的方式,对所述图像码流进行解码。A residual skip decoding unit, configured to use the reconstructed background frame corresponding to the image block using the inter-frame prediction mode as a reference frame, and skip the motion vector residual information corresponding to the image block using the inter-frame prediction mode. , and decode the image code stream.
  13. 如权利要求8所述的视频解码系统,其中,所述背景帧码流为根据所述目标视频中的视频图像经过背景建模获取的背景帧进行编码所得。The video decoding system according to claim 8, wherein the background frame code stream is obtained by encoding a background frame obtained by background modeling according to the video image in the target video.
  14. 如权利要求13所述的视频解码系统,其中,所述对所述视频图像进行背景建模的方式包括单帧生成方式、中值滤波方式与均值滤波方式。The video decoding system according to claim 13, wherein the method for performing background modeling on the video image includes a single frame generation method, a median filtering method and an average filtering method.
  15. 一种视频解码设备,其中,所述视频解码设备包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机可读指令,所述计算机可读指令被所述处理器执行时实现如下步骤:A video decoding apparatus, wherein the video decoding apparatus includes a memory, a processor, and computer-readable instructions stored on the memory and executable on the processor, the computer-readable instructions being processed by the processor The following steps are implemented when the device is executed:
    接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
    在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。When a playback viewpoint selection instruction is received, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, multiple image streams corresponding to different viewpoints of the target video are received, and based on a plurality of the reconstructed background frames, the different The multi-channel image code stream of the viewpoint is decoded to obtain the multi-channel target video.
  16. 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有计算机可读指令,所述计算机可读指令被处理器执行时实现如下步骤:A computer-readable storage medium, wherein computer-readable instructions are stored on the computer-readable storage medium, and when the computer-readable instructions are executed by a processor, the following steps are implemented:
    接收编码端发送的不同视点对应的多个背景帧码流,并将多个所述背景帧码流进行解码得到多个重建背景帧;receiving multiple background frame code streams corresponding to different viewpoints sent by the encoding end, and decoding the multiple background frame code streams to obtain multiple reconstructed background frames;
    在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频。When a playback viewpoint selection instruction is received, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, multiple image streams corresponding to different viewpoints of the target video are received, and based on a plurality of the reconstructed background frames, the different The multi-channel image code stream of the viewpoint is decoded to obtain the multi-channel target video.
  17. 如权利要求16所述的计算机可读存储介质,其中,所述在接收到播放视点选择指令时,按照所述播放视点选择指令所确定的视点选择顺序,接收目标视频对应在不同视点的多路图像码流,并基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码,以得到多路目标视频的步骤包括:The computer-readable storage medium according to claim 16, wherein when receiving a playback viewpoint selection instruction, according to the viewpoint selection sequence determined by the playback viewpoint selection instruction, the received target video corresponds to multiple channels of different viewpoints. The steps of decoding the multi-channel image code streams of the different viewpoints based on a plurality of the reconstructed background frames to obtain the multi-channel target video include:
    在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流;When receiving the playback viewpoint selection instruction sent by the user, determine the target playback viewpoints currently selected by the user one by one based on the playback viewpoint selection instruction, and acquire an image code stream corresponding to the target playback viewpoint generated by the encoder ;
    从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码,以得到并播放与所述目标播放视点对应的一路目标视频。A target reconstructed background frame corresponding to the target playback viewpoint is selected from a plurality of the reconstructed background frames, and based on the target reconstructed background frame, one image stream corresponding to the target playback viewpoint is decoded, to obtain and play a target video corresponding to the target playback viewpoint.
  18. 如权利要求17所述的计算机可读存储介质,其中,所述播放视点选择指令包括首次视点选择指令,The computer-readable storage medium of claim 17, wherein the playback viewpoint selection instruction comprises a first viewpoint selection instruction,
    所述在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流的步骤包括:When receiving the playback viewpoint selection instruction sent by the user, the target playback viewpoint currently selected by the user is determined one by one based on the playback viewpoint selection instruction, and an image generated by the encoding end corresponding to the target playback viewpoint is obtained. The steps of the code stream include:
    在接收到由用户发送的首次视点选择指令时,获取所述首次视点选择指令中的首次所选视点,并确定与所述首次所选视点对应的第一目标视频的第一视频编号;When receiving the first viewpoint selection instruction sent by the user, acquiring the first selected viewpoint in the first viewpoint selection instruction, and determining the first video number of the first target video corresponding to the first selected viewpoint;
    确定所述第一目标视频中位于初始位置的视频图像的第一图像编号;determining the first image number of the video image at the initial position in the first target video;
    从编码端接收与所述第一视频编号以及所述第一图像编号对应的第一图像码流;Receive a first image code stream corresponding to the first video number and the first image number from the encoding end;
    所述从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码的步骤包括:The target reconstructed background frame corresponding to the target playback viewpoint is selected from the plurality of reconstructed background frames, and the image code stream corresponding to the target playback viewpoint is performed based on the target reconstructed background frame. The steps of decoding include:
    根据所述第一视频编码,从多个所述重建背景帧中确定与所述第一目标视频对应的第一重建背景帧,并基于所述第一重建背景帧,对所述第一图像码流进行解码。According to the first video coding, a first reconstructed background frame corresponding to the first target video is determined from a plurality of the reconstructed background frames, and based on the first reconstructed background frame, the first image code is stream to decode.
  19. 如权利要求18所述的计算机可读存储介质,其中,所述播放视点选择指令包括视点切换指令,The computer-readable storage medium of claim 18, wherein the playback viewpoint selection instruction comprises a viewpoint switching instruction,
    所述在接收到由用户发送的播放视点选择指令时,基于所述播放视点选择指令逐一确定用户当前所选的目标播放视点,并获取由编码端生成的与所述目标播放视点对应的一路图像码流的步骤包括:When receiving the playback viewpoint selection instruction sent by the user, the target playback viewpoint currently selected by the user is determined one by one based on the playback viewpoint selection instruction, and an image generated by the encoding end corresponding to the target playback viewpoint is obtained. The steps of the code stream include:
    在接收到用户发送的视点切换指令时,获取所述视点切换指令中的目标切换视点,并确定与所述目标切换视点对应的第二目标视频的第二视频编号;When receiving the viewpoint switching instruction sent by the user, acquiring the target switching viewpoint in the viewpoint switching instruction, and determining the second video number of the second target video corresponding to the target switching viewpoint;
    获取所述第一目标视频的播放进度,并确定所述第二目标视频中与所述播放进度对应的第二视频图像的第二图像编号;Acquire the playback progress of the first target video, and determine the second image number of the second video image corresponding to the playback progress in the second target video;
    从编码端接收与所述第二视频编号以及所述第二图像编号对应的第二图像码流;Receive a second image code stream corresponding to the second video number and the second image number from the encoding end;
    所述从多个所述重建背景帧中选出与所述目标播放视点对应的目标重建背景帧,并基于所述目标重建背景帧对所述与所述目标播放视点对应的一路图像码流进行解码的步骤包括:The target reconstructed background frame corresponding to the target playback viewpoint is selected from the plurality of reconstructed background frames, and the image code stream corresponding to the target playback viewpoint is performed based on the target reconstructed background frame. The steps of decoding include:
    根据所述第二视频编码,从多个所述重建背景帧中确定与所述第二目标视频对应的第二重建背景帧,并基于所述第二重建背景帧,对所述第二图像码流进行解码。According to the second video encoding, a second reconstructed background frame corresponding to the second target video is determined from a plurality of the reconstructed background frames, and based on the second reconstructed background frame, the second image code is stream to decode.
  20. 如权利要求16所述的计算机可读存储介质,其中,所述基于多个所述重建背景帧将所述不同视点的多路图像码流进行解码的步骤包括:The computer-readable storage medium of claim 16, wherein the step of decoding the multiplexed image code streams of different viewpoints based on a plurality of the reconstructed background frames comprises:
    在对每一所述图像码流进行解码时,确定与所述图像码流对应的图像块使用帧间预测模式,将所述图像块中的运动矢量设置为0,其中,所述图像码流是由多个所述图像块的码流组成;When decoding each of the image code streams, it is determined that the image block corresponding to the image code stream uses the inter prediction mode, and the motion vector in the image block is set to 0, wherein the image code stream is composed of a plurality of code streams of the image blocks;
    将使用帧间预测模式的图像块对应的重建背景帧作为参考帧,并按照跳过所述使用帧间预测模式的图像块对应的运动矢量残差信息的方式,对所述图像码流进行解码。Taking the reconstructed background frame corresponding to the image block using the inter-frame prediction mode as a reference frame, and decoding the image code stream in a manner of skipping the motion vector residual information corresponding to the image block using the inter-frame prediction mode .
PCT/CN2020/111448 2020-07-29 2020-08-26 Video decoding method, system and device and computer-readable storage medium WO2022021519A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010748734.3A CN111800653B (en) 2020-07-29 2020-07-29 Video decoding method, system, device and computer readable storage medium
CN202010748734.3 2020-07-29

Publications (1)

Publication Number Publication Date
WO2022021519A1 true WO2022021519A1 (en) 2022-02-03

Family

ID=72827443

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/111448 WO2022021519A1 (en) 2020-07-29 2020-08-26 Video decoding method, system and device and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN111800653B (en)
WO (1) WO2022021519A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115150639A (en) * 2022-09-01 2022-10-04 北京蔚领时代科技有限公司 Weak network resisting method and device based on distributed encoder

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114449348A (en) * 2020-11-04 2022-05-06 北京金山云网络技术有限公司 Panoramic video processing method and device
CN113114985B (en) * 2021-03-31 2022-07-26 联想(北京)有限公司 Information processing method and information processing device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101742319A (en) * 2010-01-15 2010-06-16 北京大学 Background modeling-based static camera video compression method and background modeling-based static camera video compression system
CN102006473A (en) * 2010-11-18 2011-04-06 无锡中星微电子有限公司 Video encoder and encoding method, and video decoder and decoding method
CN102413332A (en) * 2011-12-01 2012-04-11 武汉大学 Multi-viewpoint video coding method based on time-domain-enhanced viewpoint synthesis prediction
CN104703027A (en) * 2015-03-17 2015-06-10 华为技术有限公司 Decoding method and decoding device for video frame
JP2017092886A (en) * 2015-11-17 2017-05-25 日本電信電話株式会社 Video encoding method, video encoder and video encoding program
CN107396138A (en) * 2016-05-17 2017-11-24 华为技术有限公司 A kind of video coding-decoding method and equipment
EP1624675B1 (en) * 2004-08-03 2020-04-22 Microsoft Technology Licensing, LLC A system and process for compressing and decompressing multiple, layered, video streams employing spatial and temporal encoding
CN111447503A (en) * 2020-04-26 2020-07-24 烽火通信科技股份有限公司 Viewpoint switching method, server and system for multi-viewpoint video

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102006475B (en) * 2010-11-18 2012-12-19 无锡中星微电子有限公司 Video coding and decoding device and method
CN102665077A (en) * 2012-05-03 2012-09-12 北京大学 Rapid and efficient encoding-transcoding method based on macro block classification
US9693055B2 (en) * 2012-12-28 2017-06-27 Electronics And Telecommunications Research Institute Video encoding and decoding method and apparatus using the same
CN103152570B (en) * 2013-03-01 2016-02-24 北京大学 A kind of video bit stream coding/decoding method and device
CN104980763B (en) * 2014-04-05 2020-01-17 浙江大学 Video code stream, video coding and decoding method and device
JP6946724B2 (en) * 2017-05-09 2021-10-06 ソニーグループ株式会社 Client device, client device processing method, server and server processing method
CN109495749A (en) * 2018-12-24 2019-03-19 上海国茂数字技术有限公司 A kind of coding and decoding video, search method and device
CN111372145B (en) * 2020-04-15 2021-07-27 烽火通信科技股份有限公司 Viewpoint switching method and system for multi-viewpoint video

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1624675B1 (en) * 2004-08-03 2020-04-22 Microsoft Technology Licensing, LLC A system and process for compressing and decompressing multiple, layered, video streams employing spatial and temporal encoding
CN101742319A (en) * 2010-01-15 2010-06-16 北京大学 Background modeling-based static camera video compression method and background modeling-based static camera video compression system
CN102006473A (en) * 2010-11-18 2011-04-06 无锡中星微电子有限公司 Video encoder and encoding method, and video decoder and decoding method
CN102413332A (en) * 2011-12-01 2012-04-11 武汉大学 Multi-viewpoint video coding method based on time-domain-enhanced viewpoint synthesis prediction
CN104703027A (en) * 2015-03-17 2015-06-10 华为技术有限公司 Decoding method and decoding device for video frame
JP2017092886A (en) * 2015-11-17 2017-05-25 日本電信電話株式会社 Video encoding method, video encoder and video encoding program
CN107396138A (en) * 2016-05-17 2017-11-24 华为技术有限公司 A kind of video coding-decoding method and equipment
CN111447503A (en) * 2020-04-26 2020-07-24 烽火通信科技股份有限公司 Viewpoint switching method, server and system for multi-viewpoint video

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115150639A (en) * 2022-09-01 2022-10-04 北京蔚领时代科技有限公司 Weak network resisting method and device based on distributed encoder
CN115150639B (en) * 2022-09-01 2022-12-20 北京蔚领时代科技有限公司 Weak network resisting method and device based on distributed encoder

Also Published As

Publication number Publication date
CN111800653B (en) 2021-06-11
CN111800653A (en) 2020-10-20

Similar Documents

Publication Publication Date Title
EP2134092B1 (en) Information processing apparatus and method, and program
US6989868B2 (en) Method of converting format of encoded video data and apparatus therefor
KR101377021B1 (en) Encoding device and method, decoding device and method, and transmission system
WO2022021519A1 (en) Video decoding method, system and device and computer-readable storage medium
US20070103558A1 (en) Multi-view video delivery
US20130022116A1 (en) Camera tap transcoder architecture with feed forward encode data
CN101690163A (en) Shutter time compensation
CN101002471A (en) Method and apparatus to encode image, and method and apparatus to decode image data
KR19990064087A (en) Video speech decoding apparatus, video speech coding apparatus and information transmission system
JP2010035133A (en) Moving image encoding apparatus and moving image encoding method
JP2020524450A (en) Transmission system for multi-channel video, control method thereof, multi-channel video reproduction method and device thereof
JP6541932B2 (en) Video system and method for displaying image data, computer program and encoding apparatus
JPH10262228A (en) Communication system, multi-point controller and video information display method
JP4767916B2 (en) Video encoded data converter
CN117579843B (en) Video coding processing method and electronic equipment
CN114513658B (en) Video loading method, device, equipment and medium
CN117291810B (en) Video frame processing method, device, equipment and storage medium
RU2628198C1 (en) Method for interchannel prediction and interchannel reconstruction for multichannel video made by devices with different vision angles
CN104702970A (en) Video data synchronization method, device and system
JP3205839B2 (en) Image coding device
JPH0294886A (en) Method and device for coding and decoding video signal
KR101609798B1 (en) moving picture replay device
JP2005159832A (en) Signal processor
JP2001045495A (en) Image compositing device
JP2000115779A (en) Image processing unit and moving image coding method applied to the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20946746

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20946746

Country of ref document: EP

Kind code of ref document: A1