WO2020135527A1 - 多媒体数据处理 - Google Patents
多媒体数据处理 Download PDFInfo
- Publication number
- WO2020135527A1 WO2020135527A1 PCT/CN2019/128377 CN2019128377W WO2020135527A1 WO 2020135527 A1 WO2020135527 A1 WO 2020135527A1 CN 2019128377 W CN2019128377 W CN 2019128377W WO 2020135527 A1 WO2020135527 A1 WO 2020135527A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- multimedia data
- track
- frame
- encoded
- data stream
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims description 59
- 238000000034 method Methods 0.000 claims abstract description 54
- 238000003672 processing method Methods 0.000 claims abstract description 23
- 230000008569 process Effects 0.000 claims abstract description 17
- 238000005538 encapsulation Methods 0.000 claims description 70
- 230000005540 biological transmission Effects 0.000 claims description 44
- 238000013480 data collection Methods 0.000 claims description 44
- 239000002131 composite material Substances 0.000 claims description 41
- 238000004891 communication Methods 0.000 claims description 28
- 238000004806 packaging method and process Methods 0.000 claims description 10
- 238000010586 diagram Methods 0.000 description 22
- 101001073193 Homo sapiens Pescadillo homolog Proteins 0.000 description 6
- 102100035816 Pescadillo homolog Human genes 0.000 description 6
- 238000012544 monitoring process Methods 0.000 description 4
- 238000013500 data storage Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 101000857634 Homo sapiens Receptor-transporting protein 1 Proteins 0.000 description 2
- 102100025426 Receptor-transporting protein 1 Human genes 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/647—Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/647—Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
- H04N21/64723—Monitoring of network processes or resources, e.g. monitoring of network load
- H04N21/64738—Monitoring network characteristics, e.g. bandwidth, congestion level
Definitions
- This application relates to video surveillance technology, and in particular to multimedia data processing.
- the present application provides a multimedia data processing method, device, electronic device, and readable storage medium.
- a multimedia data processing method is provided, which is applied to a sending-end device.
- the method includes:
- Transmitting the multimedia data stream to the receiving end device decapsulates the multimedia data stream to obtain encoded multimedia data of each track, and encodes the encoded multimedia data of each track based on the synchronization information Be processed.
- a multimedia data processing method is provided, which is applied to a receiving end device, and is characterized in that the method includes:
- the multimedia data stream includes encoded multimedia data of multiple tracks, and synchronization information for synchronizing the encoded multimedia data of each track, wherein the encoded multimedia data of each track is for The data obtained by encoding the multimedia data of each track separately;
- the encoded multimedia data of each track is processed based on the synchronization information.
- a multimedia data processing apparatus which is applied to a sending-end device, and the apparatus includes:
- An obtaining unit used to obtain the encoded multimedia data of each track
- An encapsulation unit for encapsulating the encoded multimedia data of the multiple tracks into a single multimedia data stream
- the multimedia data stream includes the encoded multimedia data of each track, and synchronization information for synchronizing the encoded multimedia data of each track;
- the transmission unit is used to transmit the multimedia data stream to the receiving end device, and the receiving end device decapsulates the multimedia data stream to obtain encoded multimedia data for each track, and based on the synchronization information for each Track encoded multimedia data for processing.
- a multimedia data processing apparatus which is applied to a receiving end device, and the apparatus includes:
- a receiving unit configured to receive a multimedia data stream sent by a sending end device;
- the multimedia data stream includes encoded multimedia data of multiple tracks, and synchronization information for synchronizing the encoded multimedia data of each track, wherein the encoding of each track
- Multimedia data is data obtained by separately encoding multimedia data for each track;
- a decapsulation unit used to decapsulate the multimedia data stream to obtain encoded multimedia data of each track
- the processing unit is configured to process the encoded multimedia data of each track based on the synchronization information.
- an electronic device including a processor, a communication interface, a storage medium, and a communication bus, wherein the processor, the communication interface, and the storage medium complete communication with each other through the communication bus;
- Storage medium for storing machine executable instructions
- the processor is configured to implement the multimedia data processing method provided in the first aspect when executing the machine-executable instructions stored on the storage medium.
- a machine-readable storage medium in which machine-executable instructions are stored, and the machine-executable instructions are implemented by a processor to implement the first aspect. Multimedia data processing method.
- an electronic device including a processor, a communication interface, a storage medium, and a communication bus, wherein the processor, the communication interface, and the storage medium complete communication with each other through the communication bus;
- Storage medium for storing machine executable instructions
- the processor is configured to implement the multimedia data processing method provided in the second aspect when executing the machine-executable instructions stored on the storage medium.
- a machine-readable storage medium in which machine-executable instructions are stored, and the machine-executable instructions are implemented by a processor to implement the second aspect. Multimedia data processing method.
- the encoded multimedia data of multiple tracks is encapsulated into a single multimedia data stream by the transmitting device, and the multimedia data stream is transmitted to the receiving device, and the multimedia data stream is deciphered by the receiving device Encapsulation, to obtain the encoded multimedia data of each track, and process the encoded multimedia data of each track, while ensuring the synchronization of the multimedia data of multiple tracks, reduce the performance requirements of the encoding end device and the decoding end device, and expand The applicable scenario of the program.
- FIG. 1 is a schematic structural diagram of a multimedia data processing system according to an exemplary embodiment of the present application.
- FIG. 2 is a schematic flowchart of a multimedia data processing method according to an exemplary embodiment of the present application.
- FIG. 3 is a schematic flowchart of a multimedia data processing method according to an exemplary embodiment of the present application.
- 4A is a schematic diagram of a common RTP package.
- FIG. 4B is a schematic diagram of encapsulation of a multimedia data stream, such as an RTP stream, according to an exemplary embodiment of the present application.
- FIG. 4C is a schematic flowchart of an NVR processing a received RTP stream according to an exemplary embodiment of the present application.
- Figure 4D is a schematic diagram of an ordinary PES package.
- FIG. 4E is a schematic diagram of encapsulation of a multimedia data stream, such as a PES stream, according to an exemplary embodiment of the present application.
- Fig. 4F is a schematic diagram showing a NVR stitching and displaying video images according to an exemplary embodiment of the present application.
- FIG. 5 is a schematic structural diagram of a multimedia data processing device according to an exemplary embodiment of the present application.
- FIG. 6 is a schematic structural diagram of a multimedia data processing apparatus according to another exemplary embodiment of the present application.
- FIG. 7 is a schematic diagram of a hardware structure of an electronic device shown in an exemplary embodiment of the present application.
- FIG. 8 is a schematic structural diagram of a multimedia data processing apparatus according to an exemplary embodiment of the present application.
- FIG. 9 is a schematic structural diagram of a multimedia data processing apparatus according to another exemplary embodiment of the present application.
- Fig. 10 is a schematic structural diagram of a multimedia data processing apparatus according to another exemplary embodiment of the present application.
- FIG. 11 is a schematic diagram of a hardware structure of an electronic device shown in an exemplary embodiment of the present application.
- FIG. 1 is a schematic structural diagram of a multimedia data processing system according to an embodiment of the present application.
- the multimedia data processing system may include a sending device 101 and a receiving device 102.
- the sending device 101 can encapsulate multiple tracks of multimedia data into a single multimedia data stream, and send the multimedia data stream to the receiving device 102.
- multiple tracks may be understood as “multi-channel”, “multi-channel”, etc., and are used to illustrate that the multimedia data received by the sending-end device 101 may come from different data sources.
- the sending end device 101 may acquire multimedia data of multiple tracks from a single target multimedia data collection device.
- the target multimedia data collection device can be integrated with multiple sensors, and the multiple sensors can shoot the same scene, and each sensor can shoot to obtain a track of multimedia data.
- the sending-end device 101 may acquire multimedia data of multiple tracks from multiple target multimedia data collection devices.
- the multiple target multimedia data collection devices are not fixed, but can be specified according to the requirements in the monitoring scene to cooperate with video monitoring to achieve a larger monitoring range.
- the multimedia data collected by the multiple target multimedia data collection devices needs to be stitched before being played.
- the sending-end device 101 may acquire super-resolution multimedia data stored locally or collected in real time, and divide the acquired super-resolution multimedia data according to a preset number of tracks to obtain multimedia data of multiple tracks.
- the sending-end device 101 may also receive a multimedia data stream encapsulated according to a preset transmission encapsulation protocol, and decapsulate the received multimedia data stream to obtain multimedia data of multiple tracks.
- the sending-end device 101 may also decapsulate the locally stored multimedia data stream encapsulated according to a preset storage encapsulation protocol to obtain multimedia data of multiple tracks.
- the multimedia data acquired by the sending end device can also be processed according to specific requirements.
- the multimedia data collection device has multiple sensors integrated to shoot the same scene. After acquiring multiple images of the same scene captured by multiple sensors from different angles, it may be necessary to extract images containing specific objects (eg, images containing "cars") from the captured multiple images for transmission . In this case, each image can be cut separately to obtain an effective image containing a specific object "car”. At this time, the image data of each image obtained by cutting is multimedia data of multiple tracks.
- the multimedia data stream may include multimedia data of each track.
- the multimedia data of each track may be multimedia data collected at the same time or multimedia data collected at different times.
- the present disclosure does not Limit it.
- the same time is not limited to the exact same time, and a preset error (which can be determined according to the actual scene, such as 2 milliseconds, 5 milliseconds, etc.) may be allowed.
- multimedia data 1 and multimedia Data 2 is determined to be multimedia data collected by the multimedia data collection device A and the multimedia data collection device B at the same time.
- the encapsulated multimedia data stream also carries synchronization information for multimedia data synchronization of each track.
- the multimedia data of each track can be synchronized with each other to facilitate data recovery at the decoding end.
- the synchronization information will be described in detail below.
- the multimedia data of each track are respectively encoded multimedia data.
- the encoding operation of the multimedia data can be performed by the sending end device. For example, after acquiring the multimedia data of each track, the transmitting-end device 101 encodes the multimedia data of each track, and then encapsulates the encoded multimedia data of each track into a single multimedia data stream.
- the encoding operation may also be performed by the multimedia data collection device.
- a multimedia data collection device corresponding to a track encodes the collected data after collecting the multimedia data, and the multimedia data obtained by the sending end device 101 from the multimedia data collection device is the encoded multimedia data of the track.
- the receiving end device 102 when the receiving end device 102 receives the multimedia data stream sent by the sending end device 101, it may decapsulate the multimedia data stream to obtain the encoded multimedia data of each track, and The encoded multimedia data is processed.
- the sender device and the receiver device only refer to the role of the device in different application scenarios, and do not limit the device type. That is, the receiver device in a certain application scenario is used in other applications. It can also be used as the sending device in the scenario.
- the sending device is a multimedia data collection device
- a device with multimedia data storage function such as NVR (Network Video Recorder, network hard disk recorder)
- NVR Network Video Recorder, network hard disk recorder
- the device with the multimedia data storage function can be used as the sending end device to send the multimedia data to the terminal device (as the receiving end device).
- the multimedia data collection device in actual scenarios, it may be a plurality of target multimedia data collection devices that are independently deployed, or an integrated deployment of multimedia data collection devices, such as a dome camera with multiple cameras integrated.
- the sending-end device is, for example, a dome camera
- the multimedia data of multiple tracks acquired by it is image data captured by multiple cameras of itself.
- the sending end can separately encode the multimedia data of multiple tracks, and encapsulate the encoded multimedia data of each track into a single multimedia data stream to send to the decoding end; the decoding end receives The obtained multimedia data stream is decapsulated and then decoded to obtain decoded images of each track, and the obtained decoded images of each track can be spliced.
- this scheme reduces the performance requirements of the encoding end device and the decoding end device.
- FIG. 2 is a schematic flowchart of a multimedia data processing method provided by an embodiment of the present application, where the multimedia data processing method may be applied to a sending end device, as shown in FIG. 2, the multimedia data processing method may include the following step.
- Step S200 Obtain the encoded multimedia data of each track.
- Step S210 Encapsulate the encoded multimedia data of the multiple tracks into a single multimedia data stream, where the multimedia data stream includes the encoded multimedia data of each track and synchronization information for synchronizing the encoded multimedia data of each track.
- Step S220 Transmit the multimedia data stream to the receiving end device, and the receiving end device decapsulates the multimedia data stream to obtain encoded multimedia data of each track, and processes the encoded multimedia data of each track based on synchronization information.
- the multimedia data may include but is not limited to video data, audio data, or picture data.
- the sending device when the sending device receives the instruction to send the multimedia data collected by the target multimedia data collection device to the receiving device, the sending device can encapsulate the multimedia data collected by multiple target multimedia data collection devices into a single Multimedia data stream, and transmit the multimedia data stream to the receiving device.
- the acquiring respective encoded multimedia data of multiple tracks includes: acquiring multimedia data of multiple tracks from a single data collection device, wherein the multimedia data of the multiple tracks is in the data collection device Multi-sensors are set to shoot the multimedia data for the same scene; separately encode the multimedia data of each track to obtain the encoded multimedia data.
- the acquiring the encoded multimedia data of each track includes: acquiring the encoded multimedia data of multiple tracks from multiple data collection devices.
- the acquiring respective encoded multimedia data of multiple tracks includes: dividing the multimedia data according to the number of the tracks to obtain multimedia data of the multiple tracks; and separately analyzing the multimedia data of each track Encoding to obtain the encoded multimedia data.
- the acquiring the encoded multimedia data of each track includes: de-encapsulating the pre-stored multimedia data according to a preset storage and encapsulation protocol to obtain the encoded multimedia data of each track.
- the multimedia data stored locally at the sending end may be data encapsulated according to a preset storage encapsulation protocol.
- the encapsulation operation using the preset storage encapsulation protocol will be described below.
- the encapsulating the encoded multimedia data of multiple tracks into a single multimedia data stream may include: encapsulating the encoded multimedia data of each track into a multimedia data stream according to a preset transmission encapsulation protocol.
- the preset transmission encapsulation protocol may include but not limited to RTP (Real-time Transport Protocol) protocol and the like.
- the sending-end device when the sending-end device is a multimedia data collection device, for example, if the sending-end device is a dome camera integrating multiple cameras, the sending-end device can separately integrate multiple target multimedia data collection devices (such as the integrated dome camera) Multiple cameras) to encode the multimedia data collected, and encapsulate the encoded multimedia data of the multiple target multimedia data collection devices into a multimedia data stream according to a preset transmission encapsulation protocol.
- target multimedia data collection devices such as the integrated dome camera
- Multiple cameras to encode the multimedia data collected, and encapsulate the encoded multimedia data of the multiple target multimedia data collection devices into a multimedia data stream according to a preset transmission encapsulation protocol.
- the foregoing encapsulating multimedia data collected by multiple target multimedia data collection devices into a multimedia data stream may include:
- the encoded multimedia data of each track is encapsulated into a multimedia data stream according to a preset transmission encapsulation protocol.
- the sending end device when the sending end device is a multimedia data storage device, such as an NVR, the sending end device may first decapsulate the stored multimedia data of the multiple target multimedia data collection devices encapsulated according to a preset storage encapsulation protocol To obtain the encoded multimedia data of multiple target multimedia data collection devices, and encapsulate the encoded multimedia data of the multiple target multimedia data collection devices into a multimedia data stream according to a preset transmission encapsulation protocol.
- the preset storage encapsulation protocol may include, but is not limited to, PS (MPEG (Moving Picture Experts Group) Program Stream, a mainstream audio and video data encapsulation method) protocol, etc.
- PS MPEG (Moving Picture Experts Group) Program Stream
- a mainstream audio and video data encapsulation method etc.
- the synchronization information carried in the multimedia data stream is synthetic frame information
- the synthetic frame information may include but is not limited to: packet type identification, frame start identification, frame end identification, frame type identification, composite frame sequence number, The total number of tracks in the composite frame.
- the data packet type identifier is used to identify the type of the data packet carried in the multimedia data stream, which includes but is not limited to video data, audio data, or picture data;
- the start of frame identifier is used to identify whether the data packet carried in the multimedia data stream is the start of frame data packet
- the end-of-frame identification is used to identify whether the data packet carried in the multimedia data stream is the end-of-frame data packet
- the frame type identifier is used to identify the frame type corresponding to the data packet carried in the multimedia data stream.
- video frames it may include, but is not limited to, I frames, P frames, or B frames;
- the synthesized frame sequence number is used to identify the track where the data packet carried in the multimedia data stream is located (one track for each target multimedia data collection device);
- the total number of tracks of the synthesized frame is used to identify the data carried by several tracks in the multimedia data stream (that is, the multimedia data of several target multimedia data collection devices).
- the synthesized frame information may be carried in the header of each data packet.
- the synthesized frame information can be carried in the RTP header of each RTP packet, and the receiving end device can determine that it belongs to the same multimedia data stream according to the synthesized frame information carried in the RTP header of the RTP packet.
- the above limitation on the content of the synthesized frame information is only a specific example in the present application, and does not belong to the limitation on the protection scope of the present application.
- the synthesized frame information may not include the frame start identifier and the frame end identifier; or, the multimedia data stream
- the data packets corresponding to the images collected by the multiple cameras at the same time are uniformly numbered, and the multimedia data collection device corresponding to each data packet can be directly determined according to the packet sequence number, without the need to additionally carry the synthetic frame sequence number.
- packets with sequence numbers 0 to 3 correspond to the packets of camera 1; packets with sequence numbers 4 to 7 correspond to the image of camera 2.
- Packets 8 to 11 are the data packets corresponding to the image of the camera 3, and the specific implementation thereof will not be repeated here.
- FIG. 3 is a schematic flowchart of a multimedia data processing method according to an embodiment of the present application.
- the multimedia data processing method may be applied to a receiving device.
- the multimedia data processing method may include the following step.
- Step S300 Receive a multimedia data stream sent by the sending end device; wherein the multimedia data stream includes encoded multimedia data of multiple tracks and synchronization information for synchronizing the encoded multimedia data of each track, wherein the encoded multimedia data of each track Data obtained by separately encoding multimedia data for each track.
- Step S310 Decapsulate the multimedia data stream to obtain encoded multimedia data of each track.
- Step S320 Process the encoded multimedia data of each track based on the synchronization information.
- the receiving end device when the receiving end device receives the multimedia data stream sent by the sending end device, it can decapsulate the multimedia data stream to obtain the encoded multimedia data of multiple tracks, and based on the synchronization information to each track The encoded multimedia data is processed.
- processing of the encoded multimedia data by the receiving end device may include, but is not limited to, segmentation, display, transmission, and/or storage.
- the multimedia data is video data.
- the above processing of the encoded multimedia data of each track based on the synchronization information may include:
- the receiving end device may first decode the encoded video frame of each track.
- the synchronization information may include composite frame information carried in a data packet of a multimedia data stream
- the composite frame information may include at least one of the following: a packet type identifier, a frame start identifier and a frame end identifier, a frame type identifier, a composite frame sequence number, and a total frame track number.
- the decoding of the encoded multimedia data of each track based on the synchronization information includes:
- the receiving end device may determine whether the data in the multimedia data stream is video data according to the packet type identifier in the multimedia data stream (such as carried in the encapsulation header of the multimedia data stream); for video data, the receiving end The device may determine the data packet corresponding to the video frame of the same track according to the composite frame sequence number, and obtain the complete video frame of the same track according to the frame start identifier and the frame end identifier.
- the receiving end device When the receiving end device obtains the complete video frame of each track, it can decode the complete video frame of each track separately.
- multiple parallel decoders may be used to decode the encoded multimedia data of multiple tracks.
- the correspondence between tracks and decoders can be configured, one track corresponds to one decoder, and the receiving end device can send multimedia data to the corresponding decoder for decoding according to the track where the multimedia data is located.
- the receiving end device decodes the video of each track to obtain a decoded video image; cuts the decoded video image according to preset conditions to obtain a cut video image;
- the video images are individually encoded to obtain the encoded video data of each track;
- the encoded video data of each track is encapsulated into a single multimedia video stream according to a preset storage encapsulation protocol for storage, or encapsulated according to a preset transmission encapsulation protocol for network transmission.
- the receiving device can analyze whether the video images meet the requirements. For example, whether each image contains a specific object (for example, a car). It is possible to encapsulate only images containing specific objects. Alternatively, the image can be cut to retain only the part containing the specific object, and then the cut image is encoded, and then the encoded data is encapsulated according to a preset packaging protocol. The encapsulated data stream can be stored locally or transmitted over the network.
- a specific object for example, a car
- the receiving end device after the receiving end device decodes the video images of each track, it can determine the position information of each video image in the mosaic according to the track where the decoded video image is located and the width and height of the decoded video image , And stitch and display the decoded video image according to the location information.
- the receiving end device after the receiving end device stitches the video images of each track, it can also zoom the stitched video image according to the set output resolution, and perform the zooming on the stitched image. display.
- the decoded video image after the receiving end device obtains the decoded video image, before stitching the decoded video image, the decoded video image may be scaled according to actual requirements, and the scaled video The images are spliced and displayed, and the specific implementation thereof will not be repeated here.
- the above-mentioned decoding of the complete video frames of each track through multiple parallel decoders is only a specific example of decoding the video frames of the multimedia data collection device, and does not belong to the present application.
- Limitation of protection scope that is, the complete video frame can also be decoded in other ways, for example, the complete video frame of each track can be serially decoded by a single decoder; or multiple decoders can be used for parallel decoding, one decoding
- the device can correspond to multiple target multimedia data collection devices, the specific implementation of which will not be repeated here.
- the above synchronization information includes the synthesized frame information carried in the data packet of the multimedia data stream; the above processing of the encoded multimedia data of each track based on the synchronization information may include:
- the receiver device after the receiver device receives the multimedia data stream sent by the sender device and decapsulates the encoded multimedia data of each track, it can send the encoded multimedia data of each track to other devices (in this case The receiving device also serves as the sending device).
- the receiving end device may encapsulate the encoded multimedia data of each track into a single multimedia data stream according to a preset transmission encapsulation protocol, and perform network transmission on the multimedia data stream encapsulated according to the preset transmission encapsulation protocol.
- the synchronization information includes the synthesized frame information carried in the data packet of the multimedia data stream; the processing of the encoded multimedia data of each track based on the synchronization information includes:
- the multimedia data stream encapsulated according to the preset storage encapsulation protocol is stored.
- the encoded multimedia data of each track may be stored.
- the receiving end device may encapsulate the encoded multimedia data of each track into a multimedia data stream according to a preset storage encapsulation protocol, and store the multimedia data stream encapsulated according to the preset storage encapsulation protocol.
- the sending device is a dome camera with multiple cameras
- the receiving device is NVR
- the preset transmission encapsulation protocol is RTP
- the preset storage encapsulation protocol is PS.
- the video data collected by multiple cameras can be packaged into a multimedia data stream (hereinafter referred to as multimedia data RTP stream) according to the RTP protocol.
- multimedia data RTP stream a multimedia data stream
- ordinary RTP encapsulation is performed in a 4-byte alignment mode.
- the RTP packet needs to be filled; where the length of the filled bytes can be 1, 2 Or 3 bytes.
- the multimedia data RTP stream uses the padding field in the common RTP header, and at least 4 bytes are filled.
- the schematic diagram can be as shown in FIG. 4B.
- These bytes include MFI (Multiple Frame Indicator). Information) byte, the MFI byte includes the packet type (audio or video, etc.) identification, frame start identification, frame end identification, frame type (I frame, P frame%) identification, synthetic frame sequence number (use this to confirm the packet The track where it is) and the total number of tracks in the synthesized frame.
- the multimedia data stream of n tracks is RTP0 header+RTP0 payload(load)+MFI0+RTP1 header+RTP1 payload+MFI1+RTPn header+...+RTPn payload+MFIn.
- the data packet type identifier can have at least 4 values, which are used to identify video, audio, private frames, etc.;
- the frame start identifier may have at least 2 values, which are used to identify whether the data packet is a data packet at the beginning of the frame;
- the identifier "1" indicates that the data packet is the data packet at the beginning of the frame
- the identifier "0" indicates that the data packet is not the data packet at the beginning of the frame.
- the end-of-frame identification may have at least 2 values, which are used to identify whether the data packet is an end-of-frame data packet, respectively;
- the identifier "1" indicates that the data packet is an end-of-frame data packet
- the identifier "0" indicates that the data packet is not an end-of-frame data packet.
- the frame type identifier may have at least 4 values, which are used to identify the I frame, P frame or B frame respectively when the packet type is video;
- the composite frame number is used to identify the track where the data packet is (one camera corresponds to one track);
- the total number of tracks in the composite frame is used to identify the total number of tracks.
- the information included in the MFI byte is not limited to the above information.
- the MFI byte may also include the position information of each track image in the mosaic in the multimedia data stream.
- the receiving end device can splice the images of each track according to the position information of each track image included in the MFI byte in the mosaic map, which is specifically implemented in I will not repeat them here.
- the multimedia data RTP stream can be sent to the NVR.
- the NVR After receiving the multimedia data RTP stream, the NVR can unpack, strip the encapsulated data header, obtain each track code stream, and perform display, storage or network transmission processing according to requirements.
- the schematic diagram can be shown in FIG. 4C.
- the NVR After the NVR decapsulates the multimedia data RTP stream to obtain each track code stream (ES stream), the NVR can encapsulate each track code stream into a multimedia data PS stream and store it.
- FIG. 4D is a schematic diagram of an ordinary PES package.
- MFI information is added to the stuff field of the PES header encapsulated by the ordinary PS, and the load information is encapsulated later.
- the schematic diagram may be as shown in FIG. 4E.
- the multimedia data stream of n tracks is PES0 header+MFI0+PES0 payload+PES1 header+MFI1+PES1 payload+...+PESn header+MFIn+PESn payload, so that the coded streams from different cameras are combined into a multimedia data stream (also Called "multi-track flow").
- the NVR After the NVR decapsulates the multimedia data RTP stream to obtain each track code stream, the NVR can re-encapsulate each track code stream into a multimedia data RTP stream and perform network transmission.
- the NVR decapsulates the multimedia data RTP stream to obtain each track code stream
- the frame type, track number, frame start, and frame end of the track are determined according to the MFI information.
- the NVR can determine the position of the video image of a single track in the mosaic according to the track number and the corresponding width and height of the frame obtained by the package analysis.
- multiple track code streams are decoded in parallel by multiple decoders. After parsing into multiple track ES streams, the ES stream with the same track number is decoded with the same decoder, and the decoded video image is stitched according to the position of the track and displayed.
- the NVR can decode the ES streams of each track synchronously (parallel decoding) after parsing multiple track ES streams.
- the processing flow of accessing the code stream to the decoding and splicing display may include:
- the access code stream Sn-i (n: the number of streams of the multimedia data stream; i: packet sequence number), stripping the encapsulation information, parsing the MFI to obtain the frame type (I frame, P frame or B frame), where Track, and frame start identification, frame end identification and other information, so as to get the En-i (I frame, P frame..., n: the number of streams of the multimedia data stream; i: packet sequence number) naked stream for each track.
- the En-i bare stream is assembled into a complete frame, it is sent to the parallel decoder n (nth decoder) according to the track where it is decoded, and the video image n (the decoded image data of the nth track is obtained) ).
- Each video image obtains the position of each video image in the mosaic according to the track information and width and height information obtained in step 1, and performs image stitching according to the position to obtain a complete video image.
- the complete video image is reduced or enlarged according to the output resolution and output to the monitor for display.
- the sending end device may encapsulate multiple tracks of multimedia data into a single multimedia data stream, and transmit the multimedia data stream to the receiving end device.
- the receiving end device decapsulates the multimedia data stream to obtain the encoded multimedia data of each track, and processes the encoded multimedia data of each track. In this way, in the case of ensuring the synchronization of multimedia data of multiple tracks, the performance requirements of the encoding end device and the decoding end device are reduced, and the applicable scenarios of the solution are expanded.
- FIG. 5 is a schematic structural diagram of a multimedia data processing apparatus according to an embodiment of the present application, where the multimedia data processing apparatus may be applied to the sending end device in the foregoing method embodiment, as shown in FIG. 5, the multimedia The data processing device may include:
- the obtaining unit 500 is used to obtain the encoded multimedia data of each track
- An encapsulating unit 510 configured to encapsulate the encoded multimedia data of the multiple tracks into a single multimedia data stream, wherein the multimedia data stream includes the encoded multimedia data of each track and the encoding of each track Synchronization information of multimedia data synchronization;
- the transmission unit 520 is configured to transmit the multimedia data stream to the receiving end device, and the receiving end device decapsulates the multimedia data stream to obtain encoded multimedia data of each track, and based on the synchronization information The encoded multimedia data of each track is processed.
- the device further includes an encoding unit 530,
- the acquiring unit 500 is specifically configured to acquire multimedia data of multiple tracks from a single data collection device, where the multimedia data of the multiple tracks is multimedia data taken by multiple sensors provided in the data collection device for the same scene;
- the encoding unit 530 is configured to separately encode the multimedia data of each track to obtain the encoded multimedia data.
- the acquiring unit 500 is specifically configured to acquire encoded multimedia data of multiple tracks from multiple data collection devices.
- the obtaining unit 500 is specifically configured to divide multimedia data according to the number of the tracks to obtain multimedia data of the multiple tracks;
- the encoding unit 530 is configured to encode the multimedia data of each track to obtain the encoded multimedia data.
- the encapsulation unit 510 is specifically configured to decapsulate pre-stored multimedia data according to a preset storage encapsulation protocol to obtain the encoded multimedia data of each track.
- the encapsulation unit 510 is further specifically configured to encapsulate the encoded multimedia data of each track into the single multimedia data stream according to a preset transmission encapsulation protocol.
- the synchronization information includes synthesized frame information carried in a data packet of the multimedia data stream
- the composite frame information includes at least one of the following: a packet type identifier, a frame start identifier and a frame end identifier, a frame type identifier, a composite frame number, and a total number of composite frame tracks.
- FIG. 7 is a schematic diagram of a hardware structure of an electronic device provided by an embodiment of the present application.
- the electronic device may include a processor 701, a communication interface 702, a storage medium 703, and a communication bus 704.
- the processor 701, the communication interface 702, and the storage medium 703 communicate with each other through the communication bus 704.
- the machine executable instructions are stored on the storage medium 703; the processor 701 can execute the multimedia data processing method shown in FIG. 2 by executing the machine executable instructions stored on the storage medium 703.
- the storage medium 703 mentioned herein may be any electronic, magnetic, optical, or other physical storage device, and may contain or store information, such as executable instructions, data, and so on.
- the storage medium 702 may be: RAM (Radom Access Memory, random access memory), volatile memory, non-volatile memory, flash memory, storage drive (such as a hard disk drive), solid-state drive, any type of storage disk (such as CD-ROM, DVD, etc.), or similar storage media, or a combination thereof.
- An embodiment of the present application further provides a machine-readable storage medium storing machine-executable instructions, such as the storage medium 703 in FIG. 7, and the machine-executable instructions may be executed by the processor 701 in the electronic device shown in FIG. 7 In order to realize the multimedia data processing method shown in FIG. 2.
- FIG. 8 is a schematic structural diagram of a multimedia data processing apparatus according to an embodiment of the present application.
- the multimedia data processing apparatus may be applied to the receiving device in the foregoing method embodiment.
- the multimedia The data processing device may include:
- the receiving unit 810 is configured to receive a multimedia data stream sent by a sending-end device; wherein the multimedia data stream includes encoded multimedia data of multiple tracks and synchronization information for synchronizing the encoded multimedia data of each track, wherein each The encoded multimedia data of the track is data obtained by separately encoding the multimedia data of each track;
- the decapsulation unit 820 is used to decapsulate the multimedia data stream to obtain encoded multimedia data of each track;
- the processing unit 830 is configured to process the encoded multimedia data of each track based on the synchronization information.
- the multimedia data is video data
- the processing unit 830 is specifically configured to decode the encoded video frames of each track based on the synchronization information to obtain a decoded video image; and process the decoded video image.
- the synchronization information includes synthesized frame information carried in a data packet of a multimedia data stream
- the composite frame information includes at least one of the following: a packet type identifier, a frame start identifier and a frame end identifier, a frame type identifier, a composite frame number, and a total number of tracks of the composite frame;
- the processing unit 830 is specifically configured to obtain the complete video frame of each track according to the packet type identifier, the frame start identifier, the frame end identifier, and the synthesized frame sequence number; and decode the complete video frame of each track.
- the processing unit 830 is specifically configured to decode a complete video frame of each track through multiple parallel decoders.
- the processing unit 830 is specifically configured to send the complete video frame of each track to the corresponding decoder for decoding.
- the processing unit 830 is specifically configured to determine the stitching position information of each video image according to the track where the decoded video image is located and the width and height of the decoded video image; The position information stitches the decoded video image to obtain a stitched video image; and displays the stitched video image.
- processing unit 830 is further configured to scale the stitched video image according to the output resolution
- the processing unit 830 is also specifically used to display the zoomed mosaic video image.
- the device further includes: a first packaging unit 840;
- the synchronization information includes synthesized frame information carried in a data packet of the multimedia data stream;
- the first encapsulation unit 840 is configured to encapsulate the encoded multimedia data of each track into a single multimedia data stream according to a preset transmission encapsulation protocol, where the multimedia data stream carries synthesized frame information, and the synthesized frame information includes at least one of the following : Data packet type identification, frame start identification and frame end identification, composite frame sequence number, total frame track number;
- the processing unit 830 is specifically configured to perform network transmission on the multimedia data stream encapsulated according to the preset transmission encapsulation protocol.
- the device further includes: a second packaging unit 850,
- the synchronization information includes synthesized frame information carried in a data packet of the multimedia data stream;
- the second encapsulation unit 850 is further configured to encapsulate the encoded multimedia data of each track into a single multimedia data stream according to a preset storage and encapsulation protocol, where the data packet of the multimedia data stream carries synthesized frame information, and the synthesized frame information includes At least one of the following: packet type identification, frame start identification and frame end identification, composite frame number, and total frame track number;
- the processing unit 830 is specifically configured to store the multimedia data stream encapsulated according to the preset storage encapsulation protocol.
- the processing unit 830 is further specifically configured to: cut the decoded video images of each track according to a preset condition to obtain a cut video image; cut the cut video of each track Images are coded separately to obtain coded video data for each track;
- the first encapsulation unit 840 is specifically configured to encapsulate the encoded video data of each track into a single multimedia video stream according to a preset transmission encapsulation protocol for network transmission, or
- the second encapsulation unit 850 is specifically configured to encapsulate the encoded video data of each track into a single multimedia video stream according to a preset storage encapsulation protocol for storage.
- FIG. 11 is a schematic diagram of a hardware structure of an electronic device provided by an embodiment of the present application.
- the electronic device may include a processor 1101, a communication interface 1102, a storage medium 1103, and a communication bus 1104.
- the processor 1101, the communication interface 1102, and the storage medium 1103 communicate with each other through the communication bus 1104.
- the machine executable instructions are stored on the storage medium 1103; the processor 1101 can execute the multimedia data processing method shown in FIG. 3 by executing the machine executable instructions stored on the storage medium 1103.
- the storage medium 1103 mentioned herein may be any electronic, magnetic, optical, or other physical storage device, and may contain or store information, such as executable instructions, data, and so on.
- the storage medium 1103 may be: RAM (Radom Access Memory, random access memory), volatile memory, non-volatile memory, flash memory, storage drive (such as a hard disk drive), solid-state drive, any type of storage disk (such as CD-ROM, DVD, etc.), or similar storage media, or a combination thereof.
- An embodiment of the present application further provides a machine-readable storage medium storing machine-executable instructions, such as the storage medium 1103 in FIG. 11, and the machine-executable instructions may be executed by the processor 1101 in the electronic device shown in FIG. 11 In order to realize the multimedia data processing method shown in FIG. 3.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Security & Cryptography (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (38)
- 一种多媒体数据处理方法,应用于发送端设备,其特征在于,该方法包括:获取多个轨道各自的编码多媒体数据;将所述多个轨道的编码多媒体数据封装为单个多媒体数据流,其中,所述多媒体数据流包括各个轨道的所述编码多媒体数据,以及用于将各个轨道的所述编码多媒体数据同步的同步信息;将所述多媒体数据流传输给接收端设备,由所述接收端设备对所述多媒体数据流进行解封装,以得到各个轨道的编码多媒体数据,并基于所述同步信息对各个轨道的编码多媒体数据进行处理。
- 根据权利要求1所述的方法,其特征在于,所述获取多个轨道各自的编码多媒体数据包括:从单个数据采集设备获取多个轨道的多媒体数据,其中,所述多个轨道的多媒体数据为所述数据采集设备中设置的多个传感器针对同一场景拍摄的多媒体数据;分别对各个轨道的多媒体数据进行编码以获取所述编码多媒体数据。
- 根据权利要求1所述的方法,其特征在于,所述获取多个轨道各自的编码多媒体数据包括:从多个数据采集设备获取多个轨道的编码多媒体数据。
- 根据权利要求1所述的方法,其特征在于,所述获取多个轨道各自的编码多媒体数据包括:将多媒体数据按照所述轨道的数量进行划分,以得到所述多个轨道的多媒体数据;分别对各个轨道的多媒体数据进行编码以获取所述编码多媒体数据。
- 根据权利要求1所述的方法,其特征在于,所述获取多个轨道各自的编码多媒体数据包括:按照预设存储封装协议对预存储的多媒体数据进行解封装,以得到各个轨道的所述编码多媒体数据。
- 根据权利要求1至5任一所述的方法,其特征在于,所述将所述多个轨道的编码多媒体数据封装为单个多媒体数据流,包括:按照预设传输封装协议将各个轨道的编码多媒体数据封装为所述单个多媒体数据流。
- 根据权利要求1至6任一所述的方法,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;其中,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、帧类型标识、合成帧序号、合成帧的轨道总数。
- 一种多媒体数据处理方法,应用于接收端设备,其特征在于,该方法包括:接收发送端设备发送的多媒体数据流;所述多媒体数据流包括多个轨道的编码多媒体数据,以及用于将各个轨道的编码多媒体数据同步的同步信息,其中,各个轨道的编码多媒体数据为分别针对各个轨道的多媒体数据单独编码获得的数据;对所述多媒体数据流进行解封装,以得到各个轨道的编码多媒体数据;基于所述同步信息对各个轨道的编码多媒体数据进行处理。
- 根据权利要求8所述的方法,其特征在于,所述多媒体数据为视频数据;所述基于所述同步信息对各个轨道的编码多媒体数据进行处理,包括:基于所述同步信息对各个轨道的编码视频帧进行解码以得到解码后的视频图像;对解码后的视频图像进行处理。
- 根据权利要求9所述的方法,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、帧类型标识、合成帧序号、合成帧的轨道总数;所述基于所述同步信息对各个轨道的编码视频帧进行解码,包括:根据所述数据包类型标识、帧开始标识和帧结束标识、以及合成帧序号,分别得到各个轨道的完整视频帧;对各个轨道的完整视频帧进行解码。
- 根据权利要求9所述的方法,其特征在于,所述对各个轨道的完整视频帧进行解码,包括:通过多个并行的解码器分别对各个轨道的完整视频帧进行解码。
- 根据权利要求11所述的方法,其特征在于,所述通过多个并行的解码器分别对各个轨道的完整视频帧进行解码,包括:将各个轨道的完整视频帧发送至与该轨道对应的解码器进行解码。
- 根据权利要求9所述的方法,其特征在于,所述对解码后的视频图像进行处理,包括:根据解码后的视频图像所在轨道,以及解码后的视频图像的宽高,确定各视频图像的拼接位置信息;根据所述位置信息对解码后的视频图像进行拼接以得到拼接视频图像;对拼接视频图像进行显示。
- 根据权利要求13所述的方法,其特征在于,所述对拼接视频图像进行显示之前,还包括:根据输出分辨率对拼接视频图像进行缩放;所述对拼接视频图像进行显示,包括:对缩放后的拼接视频图像进行显示。
- 根据权利要求8所述的方法,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;基于所述同步信息对各个轨道的编码多媒体数据进行处理,包括:按照预设传输封装协议将各个轨道的编码多媒体数据封装为单个多媒体数据流,其中所述多媒体数据流的数据包中携带有所述合成帧信息,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、合成帧序号、合成帧的轨道总数;对该按照预设传输封装协议封装的多媒体数据流进行网络传输。
- 根据权利要求8所述的方法,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;所述基于所述同步信息对各个轨道的编码多媒体数据进行处理,包括:按照预设存储封装协议将各个轨道的编码多媒体数据封装为单个多媒体数据流,其中所述多媒体数据流的数据包中携带有所述合成帧信息,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、合成帧序号、合成帧的轨道总数;对该按照预设存储封装协议封装的多媒体数据流进行存储。
- 根据权利要求9所述的方法,其特征在于,所述对解码后的视频图像进行处理,包括:根据预设条件对各个轨道的解码后的视频图像进行切割,以得到切割后的视频图像;将各个轨道的切割后的视频图像单独编码以获得各个轨道的编码视频数据;将各个轨道的编码视频数据按照预设存储封装协议封装为单个多媒体视频流以进行存储,或者按照预设传输封装协议封装以进行网络传输。
- 一种多媒体数据处理装置,应用于发送端设备,其特征在于,该装置包括:获取单元,用于获取多个轨道各自的编码多媒体数据;封装单元,用于将所述多个轨道的编码多媒体数据封装为单个多媒体数据流,其中,所述多媒体数据流包括各个轨道的所述编码多媒体数据,以及用于将各个轨道的所述编码多媒体数据同步的同步信息;传输单元,用于将所述多媒体数据流传输给接收端设备,由所述接收端设备对所述多媒体数据流进行解封装,以得到各个轨道的编码多媒体数据,并基于所述同步信息对各个轨道的编码多媒体数据进行处理。
- 根据权利要求18所述的装置,其特征在于,所述装置还包括编码单元,所述获取单元具体用于:从单个数据采集设备获取多个轨道的多媒体数据,其中, 所述多个轨道的多媒体数据为所述数据采集设备中设置的多个传感器针对同一场景拍摄的多媒体数据;所述编码单元用于:分别对各个轨道的多媒体数据进行编码以获取所述编码多媒体数据。
- 根据权利要求18所述的装置,其特征在于,所述获取单元具体用于:从多个数据采集设备获取多个轨道的编码多媒体数据。
- 根据权利要求18所述的装置,其特征在于,所述装置还包括编码单元,所述获取单元具体用于:将多媒体数据按照所述轨道的数量进行划分,以得到所述多个轨道的多媒体数据;所述编码单元用于:分别对各个轨道的多媒体数据进行编码以获取所述编码多媒体数据。
- 根据权利要求18所述的装置,其特征在于,所述封装单元具体用于:按照预设存储封装协议对预存储的多媒体数据进行解封装,以得到各个轨道的所述编码多媒体数据。
- 根据权利要求18至22任一所述的装置,其特征在于,所述封装单元具体用于:按照预设传输封装协议将各个轨道的编码多媒体数据封装为所述单个多媒体数据流。
- 根据权利要求18至23任一所述的装置,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;其中,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、帧类型标识、合成帧序号、合成帧的轨道总数。
- 一种多媒体数据处理装置,应用于接收端设备,其特征在于,该装置包括:接收单元,用于接收发送端设备发送的多媒体数据流;所述多媒体数据流包括多个轨道的编码多媒体数据,以及用于将各个轨道的编码多媒体数据同步的同步信息,其中,各个轨道的编码多媒体数据为分别针对各个轨道的多媒体数据单独编码获得的数据;解封装单元,用于对所述多媒体数据流进行解封装,以得到各个轨道的编码多媒体数据;处理单元,用于基于所述同步信息对各个轨道的编码多媒体数据进行处理。
- 根据权利要求25所述的装置,其特征在于,所述多媒体数据为视频数据;所述处理单元具体用于:基于所述同步信息对各个轨道的编码视频帧进行解码以得到解码后的视频图像;对解码后的视频图像进行处理。
- 根据权利要求26所述的装置,其特征在于,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、帧类型标识、合成帧序号、合成帧的轨道总数;所述处理单元具体用于:根据所述数据包类型标识、帧开始标识和帧结束标识、以及合成帧序号,分别得到各个轨道的完整视频帧;对各个轨道的完整视频帧进行解码。
- 根据权利要求26所述的装置,其特征在于,所述处理单元具体用于:通过多个并行的解码器分别对各个轨道的完整视频帧进行解码。
- 根据权利要求28所述的装置,其特征在于,所述处理单元具体用于:将各个轨道的完整视频帧发送至与该轨道对应的解码器进行解码。
- 根据权利要求26所述的装置,其特征在于,所述处理器具体用于:根据解码后的视频图像所在轨道,以及解码后的视频图像的宽高,确定各视频图像的拼接位置信息;根据所述位置信息对解码后的视频图像进行拼接以得到拼接视频图像;对拼接视频图像进行显示。
- 根据权利要求30所述的装置,其特征在于,所述处理单元具体用于:根据输出分辨率对拼接视频图像进行缩放;对缩放后的拼接视频图像进行显示。
- 根据权利要求25所述的装置,其特征在于,所述装置还包括:第一封装单元,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;所述第一封装单元用于:按照预设传输封装协议将各个轨道的编码多媒体数据封装为单个多媒体数据流,其中所述多媒体数据流的数据包中携带有所述合成帧信息,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、合成帧序号、合成帧的轨道总数;所述处理单元具体用于:对该按照预设传输封装协议封装的多媒体数据流进行网络传输。
- 根据权利要求25所述的装置,其特征在于,所述装置还包括:第二封装单元,所述同步信息包括所述多媒体数据流的数据包中携带的合成帧信息;所述第二封装单元用于:按照预设存储封装协议将各个轨道的编码多媒体数据封装为单个多媒体数据流,其中所述多媒体数据流的数据包中携带有所述合成帧信息,所述合成帧信息包括以下至少一者:数据包类型标识、帧开始标识和帧结束标识、合成帧序号、合成帧的轨道总数;所述处理单元具体用于对该按照预设存储封装协议封装的多媒体数据流进行存储。
- 根据权利要求26所述的装置,其特征在于,所述处理单元具体还用于:根据预设条件对各个轨道的解码后的视频图像进行切 割,以得到切割后的视频图像;将各个轨道的切割后的视频图像单独编码以获得各个轨道的编码视频数据;所述第一封装单元具体用于将各个轨道的编码视频数据按照预设传输封装协议封装为单个多媒体视频流以进行网络传输,或者所述第二封装单元具体用于将各个轨道的编码视频数据按照预设存储封装协议将封装为单个多媒体视频流以进行存储。
- 一种电子设备,其包括:处理器;通信接口、非暂时性存储介质、以及通信总线,其中,所述处理器、通信接口、存储介质通过所述通信总线完成相互间的通信,所述处理器通过执行所述存储介质上的机器可执行指令,实现如权利要求1至7任一所述的方法。
- 一种电子设备,其包括:处理器;通信接口、非暂时性存储介质、以及通信总线,其中,所述处理器、通信接口、存储介质通过所述通信总线完成相互间的通信,所述处理器通过执行所述存储介质上的机器可执行指令,实现如权利要求8至17任一所述的方法。
- 一种机器可读存储介质,其特征在于,所述机器可读存储介质内存储有机器可执行指令,所述机器可执行指令被处理器执行时实现权利要求1-7任一所述的方法。
- 一种机器可读存储介质,其特征在于,所述机器可读存储介质内存储有机器可执行指令,所述机器可执行指令被处理器执行时实现权利要求8-17任一所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811589685.2 | 2018-12-25 | ||
CN201811589685.2A CN111372035A (zh) | 2018-12-25 | 2018-12-25 | 多媒体数据处理方法、装置、电子设备及可读存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020135527A1 true WO2020135527A1 (zh) | 2020-07-02 |
Family
ID=71128735
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/128377 WO2020135527A1 (zh) | 2018-12-25 | 2019-12-25 | 多媒体数据处理 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111372035A (zh) |
WO (1) | WO2020135527A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114466246A (zh) * | 2022-02-14 | 2022-05-10 | 维沃移动通信有限公司 | 一种视频处理方法及其装置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5995146A (en) * | 1997-01-24 | 1999-11-30 | Pathway, Inc. | Multiple video screen display system |
CN102301730A (zh) * | 2011-07-18 | 2011-12-28 | 华为技术有限公司 | 多路音视频传输和处理方法、装置及系统 |
CN102547212A (zh) * | 2011-12-13 | 2012-07-04 | 浙江元亨通信技术股份有限公司 | 多路视频图像的拼接方法 |
CN103581609A (zh) * | 2012-07-23 | 2014-02-12 | 中兴通讯股份有限公司 | 一种视频处理方法及装置、系统 |
CN104243920A (zh) * | 2014-09-04 | 2014-12-24 | 浙江宇视科技有限公司 | 一种基于基本流视频数据封装的图像拼接方法及装置 |
CN107645647A (zh) * | 2017-09-21 | 2018-01-30 | 京信通信系统(中国)有限公司 | 一种多路音视频传输方法及装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102131097B (zh) * | 2011-03-21 | 2016-05-18 | 中国电信股份有限公司 | 3d流媒体传输方法及系统 |
CN103546827B (zh) * | 2013-10-30 | 2017-01-18 | 北京视博数字电视科技有限公司 | 用于传输电视系统中的传输流的方法、装置和系统 |
US10104361B2 (en) * | 2014-11-14 | 2018-10-16 | Samsung Electronics Co., Ltd. | Coding of 360 degree videos using region adaptive smoothing |
CN107124618A (zh) * | 2016-02-24 | 2017-09-01 | 深圳岚锋创视网络科技有限公司 | 实时全景视频录播方法和系统 |
CN107872422B (zh) * | 2016-09-23 | 2020-01-10 | 杭州海康威视数字技术股份有限公司 | 一种数据传输方法、装置及电子设备 |
CN107426491B (zh) * | 2017-05-17 | 2021-05-07 | 西安邮电大学 | 一种360度全景视频的实现方法 |
CN107529064A (zh) * | 2017-09-04 | 2017-12-29 | 北京理工大学 | 一种基于vr终端反馈的自适应编码方法 |
CN207443024U (zh) * | 2017-11-02 | 2018-06-01 | 深圳市裂石影音科技有限公司 | 全景音视频录制设备及系统 |
-
2018
- 2018-12-25 CN CN201811589685.2A patent/CN111372035A/zh active Pending
-
2019
- 2019-12-25 WO PCT/CN2019/128377 patent/WO2020135527A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5995146A (en) * | 1997-01-24 | 1999-11-30 | Pathway, Inc. | Multiple video screen display system |
CN102301730A (zh) * | 2011-07-18 | 2011-12-28 | 华为技术有限公司 | 多路音视频传输和处理方法、装置及系统 |
CN102547212A (zh) * | 2011-12-13 | 2012-07-04 | 浙江元亨通信技术股份有限公司 | 多路视频图像的拼接方法 |
CN103581609A (zh) * | 2012-07-23 | 2014-02-12 | 中兴通讯股份有限公司 | 一种视频处理方法及装置、系统 |
CN104243920A (zh) * | 2014-09-04 | 2014-12-24 | 浙江宇视科技有限公司 | 一种基于基本流视频数据封装的图像拼接方法及装置 |
CN107645647A (zh) * | 2017-09-21 | 2018-01-30 | 京信通信系统(中国)有限公司 | 一种多路音视频传输方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
CN111372035A (zh) | 2020-07-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101073777B1 (ko) | 다중화 방식 변환 장치 | |
CN104243920B (zh) | 一种基于基本流视频数据封装的图像拼接方法及装置 | |
KR101733501B1 (ko) | 방송신호 전송방법, 방송신호 수신방법, 방송신호 전송장치, 방송신호 수신장치 | |
KR101721884B1 (ko) | 방송신호 전송방법, 방송신호 수신방법, 방송신호 전송장치, 방송신호 수신장치 | |
CN101977305A (zh) | 一种视频处理方法及装置和系统 | |
KR101861941B1 (ko) | 완벽 스플라이싱을 위한 인코딩된 오디오의 전송 스트림에의 삽입 | |
WO2012106898A1 (zh) | 多路音视频传输和处理方法、装置及系统 | |
KR101764634B1 (ko) | 방송신호 전송방법, 방송신호 수신방법, 방송신호 전송장치, 방송신호 수신장치 | |
CN115567661B (zh) | 视频数据的处理方法、系统、计算机设备和存储介质 | |
CN112565224B (zh) | 一种视频处理方法及装置 | |
US12120380B2 (en) | Transmission apparatus, transmission method, encoding apparatus, encoding method, reception apparatus, and reception method | |
WO2020135527A1 (zh) | 多媒体数据处理 | |
JP2005123907A (ja) | データ再構成装置 | |
WO2024109317A1 (zh) | 一种传输视频帧及摄像参数信息的方法与设备 | |
CN107205158A (zh) | 一种基于时间戳的多路音视频流同步解码方法 | |
CN110719496A (zh) | 一种多路码流封装、播放方法、装置及系统 | |
CN115225542A (zh) | 视频信息处理方法、装置、电子设备和存储介质 | |
TWI713364B (zh) | 藉由現有hd視訊架構編碼原始高幀率視訊的方法 | |
CN113784094A (zh) | 视频数据处理方法、网关、终端设备及存储介质 | |
CN113950825A (zh) | 数据传输、处理方法、可移动平台、终端设备及芯片 | |
WO2020001431A1 (zh) | 摄像装置和播放装置 | |
Han et al. | An implementation of capture and playback for ip-encapsulated video in professional media production | |
KR20200011305A (ko) | 영상을 전송하는 방법 및 장치, 영상을 수신하는 방법 및 장치 | |
US11758108B2 (en) | Image transmission method, image display device, image processing device, image transmission system, and image transmission system with high-transmission efficiency | |
JP5367771B2 (ja) | 映像伝送システム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19905015 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19905015 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19905015 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03/02/2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19905015 Country of ref document: EP Kind code of ref document: A1 |