CN115225928B - Multi-type audio and video mixed broadcasting system and method - Google Patents

Multi-type audio and video mixed broadcasting system and method Download PDF

Info

Publication number
CN115225928B
CN115225928B CN202210813711.5A CN202210813711A CN115225928B CN 115225928 B CN115225928 B CN 115225928B CN 202210813711 A CN202210813711 A CN 202210813711A CN 115225928 B CN115225928 B CN 115225928B
Authority
CN
China
Prior art keywords
audio
video
playing
video data
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210813711.5A
Other languages
Chinese (zh)
Other versions
CN115225928A (en
Inventor
陈祥
毕江
芮浩
赵涛
冯澈
李浩楠
欧晓霄
宋峥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Radio And Television Station
Original Assignee
Beijing Radio And Television Station
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Radio And Television Station filed Critical Beijing Radio And Television Station
Publication of CN115225928A publication Critical patent/CN115225928A/en
Application granted granted Critical
Publication of CN115225928B publication Critical patent/CN115225928B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234309Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4 or from Quicktime to Realvideo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234363Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234381Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the temporal resolution, e.g. decreasing the frame rate by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2662Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440263Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the spatial resolution, e.g. for displaying on a connected PDA
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/65Transmission of management data between client and server
    • H04N21/658Transmission by the client directed to the server
    • H04N21/6587Control parameters, e.g. trick play commands, viewpoint selection

Abstract

The invention relates to a multi-type audio and video mixed broadcasting system and a method, wherein the system comprises the following steps: the system comprises an audio and video acquisition and transmission module, an audio and video processor and a playing interaction terminal, wherein the audio and video acquisition and transmission module is respectively connected with the audio and video processor in a signal transmission manner, and the playing interaction terminal is connected with the audio and video processor in a signal transmission manner; the audio and video acquisition and transmission module comprises a first audio and video acquisition device, an audio and video encoder and an audio and video transmission device; the first audio and video collector comprises a hundred million-level pixel array image sensor camera; the audio and video processor comprises an audio and video type analysis module, a video transcoding module, a video fusion device and an information stream sending and receiving module; the playing interaction terminal comprises a user interaction module and an audio and video playing module, and can display various different types of audios in a unified information stream in a mixed mode, so that a user can watch various types of audios and videos in the same information stream.

Description

Multi-type audio and video mixed broadcasting system and method
The present application claims priority from the chinese patent application filed at month 11 of 2022, under the application number 202210532881.6, entitled "a multi-type audio/video mixed broadcast system", the entire contents of which are incorporated herein by reference.
Technical Field
The invention relates to the technical field of audio and video playing, in particular to a multi-type audio and video mixed playing system.
Background
By installing a playing client capable of playing audio and video in a terminal such as an electronic device, a user can enjoy the audio and video through the client. In the related art, a main interface of the client may display a designated category, such as "VR channel", "live channel", etc., and the user may view the audio file of the type in the corresponding channel and click to play. However, in the related art, it is difficult to mix and display various different types of audio such as slow live broadcast, 4K, 8K, VR gasket video, interactive video, multi-site live broadcast and the like in a unified information stream, and a user can usually only view a single type of audio and video in the same information stream, and cannot play various different types of audio and video in a client interface at the same time, so that the user is influenced to obtain comprehensive and rich viewing experience.
Disclosure of Invention
Aiming at the defects that different types of audios and videos cannot be displayed in a mixed mode in the same information stream in the prior art and the like, the invention provides a multi-type audio/video mixed broadcasting system, which at least partially realizes the purpose of playing multiple types of audios and videos in the same information stream, comprising the following steps: the system comprises an audio and video acquisition and transmission module, an audio and video processor and a playing interaction terminal, wherein the audio and video acquisition and transmission module is respectively connected with the audio and video processor in a signal transmission manner, and the playing interaction terminal is connected with the audio and video processor in a signal transmission manner; the audio and video acquisition and transmission module comprises a first audio and video acquisition device, an audio and video encoder and an audio and video transmission device; the first audio and video collector comprises a hundred million-level pixel array image sensor camera; the audio and video processor comprises an audio and video type analysis module, a video transcoding module, a video fusion device and an information stream sending and receiving module; the playing interaction terminal comprises a user interaction module and an audio and video playing module.
Optionally, the audio-video encoder comprises a 4K encoder and/or an 8K encoder.
Optionally, the audio-video encoder further includes at least one of a standard definition encoder, a high definition encoder and a 2K encoder.
Optionally, the user interaction module comprises a user habit analysis device and a second audio/video collector.
Optionally, the audio/video playing module at least comprises one of a slow live audio/video playing unit, an interactive video playing unit and a multi-position live video playing unit.
Optionally, the multi-type audio and video mixed broadcasting system further comprises a memory, and the memory is at least connected with one of the audio and video acquisition and transmission module, the audio and video processor and the playing interaction terminal.
The technical scheme provided by the embodiment of the invention has the beneficial effects that at least:
the audio and video are fused by collecting the type information of the audio and video, the audio and video are displayed on the same interface in the same information stream according to the user requirement, and various different types of audio such as slow live broadcast, 4K, 8K, VR gasket video, interactive video, multi-position live broadcast and the like are mixed and displayed in the unified information stream, so that the user can watch various types of audio and video in the same information stream, the watching experience is greatly improved, the watching efficiency is improved, and further, the time waiting for the loading of the audio and video is shortened.
As a first technical solution of the present invention, a multi-type audio/video mixed broadcast system includes: an audio/video processor 200, a play interactive terminal 300,
the audio/video processor 200 includes an information stream receiving module 204a, an information stream sending module 204b, a tag adding module, a video transcoding module 202, and a video fusion module.
The information stream receiving module 204a receives audio/video data sent by a plurality of audio/video sources, and at least a part of the audio/video data is audio/video data of different types.
And the tag adding module adds tags to the audio and video data, and the tags are associated with the corresponding audio and video data.
The video transcoding module 202 performs unified transcoding on the video in each audio/video data, and each audio/video data at least obtains two audio/video data with different resolutions, and the audio/video data with low resolution and/or high compression rate is used as audio/video data for browsing, and the audio/video data with high resolution and/or low compression rate is used as audio/video data for playing.
And the video fusion module is used for carrying out fusion processing on the plurality of audio and video data for browsing, and splicing the audio and video data into an audio and video data stream which is used as the audio and video data stream for browsing.
The information stream sending module 204b sends the browsing audio/video data stream and the playing audio/video data stream to the playing interactive terminal 300 in different information streams, the information stream sending the browsing audio/video data stream is a browsing data information stream, and the information stream sending the playing audio/video data stream is a playing data information stream.
The playing interaction terminal 300 comprises a communication module 304, an audio/video playing module 302 and a user operation identification module 303.
The communication module 304 receives the browsing data information stream and the playing data information stream sent by the audio/video processor 200.
The audio/video playing module 302 selects to browse the data information stream or play the data information stream for playing according to the preset condition.
The user operation identification module 303 identifies an operation mode of a user when the audio/video playing module 302 plays the browsing data information stream, including a video moving operation and a video skipping operation, when the video moving operation is identified, the audio/video playing module 302 plays the next or previous audio/video of the browsing data information stream, and when the video skipping operation is identified, the audio/video playing module 302 skips to the corresponding playing data information stream according to the label corresponding to the operation to play.
Because 1) the audio and video data of different types are subjected to unified transcoding to generate audio and video data with unified video formats, and the audio and video data with at least two video formats which are the same and have different resolutions and/or compression ratios after transcoding are corresponding to the audio and video data before transcoding, the audio and video data with low resolution and/or high compression ratio is used as audio and video data for browsing, the audio and video data with high resolution and/or low compression ratio is used as audio and video data for playing, 2) each audio and video data for browsing is fused into one audio and video data stream which is sent to the playing interaction terminal 300 through the information stream, and the playing audio and video data for playing is sent to the playing interaction terminal 300 through other information streams, when the playing interaction terminal 300 plays the audio and video information stream, for example, the original different types of audio and video can be browsed through sliding operation, and when the high-quality video needs to be watched, for example, the user can jump to the high-quality video by clicking the tag.
In the whole playing process, no matter the audio and video data for browsing or playing, because the audio and video formats are the same, the audio and video playing module 302 does not need to reload for different formats, and when browsing or jumping, the audio and video is not required to be delayed in time for loading data, so that the connection of the audio and video is smooth, and the watching experience of a user is improved.
Preferably, in the recognition mode of the user operation recognition module 303, a return operation is included,
in the state of playing the playing data information stream, when the user operation identification module 303 identifies a return operation, the audio/video playing module 302 returns the browsing data information stream to play.
Therefore, the user can return to the audio and video for browsing when needed, and random switching between browsing the video and playing the video is realized.
Preferably, the tag adding module analyzes each audio and video data, and sets a pattern of the tag according to the type of the audio and video, and one or more tags are provided.
Therefore, the user can confirm the type of the audio and video according to the pattern of the tag.
Preferably, the tag adding module adds the tag to the lower side of the video.
Therefore, the tag is not only easy to operate, but also does not obstruct the user's video viewing.
Preferably, the video transcoding module 202 adjusts the audio/video data for browsing and the audio/video data for playing to the same resolution when performing unified transcoding on the video in each of the audio/video data.
Therefore, when browsing or watching, no offensive sense is generated due to the switching of the video.
Preferably, the video transcoding module 202 generates a plurality of audio/video data for playing with different resolutions when performing unified transcoding on the video in the audio/video data, and the information stream sending module 204b sends the audio/video data for playing with different resolutions to the playing interactive terminal 300 with different information streams.
Therefore, in playing the interactive terminal 300, the user can select a suitable resolution for playing according to actual needs, such as processing capability, network speed, etc., so as to avoid the problem of jamming, etc.
As a second technical solution of the present invention, a multi-type audio/video mixed broadcasting method is executed by the audio/video processor 200 and the playing interactive terminal 300, and includes the following steps: in the audio/video data receiving step S10, the audio/video processor 200 receives audio/video data sent by a plurality of audio/video sources, and at least a part of the audio/video data is audio/video data of different types.
And a tag adding step S20, wherein the audio/video processor 200 adds a tag to each audio/video data, and the tag is associated with the corresponding audio/video data.
In the video transcoding step S40, the audio/video processor 200 performs unified transcoding on the video in each audio/video data, and each audio/video data at least obtains two audio/video data with different resolutions, where the audio/video data with low resolution and/or high compression ratio is used as audio/video data for browsing, and the audio/video data with high resolution and/or low compression ratio is used as audio/video data for playing.
In the video fusion step S50, the audio/video processor 200 performs fusion processing on the plurality of audio/video data for browsing, and splices the audio/video data into an audio/video data stream as the audio/video data stream for browsing.
In the information stream transmission step S60, the audio/video processor 200 transmits the browsing audio/video data stream and the playing audio/video data stream to the playing interactive terminal 300 with different information streams, and the information stream for transmitting the browsing audio/video data stream is a browsing data information stream, and the information stream for transmitting the playing audio/video data is a playing data information stream.
In the information stream receiving step S110, the playing interaction terminal 300 receives the browsing data information stream and the playing data information stream sent by the audio/video processor 200.
In the audio/video playing step S120, the playing interactive terminal 300 selects to browse the data information stream or play the data information stream according to the preset condition for playing.
The step S130 of identifying user operation, the step S130 of identifying the operation mode of the user when playing the browsing data information stream by the playing interactive terminal 300 includes a video moving operation and a video skipping operation, the steps S130, S140, S150 and S160 of switching audio and video, when identifying the video moving operation, the step S130, S170 and S180 of skipping audio and video of the playing interactive terminal 300 playing the next or last audio and video in the browsing data information stream, and when identifying the video skipping operation, the playing interactive terminal 300 skipping to the corresponding playing data information stream to play according to the tag corresponding to the operation.
Preferably, the method further includes a return operation identifying step S190, and when the return operation is identified by the playing interactive terminal 300 in the state of playing the playing data information stream, the browsing data information stream is returned to play.
Preferably, in the step S20 of adding a tag, the playing interaction terminal 300 analyzes each of the audio and video data, and sets a pattern of the tag according to the type of the audio and video, where the tag includes one or more tags.
Preferably, in the video transcoding step S40, when the audio/video processor 200 performs unified transcoding on the video in each of the audio/video data, a plurality of audio/video data for playing with different resolutions are generated, in the video merging step S50, when the video in each of the audio/video data is subjected to unified transcoding, a plurality of audio/video data for playing with different resolutions are generated, and in the information stream transmitting step S60, the audio/video processor 200 transmits the audio/video data for playing with different resolutions to the playing interaction terminal 300 with different information streams.
Drawings
Fig. 1 is a schematic structural diagram of a multi-type audio/video mixed broadcast system according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a memory in a multi-type audio/video mixed broadcast system according to an embodiment of the present invention;
FIG. 3 is a screenshot of a mixed playing interface of multiple types of audio and video in a playing interactive terminal in an embodiment of the invention;
fig. 4 shows a schematic diagram of a multi-type audio/video mixed broadcasting system as a second embodiment;
fig. 5 shows a structural diagram of a multi-type audio/video mixed broadcast system of the second embodiment;
fig. 6 shows a flowchart of operations in an audio-video processor in a second embodiment;
FIG. 7 shows a workflow diagram in a smartphone in a second embodiment;
fig. 8 shows a diagram illustrating movement of video when browsing video in the second embodiment.
Description of the embodiments
Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein reference numerals refer to the components and techniques of the present invention so that the advantages and features of the present invention may be more readily understood as implemented in a suitable environment. The following description is a specific embodiment of the present claims, and other specific implementations not explicitly described in connection with the claims also fall within the scope of the claims. Techniques, methods, and apparatus known to one of ordinary skill in the relevant art may not be discussed in detail, but are intended to be part of the specification where appropriate.
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the present invention without making any inventive effort, shall fall within the scope of the present invention.
In order to solve the problems in the prior art, the embodiment of the invention provides a multi-type audio/video mixed broadcasting system.
The first embodiment of the invention provides a multi-type audio and video mixed broadcasting system, the structure of which is shown in fig. 1, an audio and video collecting and transmitting module 100, an audio and video processor 200 and a playing interaction terminal 300, wherein the audio and video collecting and transmitting module 100 is respectively connected with the audio and video processor 200 in a signal transmission manner, and the playing interaction terminal 300 is connected with the audio and video processor 200 in a signal transmission manner;
the audio and video acquisition and transmission module 100 comprises a first audio and video acquisition device 101, an audio and video encoder 102 and an audio and video transmission device 103;
the first audio and video collector 101 can collect live audio and video from live equipment including a digital camera, a smart phone and the like through an audio and video transmission interface, and can also collect audio and video from storage equipment including a server, a memory and the like through a data transmission interface or a network transmission interface.
The first audio/video collector 101 includes a one hundred million-level pixel array image sensor camera 1011. The total resolution of the hundred million-level pixel array image sensor camera 1011 is 1.2 hundred million pixels, the resolution of the output video of each lens of the multi-lens array camera reaches 4K (3840 x 2160), the total horizontal viewing angle is 40 degrees, the total vertical viewing angle is 13 degrees, the frame rate is not lower than 25fps, the total code rate is 200-300 Mbps, the client mode is to support standard resolution live broadcast such as 8K, 4K, 1080P, 720P and the like, support arbitrary resolution live broadcast below 8K, the interaction delay is less than 300ms, the real-time screenshot can be realized, and the single picture reaches 1.2 hundred million pixels.
Optionally, the audio/video encoder 102 includes a 4K encoder and/or an 8K encoder.
Optionally, the audio/video encoder 102 further includes at least one of a standard definition encoder, a high definition encoder, and a 2K encoder.
The audio/video transmission device 103 transmits the encoded audio/video to the audio/video processor 200 in various transmission modes. When an abnormality occurs in the transmission to the audio/video processor 200, the encoded audio/video is transmitted to the memory, and the audio/video transmission information, the memory address, and the like are transmitted to the audio/video processor 200.
The audio and video processor 200 comprises an audio and video type analysis module 201, a video transcoding module 202, a video fusion device 203 and an information stream sending and receiving module 204;
the audio and video type analysis module 201 will analyze the audio and video type and add a corresponding tag. The user can click on the playing icon or the label information to jump to the full screen to further view the video details and display the technical label information of the video, such as 8K, HDR and frame rate. Labels include, but are not limited to: 8K, 4K, VR, slow live broadcast, multi-site live broadcast and interactive video.
The video transcoding module 202 performs intelligent transcoding processing on an input video file to realize that input formats supported by decoding include but are not limited to H.264, H.265 and VP9, output formats supported by encoding include but are not limited to H.264, H.265 and VP9, and video encapsulation formats supported by encoding include but are not limited to mp4 and mov, mkv, avi; the coding rate of the output video is determined in a self-adaptive manner according to the texture complexity, the motion complexity and the precoding result of the input video by self-adaptive rate control of the content complexity; by encoding the region of interest, code rate control based on scenes is realized; the video quality is improved through video preprocessing based on deep learning, and self-adaptive picture restoration processing of the video can be carried out, wherein the self-adaptive picture restoration processing comprises sharpening, noise reduction, decompression distortion and the like; including multi-resolution outputs, including same resolution outputs, upsampling, and downsampling. Specifically, in the aspect of 8K transcoding, transcoding covering 8K, 4K, 1080P and 720P resolutions is supported, transcoding processing can be performed on 8K input content at most, multi-resolution files are output at the same time, and 8K encoding and decoding capabilities covering H.264, hevc and AV1 formats are supported; HLS format output supporting 8K resolution content; support the input and processing of 8K ultra-high definition video, support 24fps, 25fps, 30fps, 50fps, 60fps frame rate input; supporting SDR and HDR inputs; mp4, mov, mkv, avi video format inputs are supported.
The video fusion device 203 comprises a single CPU, two independent display cards and a tera-network port, wherein the CPU is not lower than twenty cores, the main frequency is not lower than 2.5GHz, a storage space not lower than 48TB is provided, and RAID10 is supported; billions of pixel video data processing software are built in; the single server supports more than 16 paths of 4K video input and synchronously decodes the video; the video with the size of 1.2 hundred million pixels is spliced and fused, so that canvas roaming remote viewing is realized, real-time video is browsed by pure Web, no plug-in is required to be installed, 20 people can simultaneously view any different shooting areas, independent free zooming and translation interaction is realized, interaction time delay is lower than 300ms, and multipath video forwarding is realized. In the hundred million-level pixel array image sensor camera 1011 and the corresponding video fusion device 203, the hundred million-level pixel array image sensor camera 1011 and the corresponding video fusion device 203 provide video stream output of hundred million-level pixel fusion, the camera provides high resolution shooting input capability, and the fusion device provides splicing fusion capability; providing an SDK on the terminal, supporting canvas roaming remote viewing, supporting web terminal browsing real-time video, and no corresponding plug-in is required to be installed; the terminal is supported to access through the SDK, 20 persons are supported to watch any different shooting areas at the same time, and independent self-contained zooming and translation interaction is supported; and supporting multipath video forwarding.
The information stream sending and receiving module 204 sends information such as audio and video data processed by the audio and video processor 200 to the playing interaction terminal 300, and receives information such as audio and video data and user habit information from the playing interaction terminal 300.
The playing interaction terminal 300 comprises a user interaction module 301 and an audio/video playing module 302.
Optionally, the user interaction module 301 includes a user habit analysis device and a second audio/video collector.
The user habit analysis device is combined with user specified login state information, provides accurate content recommendation service according to user portraits, and helps a client to finish cold start of a new user; meanwhile, through continuous algorithm learning, the content preference of the user is continuously updated, the interested content is continuously recommended, and the user is better preserved.
The second audio and video collector provides an interactive video cloud service, and comprises an interactive video playing SDK, an interactive video creation subsystem, an interactive video playing control and media distribution subsystem, an interactive video supporting subsystem and the like, wherein the interactive video creation subsystem supports online creation, release and audit of the interactive video and can carry out interface customization according to the project requirement; the interactive video creation subsystem provides a one-stop interactive video content production platform for an creator, supports a user to add interactive components such as branch selection, visual angle switching, picture interaction and the like for the video through a cloud visual production tool, and enables the video to have interaction capability; providing the functions of account management, media resource management (uploading, transcoding, cataloging, auditing, publishing), convergence distribution, statistical analysis and the like of the interactive video; the creator can finish the whole processes of creating, editing, auditing, publishing projects, browsing and analyzing data of the interactive video works on the platform in one step; and the interactive video support subsystem provides functions such as interactive video statistics and the like.
Optionally, the audio/video playing module 302 includes at least one of a slow live audio/video playing unit, an interactive video playing unit, and a multi-site live video playing unit.
The slow live audio/video playing unit supports roaming remote watching of canvas and browsing of real-time video by a web end by providing SDK on the playing interactive terminal 300 without installing corresponding plug-ins; through SDK access, 20 people can watch any different shooting areas at the same time, and the independent self-contained zooming, translation interaction and multipath video forwarding are realized.
The interactive video playing unit is used for playing the interactive video by providing double-end SDKs and realizing the functions of playing, buffering, rendering, interaction and the like of the interactive video content on the end; different interaction modes including time domain interaction (such as branch selection), space domain interaction (such as visual angle switching) and event type interaction (such as picture interaction) and the like; user interaction data can be collected for statistical analysis of the interaction video creation platform and guiding the creator to re-create.
And the multi-machine-position live video playing unit supports RTC synchronous distribution of the multi-view live signals based on the Internet by pushing RTMP or SRT of the multi-view live signals based on the Internet. Because of delay uncertainty of the Internet, different visual angle signals of a live broadcast have different degrees of delay in the transmission and distribution processes, and the solution scheme should support automatic synchronous alignment of multi-visual angle signals according to SEI time code information of a source; when a client watches multi-view live broadcast, listing information of the multi-view and real-time screenshot of each view, wherein the real-time screenshot automatic refreshing interval is 10s; when the playing interaction terminal 300 watches the multi-view live broadcast, the seamless view angle is switched: in the process of switching the visual angles, live audio is continuous and can not be interrupted at all; the switched visual angle video is automatically aligned with the audio; supporting various deployment modes, such as a privately-owned deployment mode, and automatically calling the RTC live broadcast delivery capability of the third-party CDN when the delivery capability of the platform is insufficient; the system supports 4-path multi-view synchronous switching and 100-path concurrent viewing.
Optionally, the multi-type audio/video mixed broadcasting system further includes a memory, as shown in fig. 2, where the memory is at least connected to one of the audio/video acquisition and transmission module 100, the audio/video processor 200, and the playing interaction terminal 300. When the network condition is bad, the playing fluency is improved by delaying playing, the video is prompted to be loaded on the interface, and meanwhile, any one of the audio and video acquisition and transmission module 100, the audio and video processor 200 and the playing interaction terminal 300 can realize the loading by reading the audio and video in the memory, so that the time for waiting for the loading of the audio and video is shortened.
As shown in fig. 3, by adopting the technical scheme, the inventor realizes the same information flow mixed play of multiple types of audios and videos in the mobile phone APP 8, fig. 3 shows the mixed play of VR video, 4K video, naked eye 3D video and interactive video, and different types of audios and videos in the mobile client have uniform and harmonious play interfaces, so that better viewing experience is formed. Meanwhile, each type of audio and video can display the corresponding label on the interface. VR videos will show "VR 60 panorama re-recorded world" tags, 4K videos show "4K ultra-high definition experience tags", etc. When the user clicks on these tabs, the video detail page of this type will be entered for further viewing.
According to the system provided by the embodiment of the invention, the audio and video are fused by collecting the type information of the audio and video, the audio and video are displayed on the same interface in the same information stream according to the user requirement, and various different types of audios such as slow live broadcast, 4K, 8K, VR gasket videos, interactive videos and multi-position live broadcast are mixed and displayed in the unified information stream, so that the user can watch various types of audio and video in the same information stream, the watching experience is greatly improved, the watching efficiency is improved, and further, the time waiting for loading the audio and video is shortened.
A second embodiment of the present invention will be described below.
Fig. 4 shows a schematic diagram of a multi-type audio-video mixed broadcasting system as a second embodiment.
The audio and video are generated by VR360 camera 100a, 4K camera 100b, naked eye 3D video database 100c, and interactive video database 100D, respectively. The 4 paths of audios and videos are respectively sent to the audio and video processor 200 through the Internet and pushed to the smart phone 300 after being processed. The smart phone 300 corresponds to a play interactive terminal.
4 paths of videos acquired by the audio/video processor 200: VR video, 4K video pass through the camera and gather directly, bore hole 3D video, interactive video pass through data transmission interface and gather from the server. These videos have been encapsulated by the audio and video encoders of the audio and video collectors, respectively, and have different types: different codec formats and encapsulation formats, and resolution.
For example: VR video coding and packaging format is H.264, avi;4K video coding, encapsulation format VP9, avi; the naked eye 3D video is encoded and packaged in an H.265, mp4 format; the interactive video coding and packaging format is H.264, mkv.
As shown in fig. 5, the VR360 camera 100a, the 4K camera 100b, the naked eye 3D video database 100c and the interactive video database 100D respectively transmit the audio and video to the information stream receiving module 204a through respective audio and video transmitting devices, and the information stream receiving module 204a performs type analysis and tag addition by the audio and video type analysis module 201 after receiving the audio and video information.
The audio-video type analysis module 201 corresponds to an add tag module.
The video transcoding module 202 performs unified transcoding on the video in each audio/video data, and each audio/video data at least obtains two audio/video data with different resolutions, and the audio/video data with low resolution and/or high compression rate is used as audio/video data for browsing, and the audio/video data with high resolution and/or low compression rate is used as audio/video data for playing.
The video fusion module 203 fuses the plurality of audio/video data for browsing, and splices the audio/video data into one audio/video data stream as the audio/video data stream for browsing.
The information stream transmission module 204b transmits the audio/video data stream for browsing and the audio/video data for playing to the smart phone 300 via the internet in different information streams, and the information stream for transmitting the audio/video data stream for browsing is called a browsing data information stream, and the information stream for transmitting the audio/video data for playing is called a playing data information stream.
As shown in fig. 5, the smart phone 300 includes a communication module 304, an audio/video playing module 302, and a user operation identification module 303.
The communication module 304 accepts the browse data information stream and the play data information stream sent by the audio/video processor 200. The audio/video playing module 302 selects to browse the data information stream or play the data information stream for playing according to the preset condition. In this embodiment, when the audio/video playing module 302 is initially started, the browsing data information stream is automatically selected to play.
The user operation recognition module 303 recognizes an operation mode of a user, including a video moving operation and a video jumping operation, when the audio/video playing module 302 plays the browsing data information stream. In this embodiment, in order to perform a moving operation as an operation when the user slides up and down on the touch display screen, the video skip operation is an operation when the user clicks the tag.
In the state of playing the browsing data information stream, when the video moving operation is identified, the audio/video playing module 302 plays the next or previous audio/video of the browsing data information stream, and when the video skipping operation is identified, the audio/video playing module 302 skips to the corresponding playing data information stream to play according to the label corresponding to the operation.
Fig. 6 shows a flowchart of operations in the audio-video processor in the second embodiment. As shown in figure 6 of the drawings,
in step S10, the information stream accepting module 204a accepts different types of audio and video information streams sent by the VR360 camera 100a, the 4K camera 100b, the naked-eye 3D video database 100c, and the interactive video database 100D.
In step S20, the audio/video type analysis module 201 analyzes the types of the audio/video to obtain an analysis result. For example, "VR360 panorama re-records world", "4K ultra-high definition experience", "naked eye 3D cool visual feast", "interactive video interactive high definition experience", information release name, and the like are obtained.
In step S30, the audio-video type analysis module 201 analyzes, based on the analysis result,
for example, graphics of a 'VR 360 panorama re-recording world', '4K ultra-high definition experience', 'naked eye 3D cool visual feast', 'interactive video interactive high definition experience' are generated, and the graphics are respectively added to the lower sides of videos as labels. That is, the "VR360 panorama re-recorded world" tag is added to VR video, the "4K ultra-high definition experience" tag is added to 4K video, the "naked eye 3D cool vision feast" tag is added to naked eye 3D video, and the "interactive video interactive high definition experience" tag is added to interactive video. Each tag is associated with the added audio and video, and when the user clicks on the tag, the user can jump to the detail page of the video for further viewing.
In step S40, the video transcoding module 202 transcodes VR video, 4K video, naked eye 3D video, and interactive video into a unified format video, for example: the coding format is H.264, and the video with avi format is packaged. Transcoding is only for video, and the association with tags and audio remains unchanged. During transcoding, each audio/video data obtains at least three video data with different resolutions: 4K, 1080P, 720P video. Wherein 1080P is used as audio/video data for browsing, and the rest is used as audio/video data for playing. The audio and video data for playing can be further processed, such as processing of adding caption identification and the like, which can improve user experience.
In step S50, the video fusion module 203 performs fusion processing on the 4 audio/video data for browsing with 1080 resolution, and splices the audio/video data streams into one audio/video data stream for browsing. That is, the video fusion module 203 fuses the VR video, the 4K video, and the naked eye 3D video with the resolution of 1080P and the format of h.264 and avi to generate 1080P fusion video as the browsing audio/video data stream.
In step S60, the information stream transmission module 204b transmits the audio/video data stream for browsing and the audio/video data for playing to the smart phone 300 in different information streams, and the information stream transmitting the audio/video data stream for browsing is referred to as a browsing data information stream, and the information stream transmitting the audio/video data for playing is referred to as a playing data information stream. In this embodiment, there are three play data information streams, and the corresponding resolutions are 4K, 1080P, 720P videos, respectively.
Fig. 7 shows a workflow diagram in a smartphone in a second embodiment.
The smartphone 30 employs a touch screen display, and a user can operate the displayed video.
In step S110, initially, the communication module 304 in the smart phone 300 accepts the browsing data information flow sent by the information flow sending module 204 b.
In step S120, the audio/video playing module 302 plays the audio/video of the first sequence in the browsing data information stream. At the moment, a page for watching the video is displayed on a mobile phone interface opened by a user, a first video in the information stream is displayed on the page, and when the user needs to play the next video in the playing process, the user slides up and down on the touch screen.
In step S130, the user operation recognition module 303 recognizes whether the user has an operation of sliding up and down on the touch screen or an operation of clicking a tab. Step S140 is performed when the up-down slide operation is performed, and step S170 is performed as a jump operation when the tab click operation is performed.
In step S140, the user operation identification module 303 determines whether the operation of the user on the touch screen is a sliding operation, and the step S150 is entered during the sliding operation, and after the audio/video playing module 302 plays the next audio/video in the browsing data information stream, the step S130 is returned.
If it is determined that the sliding operation is not performed, the process proceeds to step S160, and after the audio/video playing module 302 plays the last audio/video in the browsing data information stream, the process returns to step S130. At the beginning, the first audio and video is played, and the last audio and video is not changed. As a modification, the method may be configured to move cyclically, and initially, play the last audio/video in the information stream.
Fig. 8 shows a movement explanatory diagram of a video when browsing the video. The user provides a downward sliding, and the videos displayed on the smart phone 300 are respectively switched according to the sequence of 'VR 360 panorama re-recording world', '4K ultra-high definition experience', 'naked eye 3D cool vision feast', 'interactive video interactive high definition experience', and when the user also slides upwards, the videos are switched reversely.
Labels of all videos are added to the lower side areas b 1-b 4 of the videos P1-P4. As a modification, the compression ratio may be changed at the same time when the transcoding is unified.
The second embodiment of the present invention is explained above.
In the prior art, a specified category, such as "VR channel" and "live channel", is typically displayed on a client side main interface, and a user views an audio/video file of the category in a corresponding channel and clicks to play the audio/video file. Because only a single type can be checked in the same information stream, various different types of audios and videos cannot be played simultaneously in one client interface, when the videos are watched, if the types are different, the information stream is not required to be switched, so that the operation is complex, the loading time of the audios and the videos is required to be waited, and the watching experience of a user is reduced.
The invention uses 1) the audio and video data with different types to carry out unified transcoding to generate audio and video data with unified video format, and corresponds to the audio and video data before transcoding, at least two audio and video data with the same video format, different resolution and/or compression ratio are provided after transcoding, the audio and video data with low resolution and/or high compression ratio are used as the audio and video data for browsing, and the audio and video data with high resolution and/or low compression ratio are used as the audio and video data for playing.
2) Each browsing audio and video data is fused into one audio and video data stream and sent to the smart phone through the information stream, and the playing audio and video data is sent to the smart phone through other information streams, so that when the smart phone plays the browsing audio and video information stream, for example, the smart phone can browse different types of original audio and video through sliding operation, and when high-quality video is required to be watched, for example, the smart phone can jump to the high-quality video by clicking a tag.
In the whole playing process, no matter the audio and video data for browsing or playing, because the audio and video formats are the same, the audio and video playing module 302 does not need to reload for different formats, and when browsing or jumping, the audio and video is not required to be delayed in time for loading data, so that the connection of the audio and video is smooth, and the watching experience of a user is improved.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim.

Claims (10)

1. A multi-type audio and video mixed broadcast system, comprising: an audio/video processor (200) and a playing interaction terminal (300),
the audio and video processor (200) comprises an information stream receiving module (204 a), an information stream sending module (204 b), an adding tag module (201), a video transcoding module (202) and a video fusion module (203);
the information flow receiving module (204 a) receives audio and video data sent by a plurality of audio and video sources, and at least one part of the audio and video data is audio and video data of different types;
the tag adding module (201) adds tags to the audio and video data, and the tags are associated with the corresponding audio and video data;
the video transcoding module (202) is used for carrying out unified transcoding on videos in all the audio and video data, each audio and video data at least obtains two audio and video data with different resolutions, the audio and video data with low resolution and/or high compression ratio are used as audio and video data for browsing, and the audio and video data with high resolution and/or low compression ratio are used as audio and video data for playing;
the video fusion module (203) is used for carrying out fusion processing on the plurality of audio and video data for browsing, and splicing the audio and video data into an audio and video data stream which is used as the audio and video data stream for browsing;
the information stream sending module (204 b) sends the browsing audio/video data stream and the playing audio/video data to the playing interaction terminal (300) in different information streams, the information stream sending the browsing audio/video data stream is a browsing data information stream, and the information stream sending the playing audio/video data is a playing data information stream;
the playing interaction terminal (300) comprises a communication module (304), an audio and video playing module (302) and a user operation identification module (303);
the communication module (304) receives the browse data information flow and the play data information flow sent by the audio/video processor (200);
the audio/video playing module (302) selects browsing data information flow or playing data information flow for playing according to preset conditions;
the user operation identification module (303) identifies the operation modes of the user when the audio/video playing module (302) plays the browsing data information stream, including video moving operation and video jumping operation,
when the video moving operation is identified, the audio/video playing module (302) plays the next or the last audio/video of the browsing data information flow,
when the video skip operation is identified, the audio and video playing module (302) skips to the corresponding playing data information stream to play according to the label corresponding to the operation.
2. The multi-type audio-video mixed broadcasting system according to claim 1, wherein,
in the recognition mode of the user operation recognition module (303), a return operation is included,
and when the user operation identification module (303) identifies a return operation in the state of playing the playing data information stream, the audio/video playing module (302) returns the browsing data information stream to play.
3. A multi-type audio/video mixed broadcasting system according to claim 2, wherein said tag adding module (201) analyzes each of said audio/video data, sets a pattern of said tag according to a type of said audio/video, and said tag has one or more.
4. A multi-type audio-video mixed broadcasting system according to claim 3, characterized in that said adding tag module (201) adds said tag on the underside of the video.
5. The system according to claim 4, wherein the video transcoding module (202) adjusts the audio/video data for browsing and the audio/video data for playing to the same resolution when performing unified transcoding on the video in each of the audio/video data.
6. The system according to any one of claims 1 to 5, wherein the video transcoding module (202) generates a plurality of audio/video data for playback with different resolutions when performing unified transcoding of the video in each of the audio/video data,
the information stream sending module (204 b) sends the audio and video data for playing with different resolutions to the playing interaction terminal (300) in different information streams.
7. The multi-type audio and video mixed broadcasting method is characterized by being executed by an audio and video processor (200) and a broadcasting interactive terminal (300), and comprises the following steps of:
an audio/video data receiving step (S10) in which the audio/video processor (200) receives audio/video data transmitted from a plurality of audio/video sources, at least some of the audio/video data being of different types;
a tag adding step (S20) in which the audio/video processor (200) adds a tag to each piece of audio/video data, and the tag is associated with the corresponding audio/video data;
a video transcoding step (S40), in which the audio/video processor (200) performs unified transcoding on the video in each audio/video data, each audio/video data at least obtains two audio/video data with different resolutions, the audio/video data with low resolution and/or high compression ratio is used as audio/video data for browsing, and the audio/video data with high resolution and/or low compression ratio is used as audio/video data for playing;
a video fusion step (S50) in which the audio/video processor (200) fuses the plurality of browsing audio/video data to form an audio/video data stream as a browsing audio/video data stream;
an information stream transmission step (S60) in which the audio/video processor (200) transmits the browsing audio/video data stream and the playing audio/video data stream to the playing interactive terminal (300) with different information streams, and the information stream for transmitting the browsing audio/video data stream is a browsing data information stream and the information stream for transmitting the playing audio/video data is a playing data information stream;
a stream receiving step (S110), in which the playback interactive terminal (300) receives the browsing data stream and the playback data stream sent by the audio/video processor (200);
an audio/video playing step (S120), wherein the playing interaction terminal (300) selects browsing data information flow or playing data information flow for playing according to preset conditions;
a user operation recognition step (S130) of recognizing, by the play interactive terminal (300), an operation mode of a user when playing the browsing data information stream, including a video moving operation and a video jumping operation,
an audio/video switching step (S130, S140, S150, S160) in which, when the video moving operation is identified, the playing interactive terminal (300) plays the next or previous audio/video in the browsing data information stream,
and an audio/video skipping step (S130, S170 and S180), wherein when the video skipping operation is identified, the playing interaction terminal (300) skips to the corresponding playing data information stream for playing according to the label corresponding to the operation.
8. The method for mixed playback of multiple types of audio and video according to claim 7, wherein,
the method also comprises a return operation identification step (S190), and when the return operation is identified by the play interactive terminal (300) in the play data information stream state, the browse data information stream is returned to play.
9. The method for mixed playback of multiple types of audio and video according to claim 8, wherein,
in the step of adding the tag (S20), the playing interaction terminal (300) analyzes each audio/video data, and sets a pattern of the tag according to the type of the audio/video, where the tag includes one or more tags.
10. The method for mixed broadcasting of multi-type audio and video according to any one of claims 7 to 9, wherein,
in the video transcoding step (S40), when the audio/video processor (200) performs unified transcoding on the video in each of the audio/video data, a plurality of audio/video data for playback with different resolutions are generated,
in the information stream sending step (S60), the audio/video processor (200) sends the audio/video data for playing with different resolutions to the playing interaction terminal (300) with different information streams.
CN202210813711.5A 2022-05-11 2022-07-11 Multi-type audio and video mixed broadcasting system and method Active CN115225928B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210532881 2022-05-11
CN2022105328816 2022-05-11

Publications (2)

Publication Number Publication Date
CN115225928A CN115225928A (en) 2022-10-21
CN115225928B true CN115225928B (en) 2023-07-25

Family

ID=83611043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210813711.5A Active CN115225928B (en) 2022-05-11 2022-07-11 Multi-type audio and video mixed broadcasting system and method

Country Status (1)

Country Link
CN (1) CN115225928B (en)

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1775959A2 (en) * 2005-10-11 2007-04-18 Broadcom Corporation Video processing system with simultaneous multiple outputs each with unique formats
CN101669361A (en) * 2007-02-16 2010-03-10 马维尔国际贸易有限公司 Methods and systems for improving low resolution and low frame rate video
CN102625150A (en) * 2012-03-16 2012-08-01 中国科学院计算技术研究所 Media playing system and method
CN103428488A (en) * 2012-04-18 2013-12-04 Vixs系统公司 Video processing system with pattern detection and method for use thereof
CN103873888A (en) * 2012-12-12 2014-06-18 深圳市快播科技有限公司 Live broadcast method of media files and live broadcast source server
CN104160408A (en) * 2011-12-29 2014-11-19 派尔高公司 Method and system for video composition
CN104410917A (en) * 2014-09-16 2015-03-11 东方有线网络有限公司 Realization method for cable interactive television cross-domain video service connection system
CN104796732A (en) * 2015-04-21 2015-07-22 无锡天脉聚源传媒科技有限公司 Audio and video editing method and device
CN105141985A (en) * 2015-07-30 2015-12-09 世纪龙信息网络有限责任公司 System and method for smoothly playing film and television stream
CN108391127A (en) * 2018-04-27 2018-08-10 腾讯科技(深圳)有限公司 Method for video coding, device, storage medium and equipment
CN108712654A (en) * 2018-05-18 2018-10-26 网宿科技股份有限公司 A kind of code-transferring method and equipment of audio/video flow
EP3410728A1 (en) * 2017-05-30 2018-12-05 Vestel Elektronik Sanayi ve Ticaret A.S. Methods and apparatus for streaming data
CN109547839A (en) * 2018-12-11 2019-03-29 网易(杭州)网络有限公司 The dynamic previewing processing method and processing device of live content, medium, electronic equipment
WO2019072096A1 (en) * 2017-10-10 2019-04-18 腾讯科技(深圳)有限公司 Interactive method, device, system and computer readable storage medium in live video streaming
CN110662086A (en) * 2019-10-31 2020-01-07 成都威爱新经济技术研究院有限公司 5G high-definition live broadcast system and video processing method
CN113573101A (en) * 2021-07-09 2021-10-29 百果园技术(新加坡)有限公司 Video encoding method, device, equipment and storage medium
WO2021236059A1 (en) * 2020-05-19 2021-11-25 Google Llc Dynamic parameter selection for quality-normalized video transcoding

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10951906B2 (en) * 2018-08-16 2021-03-16 Bombbomb, Inc. Method and system for implementing split and parallelized encoding or transcoding of audio and video content

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1949839A (en) * 2005-10-11 2007-04-18 美国博通公司 Video processing system, set-top box and method for treating audio/video information
EP1775959A2 (en) * 2005-10-11 2007-04-18 Broadcom Corporation Video processing system with simultaneous multiple outputs each with unique formats
CN101669361A (en) * 2007-02-16 2010-03-10 马维尔国际贸易有限公司 Methods and systems for improving low resolution and low frame rate video
CN104160408A (en) * 2011-12-29 2014-11-19 派尔高公司 Method and system for video composition
CN102625150A (en) * 2012-03-16 2012-08-01 中国科学院计算技术研究所 Media playing system and method
CN103428488A (en) * 2012-04-18 2013-12-04 Vixs系统公司 Video processing system with pattern detection and method for use thereof
CN103873888A (en) * 2012-12-12 2014-06-18 深圳市快播科技有限公司 Live broadcast method of media files and live broadcast source server
CN104410917A (en) * 2014-09-16 2015-03-11 东方有线网络有限公司 Realization method for cable interactive television cross-domain video service connection system
CN104796732A (en) * 2015-04-21 2015-07-22 无锡天脉聚源传媒科技有限公司 Audio and video editing method and device
CN105141985A (en) * 2015-07-30 2015-12-09 世纪龙信息网络有限责任公司 System and method for smoothly playing film and television stream
EP3410728A1 (en) * 2017-05-30 2018-12-05 Vestel Elektronik Sanayi ve Ticaret A.S. Methods and apparatus for streaming data
CN110710220A (en) * 2017-05-30 2020-01-17 韦斯特尔电子工业和贸易有限责任公司 Method and apparatus for streaming data
CN109660853A (en) * 2017-10-10 2019-04-19 腾讯科技(北京)有限公司 Interactive approach, apparatus and system in net cast
WO2019072096A1 (en) * 2017-10-10 2019-04-18 腾讯科技(深圳)有限公司 Interactive method, device, system and computer readable storage medium in live video streaming
CN108391127A (en) * 2018-04-27 2018-08-10 腾讯科技(深圳)有限公司 Method for video coding, device, storage medium and equipment
CN108712654A (en) * 2018-05-18 2018-10-26 网宿科技股份有限公司 A kind of code-transferring method and equipment of audio/video flow
CN109547839A (en) * 2018-12-11 2019-03-29 网易(杭州)网络有限公司 The dynamic previewing processing method and processing device of live content, medium, electronic equipment
CN110662086A (en) * 2019-10-31 2020-01-07 成都威爱新经济技术研究院有限公司 5G high-definition live broadcast system and video processing method
WO2021236059A1 (en) * 2020-05-19 2021-11-25 Google Llc Dynamic parameter selection for quality-normalized video transcoding
CN113573101A (en) * 2021-07-09 2021-10-29 百果园技术(新加坡)有限公司 Video encoding method, device, equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于Hi3515C的车载驾驶音视频监控系统设计;王灵芝;李伟滨;;闽南师范大学学报(自然科学版)(第02期);36-42 *

Also Published As

Publication number Publication date
CN115225928A (en) 2022-10-21

Similar Documents

Publication Publication Date Title
US11073969B2 (en) Multiple-mode system and method for providing user selectable video content
Niamut et al. MPEG DASH SRD: spatial relationship description
CA2466924C (en) Real time interactive video system
US9870801B2 (en) Systems and methods for encoding multimedia content
US9271048B2 (en) Systems and methods for immersive viewing experience
US10623816B2 (en) Method and apparatus for extracting video from high resolution video
US7644425B2 (en) Picture-in-picture mosaic
US10250838B1 (en) System and method for converting live action alpha-numeric text to re-rendered and embedded pixel information for video overlay
US20140219634A1 (en) Video preview creation based on environment
US20080101456A1 (en) Method for insertion and overlay of media content upon an underlying visual media
US20100272187A1 (en) Efficient video skimmer
AU2020201003A1 (en) Selective capture and presentation of native image portions
US8650591B2 (en) Video enabled digital devices for embedding user data in interactive applications
CN101917568A (en) Streaming media technology-based network video player and playing system
US20150350565A1 (en) Techniques for magnifying a high resolution image
US20180091860A1 (en) Enhancing A Media Recording Comprising A Camera Recording
JP2020524450A (en) Transmission system for multi-channel video, control method thereof, multi-channel video reproduction method and device thereof
CN115225928B (en) Multi-type audio and video mixed broadcasting system and method
US10764655B2 (en) Main and immersive video coordination system and method
CN115580738B (en) High-resolution video display method, device and system for on-demand transmission
CN114866841A (en) Ultra-high-definition multi-picture display system and method with echoing function
KR101452902B1 (en) Broadcasting receiver and controlling method thereof
CN117291810B (en) Video frame processing method, device, equipment and storage medium
KR101827967B1 (en) Server and Service for Providing Video Content
CN117939183A (en) Multi-machine-position free view angle guided broadcasting method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant