CN115225928A - Multi-type audio and video mixed broadcasting system and method - Google Patents

Multi-type audio and video mixed broadcasting system and method Download PDF

Info

Publication number
CN115225928A
CN115225928A CN202210813711.5A CN202210813711A CN115225928A CN 115225928 A CN115225928 A CN 115225928A CN 202210813711 A CN202210813711 A CN 202210813711A CN 115225928 A CN115225928 A CN 115225928A
Authority
CN
China
Prior art keywords
audio
video
playing
video data
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210813711.5A
Other languages
Chinese (zh)
Other versions
CN115225928B (en
Inventor
陈祥
毕江
芮浩
赵涛
冯澈
李浩楠
欧晓霄
宋峥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Radio And Television Station
Original Assignee
Beijing Radio And Television Station
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Radio And Television Station filed Critical Beijing Radio And Television Station
Publication of CN115225928A publication Critical patent/CN115225928A/en
Application granted granted Critical
Publication of CN115225928B publication Critical patent/CN115225928B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234309Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4 or from Quicktime to Realvideo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234363Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234381Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the temporal resolution, e.g. decreasing the frame rate by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2662Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440263Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the spatial resolution, e.g. for displaying on a connected PDA
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/65Transmission of management data between client and server
    • H04N21/658Transmission by the client directed to the server
    • H04N21/6587Control parameters, e.g. trick play commands, viewpoint selection

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention relates to a multi-type audio and video mixed broadcasting system and a method thereof, wherein the system comprises: the system comprises an audio and video acquisition and transmission module, an audio and video processor and a playing interactive terminal, wherein the audio and video acquisition and transmission module is respectively in signal transmission connection with the audio and video processor, and the playing interactive terminal is in signal transmission connection with the audio and video processor; the audio and video acquisition and transmission module comprises a first audio and video acquisition device, an audio and video encoder and an audio and video transmission device; the first audio and video collector comprises a hundred million-level pixel array pixel sensor camera; the audio and video processor comprises an audio and video type analysis module, a video transcoding module, a video fusion device and an information stream sending and receiving module; the playing interactive terminal comprises the user interactive module and the audio and video playing module, and can be used for mixedly displaying different types of audio in a unified information stream, so that a user can watch various types of audio and video in the same information stream.

Description

Multi-type audio and video mixed broadcasting system and method
The application claims priority of Chinese patent application entitled "a multi-type audio and video mixed broadcasting system" filed by the Chinese intellectual property office of China at 11/05/2022 and having an application number of 202210532881.6, the entire contents of which are described in the application.
Technical Field
The invention relates to the technical field of audio and video playing, in particular to a multi-type audio and video mixed playing system.
Background
By installing the playing client capable of playing the audio and video in the terminals such as the electronic equipment, the user can enjoy the audio and video through the client. In the related art, a main interface of the client may display a designated category, such as "VR channel" and "live channel", and a user may view an audio file of the type in a corresponding channel and click to play the audio file. However, in the related art, it is difficult to mix and display various different types of audio such as slow live broadcast, 4K, 8K, VR pad video, interactive video, multi-camera live broadcast in a unified information stream, and a user can only view a single type of audio and video in the same information stream, and cannot play various different types of audio and video in one playing client interface at the same time, which affects the user to obtain comprehensive and rich viewing experience.
Disclosure of Invention
In order to overcome the defects that different types of audios and videos cannot be mixed and displayed in the same information stream to be played in the prior art and the like, and at least partially achieve the purpose of playing the multiple types of audios and videos in the same information stream, the invention provides a multi-type audio and video mixed playing system which comprises: the system comprises an audio and video acquisition and transmission module, an audio and video processor and a playing interactive terminal, wherein the audio and video acquisition and transmission module is respectively in signal transmission connection with the audio and video processor, and the playing interactive terminal is in signal transmission connection with the audio and video processor; the audio and video acquisition and transmission module comprises a first audio and video acquisition device, an audio and video encoder and an audio and video transmission device; the first audio and video collector comprises a hundred million-level pixel array pixel sensor camera; the audio and video processor comprises an audio and video type analysis module, a video transcoding module, a video fusion device and an information stream sending and receiving module; the playing interactive terminal comprises a user interactive module and an audio and video playing module.
Optionally, the audio/video encoder includes a 4K encoder and/or an 8K encoder.
Optionally, the audio/video encoder further includes at least one of a standard definition encoder, a high definition encoder, and a 2K encoder.
Optionally, the user interaction module includes a user habit analyzing device and a second audio/video collector.
Optionally, the audio/video playing module at least includes one of a slow live audio/video playing unit, an interactive video playing unit, and a multi-machine live video playing unit.
Optionally, the multi-type audio/video mixed playing system further includes a memory, and the memory is connected with at least one of the audio/video acquisition and transmission module, the audio/video processor, and the playing interactive terminal.
The technical scheme provided by the embodiment of the invention has the beneficial effects that at least:
the audio and video are fused by collecting the type information of the audio and video, the audio and video are displayed on the same interface by the same information stream according to the user requirements, and various different types of audio such as slow live broadcast, 4K, 8K, VR gasket video, interactive video, multi-machine live broadcast and the like are displayed in a unified information stream in a mixed mode, so that the user can watch various types of audio and video in the same information stream, the watching experience is greatly improved, the watching efficiency is improved, and further, the time for waiting for audio and video loading is shortened.
As a first technical solution of the present invention, a multi-type audio/video mixed playing system includes: an audio and video processor 200, a playing interactive terminal 300,
the audio and video processor 200 comprises an information stream receiving module 204a, an information stream sending module 204b, a tag adding module 201, a video transcoding module 202 and a video fusion module 203.
The information stream receiving module 204a receives audio and video data sent by a plurality of audio and video sources, wherein at least one part of the audio and video data is different types of audio and video data.
The tag adding module 201 adds tags to the audio and video data, and the tags are associated with the corresponding audio and video data.
The video transcoding module 202 performs unified transcoding processing on videos in the audio and video data, where each audio and video data at least obtains two audio and video data with different resolutions, the audio and video data with low resolution and/or high compression ratio is used as audio and video data for browsing, and the audio and video data with high resolution and/or high compression ratio is used as audio and video data for playing.
And the video fusion module 203 is used for fusing a plurality of audio and video data for browsing to splice an audio and video data stream as the audio and video data stream for browsing.
The information stream sending module 204b sends the audio-video data stream for browsing and the audio-video data for playing to the playing interactive terminal 300 in different information streams, sends the information stream of the audio-video data stream for browsing as a browsing data information stream, and sends the information stream of the audio-video data for playing as a playing data information stream.
The playing interactive terminal 300 includes a communication module 304, an audio/video playing module 302, and a user operation identification module 303.
The communication module 304 receives the browsing data information stream and the playing data information stream sent by the audio/video processor 200.
The audio/video playing module 302 selects to browse the data information stream or play the data information stream for playing according to a preset condition.
The user operation identification module 303 identifies an operation mode of a user when the audio/video playing module 302 plays the browsing data information stream, including a video moving operation and a video skipping operation, when the video moving operation is identified, the audio/video playing module 302 plays a next or previous audio/video of the browsing data information stream, and when the video skipping operation is identified, the audio/video playing module 302 skips to a corresponding playing data information stream for playing according to the tag corresponding to the operation.
Because 1) different types of audio and video data are transcoded in a unified manner to generate audio and video data with unified video formats, and the audio and video data with the same video formats and different resolutions and/or compression ratios correspond to the audio and video data before being transcoded, the audio and video data with the low resolution and/or high compression ratio are at least provided with the audio and video data with the same video formats and different resolutions and/or compression ratios after being transcoded, the audio and video data with the high resolution and/or compression ratios are taken as the audio and video data for playing, 2) the audio and video data for browsing are fused into one audio and video data stream and are sent to the playing interactive terminal 300 through information streams, and the audio and video data for playing are sent to the playing interactive terminal 300 through other information streams, when the audio and video information streams are played, for example, different types of original audio and video can be browsed through sliding operations, and when high-quality video needs to be watched, for example, a tag is clicked to jump to high-quality video.
In the whole playing process, no matter the audio and video data for browsing or the audio and video data for playing are in the same audio and video format, the audio and video playing module 302 does not need to be reloaded for different formats, and time delay does not occur for loading data when audio and video are browsed or jumped, so that the connection of the audio and video is smooth, and the watching experience of a user is improved.
Preferably, the identification mode of the user operation identification module 303 includes a return operation,
in the state of playing the data information stream, when the user operation identification module 303 identifies a return operation, the audio/video playing module 302 returns the browsing data information stream to play.
Therefore, the user can return to the audio and video for browsing when needed, and the random switching between the browsing of the video and the playing of the video is realized.
Preferably, the tag adding module 201 analyzes each audio/video data, and sets one or more tags according to the type of the audio/video.
Therefore, the user can confirm the type of the audio and video according to the pattern of the tag.
Preferably, the tag adding module 201 adds the tag to the lower side of the video.
Therefore, the tag is not only easy to operate, but also does not obstruct the user's video viewing.
Preferably, the video transcoding module 202 adjusts each of the browsing audio/video data and the playing audio/video data to the same resolution when performing the unified transcoding processing on the video in each of the audio/video data.
Therefore, when browsing or watching, no sense of incongruity is generated due to switching of videos.
Preferably, the video transcoding module 202 generates a plurality of audio/video data for playing with different resolutions when performing a unified transcoding process on a video in each of the audio/video data, and the information stream sending module 204b sends the audio/video data for playing with different resolutions to the playing interactive terminal 300 in different information streams.
Therefore, at the playing interactive terminal 300, the user can select a suitable resolution to play according to actual needs, such as processing capability, network speed, and the like, so as to avoid problems of jamming and the like.
As a second technical solution of the present invention, a multi-type audio/video mixed playing method, executed by an audio/video processor 200 and a playing interactive terminal 300, includes the following steps: in the audio/video data receiving step S10, the audio/video processor 200 receives audio/video data sent by a plurality of audio/video sources, and at least a part of the audio/video data is different types of audio/video data.
And a tag adding step S20, in which the audio/video processor 200 adds a tag to each piece of the audio/video data, and the tag is associated with the corresponding audio/video data.
And a video transcoding step S40, in which the audio/video processor 200 performs unified transcoding processing on videos in the audio/video data, each audio/video data at least obtains two audio/video data with different resolutions, the audio/video data with low resolution and/or high compression rate is used as audio/video data for browsing, and the audio/video data with high resolution and/or high compression rate is used as audio/video data for playing.
And a video fusion step S50, wherein the audio and video processor 200 performs fusion processing on the plurality of browsing audio and video data to splice into an audio and video data stream as the browsing audio and video data stream.
An information stream sending step S60, where the audio/video processor 200 sends the browsing audio/video data stream and the playing audio/video data to the playing interactive terminal 300 in different information streams, sends the information stream of the browsing audio/video data stream as a browsing data information stream, and sends the information stream of the playing audio/video data as a playing data information stream.
In the information stream receiving step S110, the playing interactive terminal 300 receives the browsing data information stream and the playing data information stream sent by the audio/video processor 200.
In the audio/video playing step S120, the playing interactive terminal 300 selects to browse the data information stream or play the data information stream for playing according to a preset condition.
The method comprises a user operation identification step S130, wherein the playing interactive terminal 300 identifies the operation mode of a user when the browsing data information stream is played, the operation mode comprises video moving operation and video skipping operation, audio and video switching steps S130, S140, S150 and S160 are carried out, when the video moving operation is identified, the playing interactive terminal 300 plays the next or previous audio and video in the browsing data information stream, the audio and video skipping steps S130, S170 and S180 are carried out, and when the playing interactive terminal 300 identifies the video skipping operation, the playing interactive terminal jumps to the corresponding playing data information stream to play according to the corresponding tag in operation.
Preferably, the method further includes a step S190 of returning operation identification, where in the state of playing the data information stream, when the interactive playing terminal 300 identifies a return operation, the interactive playing terminal returns the browsing data information stream to play.
Preferably, in the step S20 of adding tags, the playing interactive terminal 300 analyzes each audio/video data, and sets a pattern of the tag according to the type of the audio/video, where the tag includes one or more tags.
Preferably, in the video transcoding step S40, when the audio/video processor 200 performs a unified transcoding process on the video in each of the audio/video data, a plurality of audio/video data for playing with different resolutions are generated, in the video fusion step S50, when the video in each of the audio/video data is subjected to a unified transcoding process, a plurality of audio/video data for playing with different resolutions are generated, and in the information stream sending step S60, the audio/video processor 200 sends the audio/video data for playing with different resolutions to the playing interactive terminal 300 in different information streams.
Drawings
Fig. 1 is a schematic structural diagram of a multi-type audio/video mixed broadcast system in an embodiment of the present invention;
fig. 2 is a schematic diagram of a memory in a multi-type audio/video mixed playing system according to an embodiment of the present invention;
fig. 3 is a screenshot of a mixed playing interface of multiple types of audio and video on a playing interactive terminal in the embodiment of the present invention;
fig. 4 shows a structural diagram of a multi-type audio-video mixed playing system as a second embodiment;
fig. 5 shows a structural style diagram of a multi-type audio-video mixed playing system of a second embodiment;
fig. 6 shows a flowchart of the work in the av processor in the second embodiment;
fig. 7 shows a flowchart of the work in the smartphone in the second embodiment;
fig. 8 is an explanatory diagram showing movement of a video when browsing the video in the second embodiment.
Detailed Description
In the following detailed description of the preferred embodiments of the invention, reference is made to the components and techniques in order that the advantages and features of the invention may be readily understood, and readily appreciated, when appropriately set forth. The following description is an embodiment of the claimed invention, and other embodiments related to the claims not specifically described also fall within the scope of the claims. Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail but are intended to be part of the specification where appropriate.
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
In order to solve the problems in the prior art, embodiments of the present invention provide a multi-type audio and video mixed playing system.
The invention provides a multi-type audio and video mixed playing system, which is structurally shown in fig. 1, and comprises an audio and video acquisition and transmission module 100, an audio and video processor 200 and a playing interactive terminal 300, wherein the audio and video acquisition and transmission module 100 is respectively in signal transmission connection with the audio and video processor 200, and the playing interactive terminal 300 is in signal transmission connection with the audio and video processor 200;
the audio and video acquisition and transmission module 100 comprises a first audio and video acquisition device 101, an audio and video encoder 102 and an audio and video transmission device 103;
the first audio/video collector 101 can collect live audio/video from live broadcast devices including a digital camera, a smart phone, and the like through an audio/video transmission interface, and can also collect audio/video from storage devices such as a server, a memory, and the like through a data transmission interface or a network transmission interface.
The first audio video collector 101 comprises a hundred million level pixel array image sensor camera 1011. The resolution of output video of each lens reaches 4K (3840 x 2160), the total horizontal field angle is 40 degrees, the total vertical field angle is 13 degrees, the frame rate is not lower than 25fps, the total code rate is 200-300 Mbps adjustable, the client side mode supports standard resolutions of 8K, 4K, 1080P, 720P and the like, supports live broadcast with any resolution below 8K, the interaction delay is less than 300ms, real-time screenshot can be realized, and a single picture reaches 1.2 hundred million pixels.
Optionally, the audio/video encoder 102 includes a 4K encoder and/or an 8K encoder.
Optionally, the audio/video encoder 102 further includes at least one of a standard definition encoder, a high definition encoder, and a 2K encoder.
And the audio and video transmitting device 103 transmits the encoded audio and video to the audio and video processor 200 in various transmission forms. When the transmission of the audio/video processor 200 is abnormal, the encoded audio/video is transmitted to the memory, and the audio/video transmission information and the memory address and other information are transmitted to the audio/video processor 200.
The audio and video processor 200 comprises an audio and video type analysis module 201, a video transcoding module 202, a video fusion device 203 and an information stream sending and receiving module 204;
the audio and video type analysis module 201 will analyze the audio and video type and add corresponding tags. The user can click the playing icon or the label information to jump to the full screen to further view the details of the video, and the technical label information of the video, such as 8K, HDR and frame rate, is displayed. Labels include, but are not limited to: 8K, 4K, VR, slow live broadcast, multi-station live broadcast and interactive video.
The video transcoding module 202 performs intelligent transcoding processing on the input video file to realize that the input formats supported by decoding include, but are not limited to, h.264, h.265 and VP9, the output formats supported by encoding include, but are not limited to, h.264, h.265 and VP9, and the supported video packaging formats include, but are not limited to, mp4, mov, mkv and avi; the coding rate of the output video is determined in a self-adaptive manner through the code rate control of the content complexity in a self-adaptive manner according to the texture complexity, the motion complexity and the pre-coding result of the input video; by encoding the region of interest, the code rate control based on the scene is realized; the video quality is improved through the video preprocessing based on the deep learning, and the video can be subjected to self-adaptive picture restoration processing, including sharpening, noise reduction, compression distortion removal and the like; including multi-resolution output including co-resolution output, up-sampling and down-sampling. Specifically, in the aspect of 8K transcoding, transcoding covering 8K, 4K, 1080P and 720P resolutions is supported, 8K input content can be transcoded at the highest, a multi-resolution file is output at the same time, and 8K encoding and decoding capabilities covering h.264, hevc and AV1 formats are supported; supporting HLS format output of 8K resolution content; the input and processing of 8K ultra-high definition videos are supported, and the frame rate input of 24fps, 25fps, 30fps,50fps and 60fps is supported; support SDR, HDR input; mp4, mov, mkv, avi video format input is supported.
The video fusion device 203 comprises a single CPU, two independent display cards and ten-gigabit network ports, wherein the CPU is not lower than twenty cores, the main frequency is not lower than 2.5GHz, a storage space not lower than 48TB is provided, and RAID10 is supported; video data processing software of hundred million-level pixels is arranged in the video data processing device; the single server supports more than 16 paths of 4K video input and synchronously decodes; the video is integrated into the video with more than 1.2 hundred million pixels through splicing, the roaming remote watching of canvas is realized, the real-time video is browsed through pure Web, plug-ins are not required to be installed, 20 people can watch any different shooting areas simultaneously, independent free zooming and translation interaction are realized, the interaction time delay is lower than 300ms, and multi-path video forwarding is realized. In the hundred million-level pixel array sensor camera 1011 and the corresponding video fusion device 203, the hundred million-level pixel array sensor camera 1011 and the corresponding video fusion device 203 provide video stream output of hundred million-level pixel fusion, the camera provides high-resolution shooting input capability, and the fusion device provides splicing fusion capability; the terminal provides the SDK, supports canvas roaming remote watching, supports the web terminal to browse the real-time video, and does not need to install a corresponding plug-in; the terminal is supported to access through the SDK, 20 people are supported to watch any different shooting areas at the same time, and independent zooming and translation interaction is supported; multiple video forwarding is supported.
The information stream sending and receiving module 204 sends the information such as the audio and video data processed by the audio and video processor 200 to the playing interactive terminal 300, and receives the information such as the audio and video data and the user habit information from the playing interactive terminal 300.
The playing interactive terminal 300 includes a user interactive module 301 and an audio/video playing module 302.
Optionally, the user interaction module 301 includes a user habit analyzing device and a second audio/video collector.
The user habit analyzing device is used for providing accurate content recommendation service according to the user portrait by combining the login state information specified by the user, and helping the client to finish cold start of a new user; meanwhile, through continuous algorithm learning, the content preference of the user is continuously updated, the interested content is continuously recommended, and the user is better reserved.
The second audio and video collector is used for providing interactive video cloud service and comprises an interactive video playing SDK, an interactive video creation subsystem, an interactive video playing control and media distribution subsystem, an interactive video supporting subsystem and the like, wherein the interactive video creation subsystem supports online creation, release and audit of interactive videos and can perform interface customization according to the project requirement; the interactive video creation subsystem provides a one-stop interactive video content production platform for creators, supports users to add interactive components such as branch selection, visual angle switching and picture interaction for videos through a cloud visual manufacturing tool, and enables the videos to have interactive capacity; the functions of account management, media asset management (uploading, transcoding, cataloging, auditing and releasing), convergence and distribution, statistical analysis and the like of the interactive video are provided; the creator can finish the whole processes of creation, editing, auditing, project release, data browsing and analysis and the like of the interactive video works on the platform in a one-stop manner; and the interactive video support subsystem provides functions of interactive video statistics and the like.
Optionally, the audio/video playing module 302 at least includes one of a slow live audio/video playing unit, an interactive video playing unit, and a multi-camera live video playing unit.
The slow live video and audio playing unit supports canvas roaming remote watching and web end browsing of real-time video by providing the SDK on the playing interactive terminal 300 without installing a corresponding plug-in; through SDK access, 20 people can watch any different shooting areas simultaneously, and independent zooming, translation interaction and multi-channel video forwarding are achieved.
The interactive video playing unit is used for realizing interactive video playing by providing a double-end SDK and realizing the functions of playing, buffering, rendering, interacting and the like of interactive video contents on the end; different interaction modes such as time domain interaction (such as branch selection), space domain interaction (such as view angle switching) and event type interaction (such as picture interaction) are included; user interaction data can be collected for the interactive video creation platform to perform statistical analysis and guide creators to create again.
The multi-machine-position live broadcast video playing unit supports RTC synchronous distribution of multi-view live broadcast signals based on the Internet based on the pushing of RTMP or SRT of the multi-view live broadcast signals of the Internet. Due to the delay uncertainty of the internet, different visual angle signals of one live broadcast have different degrees of delay in the transmission and distribution processes, and the solution should support the automatic synchronous alignment of the multi-visual angle signals according to the SEI time code information of the information source; when a client watches multi-view live broadcasting, listing multi-view information and real-time screenshots of all views, wherein the automatic refreshing interval of the real-time screenshots is 10s; when watching multi-view live broadcasting, the broadcast interactive terminal 300 seamlessly switches the view: in the process of switching the visual angle, the live audio is continuous and can not be interrupted; the switched visual angle video should be automatically aligned with the audio; the method supports various deployment modes, for example, a privatized deployment mode is adopted, and when the distribution capability of a platform is insufficient, the RTC live broadcast distribution capability of a third-party CDN can be automatically called; the system supports 4-channel multi-view synchronous switching and 100-channel concurrent viewing.
Optionally, the multi-type audio/video mixed playing system further includes a memory, as shown in fig. 2, the memory is connected to at least one of the audio/video acquisition and transmission module 100, the audio/video processor 200, and the playing interactive terminal 300. When the network condition is not good, the playing fluency is improved by delaying playing, the video is prompted to be loaded on the interface, meanwhile, any one of the audio and video acquisition and transmission module 100, the audio and video processor 200 and the playing interactive terminal 300 can realize loading by reading the audio and video in the memory, and the time for waiting for audio and video loading is shortened.
As shown in fig. 3, by adopting the technical scheme, the inventor realizes the mixed playing of the same information stream of multiple types of audios and videos in a mobile phone APP jing 8, fig. 3 shows the mixed playing of VR videos, 4K videos, naked eye 3D videos and interactive videos, and the different types of audios and videos in the mobile client have uniform and harmonious playing interfaces, so that better watching experience is formed. Meanwhile, each type of audio and video can display the corresponding label on the interface. The VR video will show a "VR 60 panorama rerecording world" tag, a 4K video show a "4K ultra high definition experience tag", and so on. When the user clicks on these tabs, the type of video detail page will be entered for further viewing.
According to the system provided by the embodiment of the invention, the audio and the video are fused by collecting the type information of the audio and the video, and the audio and the video are displayed on the same interface in the same information stream according to the requirements of users, so that various different types of audio such as slow live broadcast, 4K, 8K, VR gasket video, interactive video, multi-machine live broadcast and the like are mixed and displayed in the unified information stream, and the users can watch various types of audio and video in the same information stream, so that the watching experience is greatly improved, the watching efficiency is improved, and further, the time for waiting for loading the audio and the video is shortened.
The second embodiment of the present invention will be explained below.
Fig. 4 shows a structural diagram of a multi-type audio-video mixing system as a second embodiment.
The audio and video are generated by the VR360 camera 100a, the 4K camera 100b, the naked eye 3D video database 100c, and the interactive video database 100D, respectively. The 4 channels of audio and video are respectively sent to the audio and video processor 200 through the internet, and are pushed to the smart phone 300 after being processed. The smart phone 300 corresponds to a play interactive terminal.
The 4 channels of videos acquired by the audio/video processor 200: VR video, 4K video pass through the camera direct acquisition, bore hole 3D video, interactive video pass through the data transmission interface and gather from the server. These videos have been encoded and encapsulated by the respective audio-video encoders of the audio-video collector 100, and have different types: different codec formats and packing formats, and resolution.
For example: the VR video coding and packaging format is H.264, avi;4K video encoding, encapsulation format VP9, avi; the naked eye 3D video coding and packaging format is H.265, mp4; the interactive video coding and packaging format is H.264, mkv.
As shown in fig. 5, the VR360 camera 100a, the 4K camera 100b, the naked-eye 3D video database 100c, and the interactive video database 100D respectively send the audio and video to the information stream receiving module 204a through respective audio and video sending devices, and the information stream receiving module 204a performs type analysis and tag addition through the audio and video type analyzing module 201 after receiving the audio and video.
The audio-video type analysis module 201 corresponds to a tag adding module.
The video transcoding module 202 performs unified transcoding processing on videos in each piece of audio/video data, where each piece of audio/video data at least obtains two pieces of audio/video data with different resolutions, the audio/video data with low resolution and/or high compression ratio is used as browsing audio/video data, and the audio/video data with high resolution and/or high compression ratio is used as playing audio/video data.
And the video fusion module 203 is used for fusing a plurality of browsing audio and video data to splice into an audio and video data stream to be used as the browsing audio and video data stream.
The information stream sending module 204b sends the audio-video data stream for browsing and the audio-video data for playing to the smart phone 300 through the internet using different information streams, where the information stream sending the audio-video data stream for browsing is referred to as a browsing data information stream, and the information stream sending the audio-video data for playing is referred to as a playing data information stream.
As shown in fig. 5, the smartphone 300 includes a communication module 304, an audio/video playing module 302, and a user operation identification module 303.
The communication module 304 receives the browsing data information stream and the playing data information stream sent by the audio/video processor 200. The audio/video playing module 302 selects to browse the data information stream or play the data information stream for playing according to a preset condition. In this embodiment, when the audio/video playing module 302 is initially started, the browsing data information stream is automatically selected to be played.
The user operation recognition module 303 recognizes an operation mode of the user when the audio/video playing module 302 plays the browsing data information stream, including a video moving operation and a video skipping operation. In this embodiment, the moving operation is an operation performed when the user slides up and down on the touch display screen, and the video skipping operation is an operation performed when the user clicks a tag.
When the video moving operation is identified in the state of playing the browsing data information stream, the audio/video playing module 302 plays the next or previous audio/video of the browsing data information stream, and when the video skipping operation is identified, the audio/video playing module 302 skips to the corresponding playing data information stream for playing according to the tag corresponding to the operation.
Fig. 6 shows a flowchart of the work in the av processor in the second embodiment. As shown in figure 6 of the drawings,
in step S10, the information stream receiving module 204a receives different types of audio and video information streams sent by the VR360 camera 100a, the 4K camera 100b, the naked-eye 3D video database 100c and the interactive video database 100D.
In step S20, the audio/video type analyzing module 201 analyzes the type of each audio/video to obtain an analysis result. For example, the user obtains 'VR 360 panorama rerecording world', '4K ultrahigh definition experience', 'naked eye 3D cool vision feast', 'interactive video interaction high definition experience', the name of an information publisher and the like.
In step S30, the audio/video type analysis module 201 generates, for example, graphics of "VR360 panorama re-recording world", "4K ultra high definition experience", "naked eye 3D cool vision feast", "interactive video interactive high definition experience", and the images are respectively added to the lower sides of the videos as tags according to the analysis result. That is, a "VR360 panorama re-recording world" tag is added to a VR video, a "4K ultra high definition experience" tag is added to a 4K video, a "naked eye 3D cool vision feast" tag is added to a naked eye 3D video,
an "interactive video interactive high definition experience" tag is added to the interactive video. And each tag is associated with the added audio and video, and when a user clicks the tags, the user can jump to a detail page of the video for further watching.
In step S40, the video transcoding module 202 transcodes the VR video, the 4K video, the naked-eye 3D video, and the interactive video into videos with a uniform format, for example: the encoding format is H.264, and the video with the format of avi is packaged. Transcoding is only for video, and the association with both the tag and the audio remains unchanged. During transcoding, each audio/video data at least obtains three video data with different resolutions: 4K, 1080P, 720P video. 1080P is used as audio and video data for browsing, and the rest is used as audio and video data for playing. The audio and video data for playing can be further processed, such as the processing of adding caption identification and the like, which can improve the user experience.
In step S50, the video fusion module 203 fuses 4 pieces of browsing audio and video data with a resolution of 1080, and splices the data into an audio and video data stream serving as the audio and video data stream for browsing. That is, the video fusion module 203 fuses the VR video with a resolution of 1080P and a format of h.264, avi, 4K video, and naked-eye 3D video to generate 1080P fused video as an audio-video data stream for browsing.
In step S60, the information stream sending module 303 sends the audio-video data stream for browsing and the audio-video data for playing to the smart phone 300 in different information streams, where the information stream sending the audio-video data stream for browsing is referred to as a browsing data information stream, and the information stream sending the audio-video data for playing is referred to as a playing data information stream. In this embodiment, there are three playing data streams, and the corresponding resolutions are 4K, 1080P and 720P videos, respectively.
Fig. 7 shows a flowchart of the work in the smartphone in the second embodiment.
The smartphone 30 uses a touch screen for display, and a user can operate the displayed video.
In step S110, initially, the communication module 304 in the smart phone 300 receives the browsing data information stream sent by the information stream sending module 204 b.
In step S120, the audio/video playing module 302 plays the audio/video with the first order in the browsing data information stream. And displaying a page for watching the video on a mobile phone interface opened by a user, displaying a first video in the information stream on the page, and sliding up and down on the touch screen when the user needs to play a next video in the playing process.
In step S130, the user operation recognition module 303 recognizes whether the user has an operation of sliding up and down or an operation of clicking a tab on the touch screen. If the up-down slide operation is performed, the process proceeds to step S140, and if the click operation is performed, the process proceeds to step S170 as a jump operation.
In step S140, the user operation identification module 303 determines whether the operation of the user on the touch screen is a downslide operation, the step S150 is performed during the downslide operation, and the audio/video playing module 302 returns to step S130 after playing the next audio/video in the browsing data information stream.
If it is determined that the audio/video playing operation is not the downslide operation, the process proceeds to step S160, and the audio/video playing module 302 returns to step S130 after playing the previous audio/video in the browsing data information stream. At the beginning, the first audio and video is played, and the last audio and video is not played. As a variant, it may be set to move cyclically, and initially, the last audio/video in the information stream is played.
Fig. 8 is a diagram illustrating movement of a video when browsing the video. The user provides the lapse, and the video that shows on the smart mobile phone 300 switches according to the order of "360 panorama rerecording worlds", "4K superelevation experiences", "bore hole 3D dazzles cool vision feast", "interactive video high definition experiences", and when the same lapse, the video reverse switch.
The labels of the respective videos are added to the lower regions b1 to b4 of the videos P1 to P4. As a variation, the compression rate may also be changed simultaneously during the unified transcoding.
The second embodiment of the present invention has been explained above.
In the prior art, generally, a specified category, such as "VR channel" or "live channel", is displayed on a client main interface, and a user views an audio/video file of the type in a corresponding channel and clicks to play the audio/video file. Because only a single type can be checked in the same information stream, different types of audios and videos cannot be played in one client interface at the same time, when the videos are watched, if the types are different, the information streams need not to be switched, so that the operation is complex, the loading time of the audios and videos also needs to be waited, and the watching experience of a user is reduced.
The invention carries out unified transcoding on different types of audio and video data to generate audio and video data with unified video format, and at least two audio and video data with the same video format and different resolution and/or compression ratio after transcoding are corresponding to the audio and video data before transcoding, the audio and video data with low resolution and/or high compression ratio is taken as the audio and video data for browsing, and the audio and video data with high resolution and/or high compression ratio is taken as the audio and video data for playing.
2) The audio and video data for browsing are fused into an audio and video data stream, the audio and video data for playing are sent to the smart phone through the information stream, and the audio and video data for playing are sent to the smart phone through other information streams.
In the whole playing process, no matter the audio and video data for browsing or the audio and video data for playing are in the same audio and video format, the audio and video playing module 302 does not need to be reloaded for different formats, and time delay does not occur for loading data when audio and video are browsed or jumped, so that the connection of the audio and video is smooth, and the watching experience of a user is improved.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim.

Claims (10)

1. A multi-type audio and video mixed playing system is characterized by comprising: an audio and video processor (200) and a playing interactive terminal (300),
the audio and video processor (200) comprises an information stream receiving module (204 a), an information stream sending module (204 b), a tag adding module (201), a video transcoding module (202) and a video fusion module (203);
the information stream receiving module (204 a) receives audio and video data sent by a plurality of audio and video sources, wherein at least one part of the audio and video data is different types of audio and video data;
the tag adding module (201) is used for adding tags to the audio and video data, and the tags are associated with the corresponding audio and video data;
the video transcoding module (202) is used for carrying out unified transcoding processing on videos in the audio and video data, each audio and video data at least obtains two audio and video data with different resolutions, the audio and video data with low resolution and/or high compression rate is used as audio and video data for browsing, and the audio and video data with high resolution and/or high compression rate is used as audio and video data for playing;
the video fusion module (203) is used for fusing a plurality of browsing audio and video data to splice into an audio and video data stream which is used as the browsing audio and video data stream;
the information flow sending module (204 b) sends the audio-video data flow for browsing and the audio-video data for playing to the playing interactive terminal (300) by using different information flows, sends the information flow of the audio-video data flow for browsing as a browsing data information flow, and sends the information flow of the audio-video data for playing as a playing data information flow;
the playing interactive terminal (300) comprises a communication module (304), an audio and video playing module (302) and a user operation identification module (303);
the communication module (304) receives browsing data information stream and playing data information stream sent by the audio and video processor (200);
the audio and video playing module (302) selects to browse the data information stream or play the data information stream for playing according to preset conditions;
the user operation identification module (303) identifies the operation mode of the user when the audio/video playing module (302) plays the browsing data information stream, including video moving operation and video skipping operation,
when the video moving operation is identified, the audio and video playing module (302) plays the next or previous audio and video of the browsing data information stream,
when the video skipping operation is identified, the audio and video playing module (302) skips to the corresponding playing data information stream for playing according to the label corresponding to the operation.
2. The system according to claim 1, wherein the multi-type audio-video mixing system,
the identification mode of the user operation identification module (303) comprises a return operation,
and when the user operation identification module (303) identifies a return operation in the state of playing the playing data information stream, the audio/video playing module (302) returns the browsing data information stream for playing.
3. The system according to claim 2, wherein the tag adding module (201) analyzes each audio/video data, sets a pattern of the tag according to the type of the audio/video, and the tags include one or more tags.
4. The system of claim 3, wherein the tag adding module (201) adds the tag to the lower side of the video.
5. The multi-type audio/video mixed playing system according to claim 4, wherein the video transcoding module (202) respectively adjusts the audio/video data for browsing and the audio/video data for playing to the same resolution when performing the uniform transcoding processing on the video in the audio/video data.
6. The multi-type audio/video mixed playing system according to any one of claims 1 to 5, wherein the video transcoding module (202) generates a plurality of audio/video data for playing with different resolutions when performing a unified transcoding process on the video in each of the audio/video data,
the information stream sending module (204 b) sends the audio and video data for playing with different resolutions to the playing interactive terminal (300) by different information streams.
7. A multi-type audio and video mixed playing method is characterized by being executed by an audio and video processor (200) and a playing interactive terminal (300), and comprising the following steps:
an audio and video data receiving step (S10), wherein an audio and video processor (200) receives audio and video data sent by a plurality of audio and video sources, and at least one part of the audio and video data is different types of audio and video data;
a tag adding step (S20), wherein the audio and video processor (200) adds tags to the audio and video data, and the tags are associated with the corresponding audio and video data;
a video transcoding step (S40), wherein the audio and video processor (200) performs unified transcoding processing on videos in the audio and video data, each audio and video data at least obtains two audio and video data with different resolutions, the audio and video data with low resolution and/or high compression rate is used as the audio and video data for browsing, and the audio and video data with high resolution and/or high compression rate is used as the audio and video data for playing;
a video fusion step (S50), wherein the audio and video processor (200) performs fusion processing on the plurality of audio and video data for browsing to splice into an audio and video data stream as the audio and video data stream for browsing;
an information stream sending step (S60), wherein the audio and video processor (200) sends the audio and video data stream for browsing and the audio and video data for playing to the playing interactive terminal (300) in different information streams, the information stream of the audio and video data stream for browsing is sent as a browsing data information stream, and the information stream of the audio and video data for playing is sent as a playing data information stream;
an information stream receiving step (S110), wherein the playing interactive terminal (300) receives browsing data information streams and playing data information streams sent by the audio and video processor (200);
an audio and video playing step (S120), wherein the playing interactive terminal (300) selects to browse the data information stream or play the data information stream for playing according to preset conditions;
a user operation identification step (S130), in which the playing interactive terminal (300) identifies the user' S operation modes including a video moving operation and a video skipping operation when playing the browsing data information stream,
audio and video switching steps (S130, S140, S150, S160), when the video moving operation is identified, the playing interactive terminal (300) plays the next or previous audio and video in the browsing data information stream,
and audio and video skipping steps (S130, S170 and S180), wherein when the playing interactive terminal (300) identifies the video skipping operation, skipping to the corresponding playing data information stream for playing according to the label corresponding to the operation.
8. The method of claim 7, wherein the multi-type audio-video mixing playing method,
and a return operation identification step (S190) is also included, and when the play interactive terminal (300) identifies a return operation in the state of playing the play data information stream, the browse data information stream is returned to play.
9. The method of claim 8, wherein the multi-type audio-video mixing playing method,
in the step (S20) of adding the tags, the playing interactive terminal (300) analyzes the audio and video data, and sets the pattern of the tags according to the type of the audio and video, wherein the tags comprise one or more tags.
10. The multi-type audio-video mixing playing method according to any one of claims 7-9,
in the video transcoding step (S40), when the audio/video processor (200) performs the unified transcoding processing on the video in each audio/video data, a plurality of audio/video data for playing with different resolutions are generated,
in the video fusion step (S50), when the video in each audio and video data is subjected to unified transcoding processing, a plurality of audio and video data for playing with different resolutions are generated,
in the information stream sending step (S60), the audio/video processor (200) sends the audio/video data for playing with different resolutions to the playing interactive terminal (300) in different information streams.
CN202210813711.5A 2022-05-11 2022-07-11 Multi-type audio and video mixed broadcasting system and method Active CN115225928B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210532881 2022-05-11
CN2022105328816 2022-05-11

Publications (2)

Publication Number Publication Date
CN115225928A true CN115225928A (en) 2022-10-21
CN115225928B CN115225928B (en) 2023-07-25

Family

ID=83611043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210813711.5A Active CN115225928B (en) 2022-05-11 2022-07-11 Multi-type audio and video mixed broadcasting system and method

Country Status (1)

Country Link
CN (1) CN115225928B (en)

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1949839A (en) * 2005-10-11 2007-04-18 美国博通公司 Video processing system, set-top box and method for treating audio/video information
CN101669361A (en) * 2007-02-16 2010-03-10 马维尔国际贸易有限公司 Methods and systems for improving low resolution and low frame rate video
CN102625150A (en) * 2012-03-16 2012-08-01 中国科学院计算技术研究所 Media playing system and method
US20130279571A1 (en) * 2012-04-18 2013-10-24 Vixs Systems, Inc. Video processing system with stream indexing data and methods for use therewith
CN103873888A (en) * 2012-12-12 2014-06-18 深圳市快播科技有限公司 Live broadcast method of media files and live broadcast source server
CN104160408A (en) * 2011-12-29 2014-11-19 派尔高公司 Method and system for video composition
CN104410917A (en) * 2014-09-16 2015-03-11 东方有线网络有限公司 Realization method for cable interactive television cross-domain video service connection system
CN104796732A (en) * 2015-04-21 2015-07-22 无锡天脉聚源传媒科技有限公司 Audio and video editing method and device
CN105141985A (en) * 2015-07-30 2015-12-09 世纪龙信息网络有限责任公司 System and method for smoothly playing film and television stream
CN108391127A (en) * 2018-04-27 2018-08-10 腾讯科技(深圳)有限公司 Method for video coding, device, storage medium and equipment
CN108712654A (en) * 2018-05-18 2018-10-26 网宿科技股份有限公司 A kind of code-transferring method and equipment of audio/video flow
EP3410728A1 (en) * 2017-05-30 2018-12-05 Vestel Elektronik Sanayi ve Ticaret A.S. Methods and apparatus for streaming data
CN109547839A (en) * 2018-12-11 2019-03-29 网易(杭州)网络有限公司 The dynamic previewing processing method and processing device of live content, medium, electronic equipment
WO2019072096A1 (en) * 2017-10-10 2019-04-18 腾讯科技(深圳)有限公司 Interactive method, device, system and computer readable storage medium in live video streaming
CN110662086A (en) * 2019-10-31 2020-01-07 成都威爱新经济技术研究院有限公司 5G high-definition live broadcast system and video processing method
US20200059654A1 (en) * 2018-08-16 2020-02-20 BombBomb, LLC Method and System for Implementing Split and Parallelized Encoding or Transcoding of Audio and Video Content
CN113573101A (en) * 2021-07-09 2021-10-29 百果园技术(新加坡)有限公司 Video encoding method, device, equipment and storage medium
WO2021236059A1 (en) * 2020-05-19 2021-11-25 Google Llc Dynamic parameter selection for quality-normalized video transcoding

Patent Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1949839A (en) * 2005-10-11 2007-04-18 美国博通公司 Video processing system, set-top box and method for treating audio/video information
EP1775959A2 (en) * 2005-10-11 2007-04-18 Broadcom Corporation Video processing system with simultaneous multiple outputs each with unique formats
CN101669361A (en) * 2007-02-16 2010-03-10 马维尔国际贸易有限公司 Methods and systems for improving low resolution and low frame rate video
CN104160408A (en) * 2011-12-29 2014-11-19 派尔高公司 Method and system for video composition
CN102625150A (en) * 2012-03-16 2012-08-01 中国科学院计算技术研究所 Media playing system and method
US20130279571A1 (en) * 2012-04-18 2013-10-24 Vixs Systems, Inc. Video processing system with stream indexing data and methods for use therewith
CN103428488A (en) * 2012-04-18 2013-12-04 Vixs系统公司 Video processing system with pattern detection and method for use thereof
CN103873888A (en) * 2012-12-12 2014-06-18 深圳市快播科技有限公司 Live broadcast method of media files and live broadcast source server
CN104410917A (en) * 2014-09-16 2015-03-11 东方有线网络有限公司 Realization method for cable interactive television cross-domain video service connection system
CN104796732A (en) * 2015-04-21 2015-07-22 无锡天脉聚源传媒科技有限公司 Audio and video editing method and device
CN105141985A (en) * 2015-07-30 2015-12-09 世纪龙信息网络有限责任公司 System and method for smoothly playing film and television stream
EP3410728A1 (en) * 2017-05-30 2018-12-05 Vestel Elektronik Sanayi ve Ticaret A.S. Methods and apparatus for streaming data
CN110710220A (en) * 2017-05-30 2020-01-17 韦斯特尔电子工业和贸易有限责任公司 Method and apparatus for streaming data
WO2019072096A1 (en) * 2017-10-10 2019-04-18 腾讯科技(深圳)有限公司 Interactive method, device, system and computer readable storage medium in live video streaming
CN109660853A (en) * 2017-10-10 2019-04-19 腾讯科技(北京)有限公司 Interactive approach, apparatus and system in net cast
CN108391127A (en) * 2018-04-27 2018-08-10 腾讯科技(深圳)有限公司 Method for video coding, device, storage medium and equipment
CN108712654A (en) * 2018-05-18 2018-10-26 网宿科技股份有限公司 A kind of code-transferring method and equipment of audio/video flow
US20200059654A1 (en) * 2018-08-16 2020-02-20 BombBomb, LLC Method and System for Implementing Split and Parallelized Encoding or Transcoding of Audio and Video Content
CN109547839A (en) * 2018-12-11 2019-03-29 网易(杭州)网络有限公司 The dynamic previewing processing method and processing device of live content, medium, electronic equipment
CN110662086A (en) * 2019-10-31 2020-01-07 成都威爱新经济技术研究院有限公司 5G high-definition live broadcast system and video processing method
WO2021236059A1 (en) * 2020-05-19 2021-11-25 Google Llc Dynamic parameter selection for quality-normalized video transcoding
CN113573101A (en) * 2021-07-09 2021-10-29 百果园技术(新加坡)有限公司 Video encoding method, device, equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王灵芝;李伟滨;: "基于Hi3515C的车载驾驶音视频监控系统设计", 闽南师范大学学报(自然科学版), no. 02, pages 36 - 42 *

Also Published As

Publication number Publication date
CN115225928B (en) 2023-07-25

Similar Documents

Publication Publication Date Title
US11073969B2 (en) Multiple-mode system and method for providing user selectable video content
US9870801B2 (en) Systems and methods for encoding multimedia content
US10623816B2 (en) Method and apparatus for extracting video from high resolution video
US7644425B2 (en) Picture-in-picture mosaic
US9271048B2 (en) Systems and methods for immersive viewing experience
US20080101456A1 (en) Method for insertion and overlay of media content upon an underlying visual media
JP4000844B2 (en) Content distribution system, content distribution system distribution server and display terminal, and content distribution program
US10250838B1 (en) System and method for converting live action alpha-numeric text to re-rendered and embedded pixel information for video overlay
CN101917568A (en) Streaming media technology-based network video player and playing system
CN1256583A (en) Video/audio in cooperation with video/audio broadcasting and graphic demonstrating system
KR100948691B1 (en) Apparatus and method for providing 3d personal advertisement broadcasting and apparatus for receiving internet broadcasting
US11012658B2 (en) User interface techniques for television channel changes
US20010017671A1 (en) "Midlink" virtual insertion system and methods
US20110200093A1 (en) Method and apparatus for transmitting and receiving video and video links
JP4655065B2 (en) Content distribution system, content distribution system distribution server and display terminal, and content distribution program
CN115225928B (en) Multi-type audio and video mixed broadcasting system and method
KR101452902B1 (en) Broadcasting receiver and controlling method thereof
KR20180003608A (en) Method for rendering audio-video content, decoder implementing the method, and rendering device for rendering audio-video content
Deshpande et al. Omnidirectional MediA Format (OMAF): toolbox for virtual reality services
CN117939183B (en) Multi-machine-position free view angle guided broadcasting method and system
CN118283287A (en) Live broadcast method, device, medium and equipment for hundred million-level pixel video
CN117939183A (en) Multi-machine-position free view angle guided broadcasting method and system
KR20150122039A (en) System and method for operating home shopping using mobile application
Olaizola et al. Live HDR video broadcast production
KR101441867B1 (en) Method and Gateway Device for Providing Contents to Media Device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant