WO2019088853A1 - Remplacement audio en direct dans un flux numérique - Google Patents

Remplacement audio en direct dans un flux numérique Download PDF

Info

Publication number
WO2019088853A1
WO2019088853A1 PCT/NZ2018/050155 NZ2018050155W WO2019088853A1 WO 2019088853 A1 WO2019088853 A1 WO 2019088853A1 NZ 2018050155 W NZ2018050155 W NZ 2018050155W WO 2019088853 A1 WO2019088853 A1 WO 2019088853A1
Authority
WO
WIPO (PCT)
Prior art keywords
secondary audio
primary video
segment
audio
data
Prior art date
Application number
PCT/NZ2018/050155
Other languages
English (en)
Inventor
Michael Philp PRENDERGAST
Original Assignee
Klaps Limited
Spalk (Us) Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Klaps Limited, Spalk (Us) Inc. filed Critical Klaps Limited
Publication of WO2019088853A1 publication Critical patent/WO2019088853A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8211Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a sound signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23424Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2347Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving video stream encryption
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/23608Remultiplexing multiplex streams, e.g. involving modifying time stamps or remapping the packet identifiers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/23611Insertion of stuffing data into a multiplex stream, e.g. to obtain a constant bitrate
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/2365Multiplexing of several video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/167Audio streaming, i.e. formatting and decoding of an encoded audio signal representation into a data stream for transmission or storage purposes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/04Time compression or expansion
    • G10L21/055Time compression or expansion for synchronising with other signals, e.g. video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/238Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
    • H04N21/2389Multiplex stream processing, e.g. multiplex stream encrypting
    • H04N21/23895Multiplex stream processing, e.g. multiplex stream encrypting involving multiplex stream encryption
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4622Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet

Definitions

  • the invention generally relates to a system and method for live audio replacement in a digital stream, and in particular, replacing a primary audio stream with an alternative audio stream in a digital media broadcast environment.
  • Streamed media is typically a combination of video data and audio data provided by a source.
  • alternate audio content is desirable.
  • a user is not satisfied with the audio provided by the source, where a user may desire alternative audio data such as a language different to that provided by the source, prefer to listen to different announcers of a sporting event sanitised dialog, and/or alternate commentary.
  • Systems, methods and products are therefore desired to allow the retrieval of alternate audio sources to accompany video playback.
  • the invention consists in a method performed by at least one processor, the method comprising : receiving encoded primary video data via a streaming network, the primary video data comprising at least a first and second discrete segments configured for sequential playback; determining a temporal playback reference for at least the second received primary video data segment; receiving a secondary audio source; encoding the secondary audio source as a sequence of discrete segments, the first encoded secondary audio segment comprising a temporal playback reference to match the temporal playback reference of the second received primary video segment.
  • the method further comprises determining the duration of at least one of the received primary video data segments; then, encoding the secondary audio source into segments with a duration to match the duration of the at least one determined primary video data segment, or a multiple thereof.
  • the duration of a received primary video data segment is determined by one or more of: metadata associated with the primary video data, measuring a received primary video data segment, and/or calculating the segment duration from bitrate information.
  • the metadata comprises timebase data and/or presentation timestamp data.
  • the method further comprises determining time references from a clock source, the time references comprising : a variable tO corresponding to a time at which secondary audio source commences; a variable tl that is a time reference in the secondary audio source that corresponds to a time at which the second discrete segment of the primary video stream is due for playback; a variable sO corresponding to the primary video stream time at tO; and a variable si corresponding to the time at which the second discrete segment of the primary video stream is due for playback.
  • the method further comprises truncating the secondary audio source for a period between tO and tl, and encoding the secondary audio source from tl.
  • the first segment of the encoded secondary audio source comprises a temporal playback reference to match the temporal playback reference of the second received primary video segment, the temporal playback reference
  • the method further comprises: receiving a primary master playlist containing information pertaining to the availability of one or more primary video sources and one or more associated primary audio sources;
  • the method further comprises: determining the secondary audio source is unavailable, and encoding an interim audio source in place of the secondary audio source.
  • the invention consists in a system configured to provide secondary audio for playback with a primary video source, the system
  • a streaming network configured stream at least video data from an encoded primary video data source to one or more viewers, the primary video data comprising at least a first and second discrete segments configured for sequential playback; a secondary audio data source; and
  • At least one processor configured to: capture secondary audio data from the secondary audio data source; receive the encoded primary video data via the streaming network; determine a temporal playback reference for at least the second received primary video data segment; receive secondary audio from the secondary audio source; and encode the secondary audio data as a sequence of discrete segments, the first encoded secondary audio segment comprising a temporal playback reference to match the temporal playback reference of the second received primary video segment.
  • the processor is further configured to:
  • the duration of a received primary video data segment is determined by one or more of: metadata associated with the primary video data, measuring a received primary video data segment, and/or calculating the segment duration from bitrate information.
  • the metadata comprises timebase data and/or start presentation timestamp data.
  • the processor is further configured to determine time references from a clock source, the time references comprising : a variable tO corresponding to a time at which secondary audio source commences; a variable tl that is a time reference in the secondary audio source that corresponds to a time at which the second discrete segment of the primary video stream is due for playback; a variable sO corresponding to the primary video stream time at tO; and a variable si corresponding to the time at which the second discrete segment of the primary video stream is due for playback.
  • the processor is further configured to:
  • the first segment of the encoded secondary audio source comprises a temporal playback reference to match the temporal playback reference of the second received primary video segment, the temporal playback reference
  • the processor is further configured to:
  • a primary master playlist containing information pertaining to the availability of one or more primary video sources and one or more associated primary audio sources; receive one or more signals indicative of the availability of one or more secondary audio sources; publish a secondary masterlist containing information one or more primary video sources and one or more associated secondary audio sources.
  • the processor is further configured to: determine the secondary audio source is unavailable, and encode an interim audio source in place of the secondary audio source.
  • any reference to any range of numbers disclosed herein also incorporates reference to all rational numbers within that range (for example, 1, 1.1, 2, 3, 3.9, 4, 5, 6, 6.5, 7, 8, 9 and 10) and also any range of rational numbers within that range (for example, 2 to 8, 1.5 to 5.5 and 3.1 to 4.7).
  • Figure 1 is a schematic diagram of components in an environment for providing alternative audio to a primary video source.
  • Figure 2 is a diagram of a temporal process for providing synchronisation between video and audio data streams.
  • Figure 3 is a flow diagram of a process for providing synchronisation between video and audio data streams.
  • Figure 4 is an example of a viewer webpage.
  • system may comprise software, hardware, or a combination thereof.
  • the software can be machine code, firmware, embedded code, and application software.
  • the hardware can be circuitry, processor, computer, integrated circuit, integrated circuit cores, active or passive sensors or sensing equipment, or a combination thereof.
  • the subject matter may be embodied as devices, systems, methods, and/or computer program products. Accordingly, some or all of the subject matter may be embodied in hardware and/or in software (including firmware, resident software, micro-code, state machines, gate arrays, etc.) Furthermore, the subject matter may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system.
  • a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium.
  • computer readable media may comprise computer storage media and communication media.
  • Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
  • Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by an instruction execution system.
  • the computer-usable or computer-readable medium could be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, of otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
  • Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
  • modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer readable media.
  • the embodiment may comprise program modules, executed by one or more systems, computers, or other devices.
  • program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
  • functionality of the program modules may be combined or distributed as desired in various embodiments.
  • a client device may represent any type of device that may, communicate with a server or other hardware and/or software for receiving video and audio data.
  • the hardware components may represent a typical architecture of a computing device, such as a desktop or server computer.
  • the hardware components may include a processor, random access memory, nonvolatile storage or any form of computer readable medium.
  • the processor may be a single microprocessor, multi-core processor, or a group of processors.
  • the random access memory may store executable code as well as data that may be immediately accessible to the processor, while the nonvolatile storage may store executable code and data in a persistent state.
  • the hardware components may also include one or more user interface devices and network interfaces.
  • the user interface devices may include monitors, displays, keyboards, pointing devices, and any other type of user interface device.
  • the network interfaces may include hardwired and wireless interfaces through which the device may communicate with other devices.
  • the software components may include an operating system on which various applications may execute.
  • user refers to an individual such as a person, or a group or people, or a business such as a retailer or advertiser of one or more a products or services.
  • a business such as a retailer or advertiser of one or more a products or services.
  • primary video source or “primary audio source” used in this specification and claims refer to an original source of video or audio such as that recorded by a camera at an event and intended by the provider of at least the video to be provided together.
  • secondary audio source used in this specification and claims refers to a source of an alternative sound source intended to accompany the primary video source and be mixed with, be mixed with a part of, or replace the primary audio source.
  • first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first device could be termed a second device, and, similarly, a second device could be termed a first device without departing from the teachings of the disclosure.
  • Exemplary embodiments provide methods, systems, and products for searching, retrieving, and synchronizing one or more alternate audio sources with one or more video sources.
  • Exemplary embodiments identify alternate audio content that may be separately available from video content.
  • exemplary embodiments permit the user to seek out and retrieve alternate audio content from the Internet, or from any other source.
  • the video content may self-identify one or more alternate audio sources that correspond to the video content.
  • the video content for example, may be tagged or embedded with websites, server addresses, frequencies, or other information that describe the alternate audio sources.
  • Exemplary embodiments may even automatically query database servers for alternate audio sources that are associated with to the video content.
  • exemplary embodiments may then synchronize the video content and the separately- available alternate audio content.
  • the video content and the alternate audio content may be received as separate streams of data, therefore either of the streams may lead or lag the other.
  • Exemplary embodiments, then, may also synchronise the separately-received streams of data to ensure coherent and intelligible media.
  • exemplary embodiments discussed herein include a method for retrieving an audio signal.
  • video data that comprises a time reference is received and the audio signal is aligned with that time reference.
  • Substantially synchronised video and audio data streams are produced and provided to a viewer for playback. Note that substantially synchronised in this context is intended to mean that while video and audio streams are played, the perception of the viewer is that the video or audio stream does not lead or lag the other. Lead or lag time of either the video or audio streams may be up to around 100 ms before a viewer perceives a lack of synchronisation. With modern telecommunication and computing equipment, much lower lead or lag times are typically achievable.
  • the computer program product has processor-readable instructions for retrieving an audio signal.
  • a video signal is received that comprises a time reference.
  • the audio signal is aligned with the time reference of the video signal.
  • the aligned video and audio signals are then provided for playback to a user.
  • a computer program for retrieving video and audio data relating to a sporting event from a primary source.
  • the computer program is programmed to remove the retrieved audio data.
  • the computer program is further programmed to retrieve audio data from a secondary source.
  • the computer program then aligns the primary video data with the secondary audio data for playback to a user.
  • a computer program for retrieving video data relating to a sporting event from a primary source.
  • the computer program is further programmed to retrieve audio data from a secondary source.
  • the computer program then aligns the primary video data with the secondary audio data for playback to a user.
  • Video is typically produced in one location, and provided for viewing in one or more other locations.
  • Audio is typically captured with the video and provided with that video to viewers.
  • the captured audio would include a commentary of the play by play aspects of the event.
  • the captured audio may also include multiple other audio sources from the location such as stadium or ambient audio.
  • the multiple audio sources may be mixed together into a single audio feed and output for consumption. Or in some instances, multiple audio sources be output for selective consumption such as where multiple language commentary is provided, and a viewer is able to select which language they wish to listen to.
  • the event may be broadcast to multiple countries in different languages.
  • the original primary video feed may be created and produced at the sporting event location, then encoded and transmitted to a video CDN as a live video stream with accompanying audio.
  • the ambient audio soundtrack may merely be audience sounds and other sounds of the event, but without commentary.
  • commentators may produce commentary to be added to the live video stream, and consumers may download and view the video and newly created audio. Because the sporting event may be broadcast to multiple countries, each country or region may have their own commentators in their own language or to address specialised interest groups.
  • the viewer is preferably a user with a web connected client device that operates a display and an audio output.
  • the device typically comprises hardware components and software components.
  • the hardware components may represent a typical architecture of a computing device, such as a personal computer, game console, network appliance, interactive kiosk, or a portable device such as a laptop computer, netbook computer, personal digital assistant, mobile telephone, smart telephone, tablet or other mobile device.
  • the viewer will typically operate a software application that allows selection of media they wish to view.
  • the software application may include a webpage or smartphone application which displays a selection of available media.
  • the available media includes a video source and a selection or two or more audio sources, where one audio source comprises original audio content provided by the video data provider, and another audio source comprises alternative audio content provided by an alternative audio source.
  • available media is presented as a list of options from which the user may make a selection.
  • the software is configured to download a manifest of video and audio source locations enabled for download and viewing.
  • Time reference information facilitates the temporal alignment of video and audio presented to a viewer.
  • Time reference information is especially important when video and audio data is encoded and transmitted over long distances where data transmission delays are typical. Time reference information is added to the video and audio data during an encoding process, or, at least before the data is transmitted.
  • Embodiments of the invention address this problem by provision of a system, product and method for adding time reference information to a new audio stream generated at one location such that synchronisation with an existing video stream generated at some other location can occur.
  • Figure 1 is a schematic of an environment 100 in which exemplary embodiments may operate.
  • environment 100 is a simplified example of a video production and consumption system with the provision of one or more audio sources alternative to that of the primary audio provider.
  • a local environment is shown at 101.
  • the local environment comprises one or more video sources 102 and one or more audio sources 112.
  • a production studio 103 receives video and audio data and performs any mixing of the video sources 102 into a video feed, and optionally, mixing of any audio sources 112 to creates one or more audio feeds.
  • a client device 105 is configured to ultimately receive and display media to an end viewer.
  • the client device can be any device configured for media playback.
  • An encoder 106 receives the video feed and audio feed from the production studio 103. Compression may be necessary or desired in instances where the video and audio feeds are to be streamed over external networks. Uncompressed video and audio data may require unnecessarily large amounts of bandwidth which may in turn cause delayed or jittery playback to a viewer, and limit the number of streams to potential viewers. Uncompressed video and/or audio may however be transported between cameras and the production studio 103 at the local event 101, then compressed, encoded and/or packaged for internet streaming by the encoder 106. The encoding may be a multimedia codec that may compress the video and audio signals. The stream will be packaged into a multimedia container format. Examples of a multimedia container formats include, but is not limited to, MPEG, 3GP, WMA, WMV, AVI, MKV, Ogg and other formats.
  • Each multimedia container format may have different characteristics depending on its best suited purpose. For example, multimedia container format intended for streaming over the Internet may have less resolution than the original capture.
  • a multimedia container format may contain audio, video, and various other data. Some container formats may interleave different data types, and some may, support multiple types of audio or video streams. Many container formats may include metadata, which may include captioning, titles, or other metadata.
  • Container formats may have different names for various portions of the container which are generally referred to as packets and contain video and/or audio data.
  • the containers may have tags, headers or other kinds of metadata that facilitate
  • Modern streaming protocols for delivery over the internet take an audio visual signal and chunk it for delivery over the internet. These streaming protocols are collectively known as segment based streaming protocols and includes protocols such as HTTP Live
  • HLS Dynamic Adaptive Streaming over HTTP
  • URI Uniform Resource Identifier
  • a Master Playlist is prepared by the encoder 106 during the encoding process.
  • the master playlist contains references to video and audio streams in the form of URLs.
  • the master playlist is made available for download by any media player on a client device.
  • a client device is configured to reference the master playlist and be directed to a URL where video and/or audio streams can be downloaded for playback.
  • the master playlist also contains nominal metadata on each stream including resolution, codecs required to play the streams, any content protection mechanisms and stream pairing (which video streams are intended to be played with which audio streams) .
  • a Media Playlist is also prepared by the encoder 106.
  • the media playlist contains references to segments of media for streaming in the form of URLs.
  • the media playlist also attributes an ID to each segment, such as an integer ID. The ID enables alignment of segments between simultaneous playlists by indicating the orders of segments in a stream.
  • the media playlist also contains nominal duration information of each segment. The duration information is intended to inform the media player and enable playback decisions, but not be an exact reference. The exact duration of each seg ment is found by obtaining and parsing it.
  • the master playlist also contains references to a corresponding media playlists.
  • Each media segment may comprise a video data segment, an audio data segment, a metadata segment, or combination of each, or a combination of two or more of audio, video and/or metadata segments.
  • the media segments comprise media samples to be decoded, distinguished as to the type of sample they are, and rendered where appropriate by a media player such as dedicated hardware or software operating on a client device.
  • a metadata segment comprises information relevant to the one or more other streams contained in the segment, such as audio and/or video streams.
  • the form and presence of the metadata depends on the container format used. For example, an MPEG-TS container format has a list programs, each with a list of streams, and for each stream :
  • the media stream includes only audio data, there is no requirement to include fields relating to video such as the width and height fields.
  • the encoded video and audio from the encoder 106 is received by a media content distribution network (CDN) 107.
  • the video CDN is a large geographically dispersed network of servers that are configured to deliver live and on-demand videos to the web connected devices of users spread over a large geographical area. When a viewer submits a request, the server caching content nearest the user's location delivers the video.
  • the container formats have timestamp data that allows for synchronisation of streamed data segments and/or synchronisation of separately streamed video and audio data.
  • a container format may be configured so that a viewer may download audio and video data separately. Other embodiments may download audio and video data together.
  • the video and audio streams are published at a publishing point which is typically a media player application, window within a web browser or facilitated by a smartphone application.
  • a viewer page 105 is depicted by the exemplary example in Figure 1.
  • the viewer page preferably includes a video playback area, and a list of one or more available audio sources to accompany the video playback.
  • the audio sources may optionally comprise original audio content from the video source provider, and one or more streams of alternative or secondary audio from another source.
  • the viewer page 105 will typically include a decoder that may decode an incoming video and audio stream.
  • the decoder may decode the video stream so that a video presentation system may play back the video and audio.
  • the viewer page 105 will typically be operated by a client device connected to a network, which may be the Internet, wide or local area network.
  • the network may be a wired network, wireless network, or combination of wired and wireless networks.
  • the viewer page displayed on the client device 105 such as exemplified by Figure 3, will typically include a menu or set of options that is set to live update the manifest and the available alternative audio sources. These options are presented to the user.
  • the decoder on the page will use the manifest to retrieve the media playlist that represents that stream. Once the player has the media playlist the decoder will retrieve the relevant audio segments that are required to continue playback. On retrieval of the audio segments the decoder will parse and decode the audio samples to be rendered synchronised to the video.
  • additional environment components comprise a video server 108, an audio server 109, and a client device 110 (herein referred to as a secondary client device).
  • the secondary client device 110 is operable to receive and display video to the secondary audio source provider.
  • the provider may be, for example, a commentator providing a personal perspective or different language for the video event being captured by the primary video source 102.
  • the secondary client device 110 is further configured capture audio from one or more audio capturing devices 111 to thereby define a secondary audio source.
  • the audio capturing devices 111 include a microphone and an alternative commentator provides audio for capture and
  • the secondary audio source provider is a person commentating on a video.
  • the commentator page can be gated behind an authentication mechanism in order to restrict the quality of commentary and
  • the layout of these pages may take many forms, although for example a 'broadcast' button on a broadcaster webpage is provided as an interface which facilitates control over creating a new alternative audio track.
  • the viewer page may provide a list of available commentators associated with a particular video source, or a subset or filtered set of options may be determined by for example if the user is a premium member with access to exclusive content or if a user has elected to filter available options by specific language settings.
  • a video stream may need to be delayed by a corresponding amount in order for synchronisation to occur correctly. Delay is controlled by the time reference information provided to the secondary audio stream and by the player operating to control delivery of video to viewers.
  • audio synchronisation logic is handled by a JavaScript module embedded in the commentator page.
  • the secondary client device 110 is operated by a user to capture secondary audio related to a video event and is intended to be mixed with, or replace the primary audio captured at the local environment 112.
  • the secondary client device 110 may receive video data directly from the video CDN 107, or from the video server 108.
  • the video server 108 acts as an server operable to provide information on the availability of one or more primary video sources to one or more users of secondary client devices 110.
  • the audio server 109 is configured to receive secondary audio data from the secondary client device 110.
  • the audio server 109 may further receive audio from the video CDN 107, the video server 108, or the secondary client device 110.
  • the secondary audio data may contain any type of audio signal.
  • additional audio tracks may include announcers or commentators that speak various languages to the sporting event for people of different languages or nationalities. Other examples may include adding music, voice, sound effects, or other audio signal.
  • the video server 108 and audio server 109 are discussed within this specification and claims with reference to the tasks they are expected to perform. However, it should be noted that the video server 108 and audio server 109 may be operated by the same server. In such instances, a server would be configured to at least deliver video data to a secondary client device 110 and receive audio data from that same device. The server may further communicate with any number of secondary audio providers as required.
  • synchronisation of a new secondary audio stream with a primary video stream is achieved by applying time reference information to the secondary audio stream to match the timestamps in the primary video stream. This produces secondary audio stream that is able to be played in time with the video stream to the viewer.
  • the viewer's client device 105 has media playback software that is configured to retrieve a secondary master playlist URL from the server(s) 108, 109.
  • the secondary master playlist contains media playlists which provide the locations of one or more primary video streams and primary and/or secondary audio streams.
  • the various streams are provided in the form of URLs which point to internet based addresses of where video and/or audio data can be downloaded for playback.
  • the viewer's client device 105 On selection of an alternative audio track from the master playlist, the viewer's client device 105 will then download a media playlist that corresponds to the selected alternative audio track.
  • the media playlist contains URLs as references to the audio segments.
  • the master playlist can be statically generated by the server(s) 108, 109 with a predetermined, fixed number of alternative audio streams, or can be dynamically created with a variable number of alternative audio streams. Dynamic generation typically involves a database query to find the number and metadata of active
  • the master playlist can be additively generated as new streams are added through time.
  • the metadata required for alignment between primary video and secondary audio can vary between container formats.
  • the segment IDs and boundaries of new secondary audio are aligned so that each segment is time-aligned with the corresponding primary video segments.
  • the presentation timestamp container format header fields are set to align with the primary video stream segments within an error tolerance.
  • FIG. 2 shows a process to apply time reference information to a new secondary audio stream is illustrated generally by a timeline of events.
  • the upper timeline 201 represents a continuous stream of primary video segments (and optionally primary audio segments) each having a duration 205 defined by their native encoding process, such as by encoder block 106.
  • the second timeline 202 represents a secondary audio stream commencing at point in time 208 and continues for as long as the provider of the secondary audio source desires.
  • Timeline 203 is a reference clock.
  • Timeline 204 shows a resulting secondary audio stream Ro encoded with time reference information that corresponds to the segment playback times of the primary video stream.
  • ti the next segment boundary in the primary video stream referenced to the audio stream time
  • So the primary video data stream time at the start of recording
  • Si the primary video data stream time at the next video segment.
  • to and tl are determined, and ti-to calculated such that ro in the resultant encoded secondary audio stream matches Si in the primary video stream .
  • the segment duration Sv 205 of a received encoded primary video segment is recorded as a variable by the video server 108 to be matched by encoded segments of the secondary audio stream .
  • the primary video stream is already divided into segments So, Si for steaming as is typically performed by the encoder 106.
  • the duration of the segments may be measured or may be retrieved from associated metadata retrieved from the media playlist.
  • the reference clock is initiated and monitored from the commencement of the secondary audio stream at to or the first reception of the primary video stream.
  • the reference clock can be provided by any available system such as, for example, the system clock that any client device will have.
  • the reference clock is synchronised with other clocks in the system, such as system clock of the encoder 106.
  • the reference clock may be derived from any other clock in the system, such as the clock of the encoder 106.
  • initial preparation is undertaken during encoding the secondary audio stream .
  • the encoding process splits the stream of secondary audio data into segments each with a duration that matches the duration 205 of the primary video stream.
  • the segments are further assigned time reference information that ensures synchronicity with the primary video stream during a later decoding process.
  • the duration of the secondary audio segments does not need to match the duration of the primary video segments.
  • the master playlist or manifest includes time reference information indicative of the desired playback start time of the primary video and/or secondary audio relative to a time reference such as a clock.
  • the duration of the secondary audio may be, for example, an arbitrary duration, or a multiple of the time duration of the primary video source segment duration .
  • the assigned time reference information is a determined by the following exemplary process :
  • time reference information assigned to the first resulting segment Ro of the secondary audio stream therefore causes playback of a segment to be aligned with the next primary video segment.
  • reference to segment playback time could be determined by a variety of methods.
  • the above described example derives time reference from a system clock and time reference information contained in metadata associated with the primary video data.
  • time reference information could be derived from the received file sizes, such as the number of bits received by the transmission. Where the bitrate of any encoded video and/or audio data is known, a number of bits received provides a reference to the temporal progress of any data stream.
  • temporal references may be derived from.
  • the encoded segments are made available for download at a server, for example the audio server 109.
  • software operated by viewers on a client device 105 is configured to periodically update or receive notification of a new secondary audio stream being made available. The viewer is then enabled to select the new secondary audio stream for playback together with a primary video stream.
  • the software is configured to download a manifest which contains the internet address of at least secondary audio sources which are available for use. The software is then configured to use the internet addresses provided by the manifest to download the primary video and secondary audio for playback to the viewer.
  • Figure 3 outlines a process 300 as undertaken by the environmental components shown in Figure 1.
  • steps 302 - 305 occur on the secondary client device 110; and steps 306 and 307 occur on the audio server 109.
  • steps 302 - 305 occur on the secondary client device 110; and steps 306 and 307 occur on the audio server 109.
  • steps 302 - 305 occur on the secondary client device 110; and steps 306 and 307 occur on the audio server 109.
  • a primary video stream is established as a publication and received by the secondary client device 110.
  • the primary video stream is segmented as described above.
  • Time reference information is derived from the primary video stream.
  • a secondary audio source is recorded by the secondary client device 110 and
  • the client device 110 is configured to encode the secondary audio source.
  • a system clock time of the secondary client device 110, or other time reference, is determined when the video server first receives transmission of the primary video source (to).
  • the secondary client device 110 is takes a measurement of stream time (si) and system clock (ti) at some point in the future and transmits this information to the audio server 109.
  • the audio server 109 uses the difference between the secondary audio stream and the primary video source is used to determine the start of the secondary audio stream to the beginning of the next segment of the primary video stream.
  • the audio sever 109 segments the audio stream with boundaries aligned with time references taken from the primary video stream such that the secondary audio stream commences playback with temporal alignment to the next primary video segment that is due for playback after the secondary audio stream
  • steps 4 to 6 are repeated to ensure synchronisation is maintained.
  • additional audio data is generated to support of the audio feed generated by the alternative audio source.
  • the secondary audio source is shorter than the playback length of the primary video source.
  • an alternative commentary provider may generate new audio data after the primary video data has been playing for some time. It is often desirable to have a secondary audio source playback length match the playback length of the primary video source as some media codecs are naturally poor at synchronisation of audio and video data of differing playback lengths.
  • a process is implemented where audio data is added to the secondary audio data stream to increase the length of that stream to substantially match the playback length of the primary video data stream.
  • the audio server is configured to produce interim audio data segments that contain audio samples with correct metadata and codecs.
  • the audio samples may be created from the primary audio stream, derived from the primary audio stream samples by way of transformation, a selection of the primary audio stream such as one of a number of audio channels, or may be independently generated audio.
  • the resulting secondary audio stream has a playback length that matches the primary video stream, but may contain only limited audio data generated by the secondary audio data source. The resulting playback of the secondary audio data is uninterrupted and therefore minimises disruptions to a viewer playback experience.
  • audio data for backfilling is created alongside the creation of the secondary audio data stream. In this way, if the connection between the server and secondary audio source is severed for any reason, such as a network dropout, there is a segment or replacement audio data that can be readily substituted in its place for interim delivery to the viewer.
  • Substitute audio such as the primary audio data can be used to fill in the playlist before the alternative track started. In this way, if the event is played back in its entirety the viewer can hear that the alternative track hasn't started yet and that the playback is uninterrupted.
  • the substitute or backfill audio is preferably generated by or at least made available by the audio server 109 for interim use. In this way, when connection to the client device 110 is unavailable or momentarily lost, the encoded audio stream made available to the viewer is uninterrupted.
  • the video server 108 is tasked with receiving video data and audio data from a source provider 101.
  • video data from a live match and audio data that may comprise one or more audio sources including ambient noise and host commentary.
  • the video data may have audio data encoded with it such that the video data and audio data are delivered together inside a segment of data such as a segment or media container.
  • the video server 108 delivers selected video data from the source provider to a third party commentator operating a client device 110 operable to provide a secondary audio source.
  • the video server 108 also delivers selected audio data to the third party commentator separately from the video data such as ambient noise from the live match, and no primary commentary.
  • the video server 108 may be configured to strip any audio data that may be encoded together with the video data such that only video is provided to the third party commentator.
  • the third party commentator operates software that displays the video data and any audio data received from the video server 108.
  • the audio server 109 is configured to receive new audio data from the third party commentator.
  • the new audio data consists of commentary from the third party commentator only.
  • the audio server is configured to receive the new audio data comprising a mix of commentary from the third party commentator and original audio data from the secondary source provider.
  • the audio server is configured to produce new audio data comprising a mix of commentary received from the third party commentator and original audio data from the source provider.
  • the video data sourced from the sourced provider includes timestamp information.
  • New secondary audio is generated by the third party commentator is assigned timestamp information that corresponds to the video data timestamp information by determining a time when the third party commentator commences production of new audio data "start time"; determining a time difference between the start time and a time reference assigned to the video data timestamp information indicating the beginning of a new encoded segment; and, applying timestamp information to the new audio data based on the determined time difference.
  • the audio server 109 is configured to make available newly created secondary audio data to an audience. According to some embodiments, availability is facilitated by provision of a manifest containing source links for a variety of video data and audio data sources.
  • the audio data sources include new audio sources prepared by the audio server based on received third party commentary.
  • the source links are configured for selection by each audience member for download and playback.
  • the audience is provided with software that enables selection of a desired primary video data source and secondary audio data source.
  • Each audience member operates software configured to facilitate downloads and displays selected video data and audio data.
  • the video data and audio data is supplied in segments of limited length.
  • the audio client/commentator page may be configured to periodically recalculate the time offset between the secondary audio stream and the primary video stream to ensure the streams are synchronised. The recalculation may be used to compensate for factors such as drift in system clocks and random or systemic data corruption.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

L'invention concerne un système configuré pour fournir un flux audio secondaire pour la lecture de données vidéo à destination de spectateurs. Le système comprend des données vidéo qui sont divisées en segments distincts pour une lecture séquentielle et une source de données audio secondaires. Au moins un processeur est configuré pour capturer des données audio secondaires à partir de la source de données audio secondaires, recevoir les données vidéo primaires codées par l'intermédiaire du réseau de diffusion en continu, déterminer une référence de lecture temporelle pour au moins le second segment de données vidéo primaires reçu ; recevoir un flux audio secondaire provenant de la source audio secondaire ; et coder les données audio secondaires sous la forme d'une séquence de segments distincts. Le premier segment audio secondaire codé présente une référence de lecture temporelle destinée à correspondre à la référence de lecture temporelle du second segment vidéo primaire reçu.
PCT/NZ2018/050155 2017-11-03 2018-11-02 Remplacement audio en direct dans un flux numérique WO2019088853A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
NZ73698717 2017-11-03
NZ736987 2017-11-03

Publications (1)

Publication Number Publication Date
WO2019088853A1 true WO2019088853A1 (fr) 2019-05-09

Family

ID=66332150

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/NZ2018/050155 WO2019088853A1 (fr) 2017-11-03 2018-11-02 Remplacement audio en direct dans un flux numérique

Country Status (1)

Country Link
WO (1) WO2019088853A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11432035B2 (en) * 2020-07-15 2022-08-30 At&T Intellectual Property I, L.P. Adaptive streaming with demuxed audio and video tracks
CN115086708A (zh) * 2022-06-06 2022-09-20 北京奇艺世纪科技有限公司 一种视频播放方法、装置、电子设备及存储介质
US11553215B1 (en) 2017-09-25 2023-01-10 Amazon Technologies, Inc. Providing alternative live media content
EP4391559A1 (fr) * 2022-12-20 2024-06-26 Broadpeak Procede de distribution d'un contenu audio et/ou video et mandataire de transcodage mettant en oeuvre ce procede

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002097791A1 (fr) * 2001-05-25 2002-12-05 Dolby Laboratories Licensing Corporation Procede de synchronisation de signaux audio a l'aide de caracterisations fondees sur des evenements auditifs
US20050042591A1 (en) * 2002-11-01 2005-02-24 Bloom Phillip Jeffrey Methods and apparatus for use in sound replacement with automatic synchronization to images
US20120257875A1 (en) * 2008-01-11 2012-10-11 Bruce Sharpe Temporal alignment of video recordings
US20140022456A1 (en) * 2011-06-17 2014-01-23 Echostar Technologies L.L.C. Alternative audio content presentation in a media content receiver
US20170201793A1 (en) * 2008-06-18 2017-07-13 Gracenote, Inc. TV Content Segmentation, Categorization and Identification and Time-Aligned Applications

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002097791A1 (fr) * 2001-05-25 2002-12-05 Dolby Laboratories Licensing Corporation Procede de synchronisation de signaux audio a l'aide de caracterisations fondees sur des evenements auditifs
US20050042591A1 (en) * 2002-11-01 2005-02-24 Bloom Phillip Jeffrey Methods and apparatus for use in sound replacement with automatic synchronization to images
US20120257875A1 (en) * 2008-01-11 2012-10-11 Bruce Sharpe Temporal alignment of video recordings
US20170201793A1 (en) * 2008-06-18 2017-07-13 Gracenote, Inc. TV Content Segmentation, Categorization and Identification and Time-Aligned Applications
US20140022456A1 (en) * 2011-06-17 2014-01-23 Echostar Technologies L.L.C. Alternative audio content presentation in a media content receiver

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11553215B1 (en) 2017-09-25 2023-01-10 Amazon Technologies, Inc. Providing alternative live media content
US11432035B2 (en) * 2020-07-15 2022-08-30 At&T Intellectual Property I, L.P. Adaptive streaming with demuxed audio and video tracks
CN115086708A (zh) * 2022-06-06 2022-09-20 北京奇艺世纪科技有限公司 一种视频播放方法、装置、电子设备及存储介质
CN115086708B (zh) * 2022-06-06 2024-03-08 北京奇艺世纪科技有限公司 一种视频播放方法、装置、电子设备及存储介质
EP4391559A1 (fr) * 2022-12-20 2024-06-26 Broadpeak Procede de distribution d'un contenu audio et/ou video et mandataire de transcodage mettant en oeuvre ce procede
WO2024132841A1 (fr) * 2022-12-20 2024-06-27 Broadpeak Procédé de délivrance d'un contenu audio et/ou vidéo et mandataire de transcodage mettant en œuvre ledit procédé

Similar Documents

Publication Publication Date Title
US10313758B2 (en) Scheduling video content from multiple sources for presentation via a streaming video channel
US10785508B2 (en) System for measuring video playback events using a server generated manifest/playlist
US10911512B2 (en) Personalized content streams using aligned encoded content segments
US8973032B1 (en) Advertisement insertion into media content for streaming
US10397636B1 (en) Methods and systems for synchronizing data streams across multiple client devices
US11070872B2 (en) Receiving device, transmitting device, and data processing method
CN104885473B (zh) 用于经由http的动态自适应流式传输(dash)的实况定时方法
US8643779B2 (en) Live audio track additions to digital streams
US9792363B2 (en) Video display method
EP2853075B1 (fr) Identification spécifique du contenu et fonctionnement de synchronisation dans la diffusion en continu adaptative sur protocole de transfert hypertexte
WO2019088853A1 (fr) Remplacement audio en direct dans un flux numérique
US10114689B1 (en) Dynamic playlist generation
US20090106357A1 (en) Synchronized Media Playback Using Autonomous Clients Over Standard Internet Protocols
US9197944B2 (en) Systems and methods for high availability HTTP streaming
US20160358597A1 (en) Real Time Popularity Based Audible Content Acquisition
PH12014502203B1 (en) Enhanced block-request streaming system for handling low-latency streaming
TR201810487T4 (tr) İşaretleyici tabanlı varış noktaları arası ortam senkronizasyonu.
KR20120069749A (ko) Url 템플릿들 및 구성 규칙들을 이용하는 향상된 블록-요청 스트리밍
KR20120069748A (ko) 개선된 클라이언트 측 핸들링을 위한 요청 제어들 또는 블록 분할을 이용하는 강화된 블록 요청 스트리밍
KR20120069746A (ko) 확장가능한 인코딩을 이용하는 개선된 블록-요청 스트리밍
RU2656093C2 (ru) Устройство поставки контента, способ поставки контента, программа, оконечное устройство и система поставки контента
WO2018134660A1 (fr) Système et procédé de diffusion de contenu en continu
EP4111700A1 (fr) Identification d'éléments dans un groupe pour remplacement d'élément dynamique
US10652292B1 (en) Synchronization of multiple encoders for streaming content
US11856242B1 (en) Synchronization of content during live video stream

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18874539

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 24.08.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18874539

Country of ref document: EP

Kind code of ref document: A1