WO2016106360A1 - Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators - Google Patents

Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators Download PDF

Info

Publication number
WO2016106360A1
WO2016106360A1 PCT/US2015/067464 US2015067464W WO2016106360A1 WO 2016106360 A1 WO2016106360 A1 WO 2016106360A1 US 2015067464 W US2015067464 W US 2015067464W WO 2016106360 A1 WO2016106360 A1 WO 2016106360A1
Authority
WO
WIPO (PCT)
Prior art keywords
multimedia
content
format
virtual machine
content provider
Prior art date
Application number
PCT/US2015/067464
Other languages
French (fr)
Inventor
Edwin A. Hernandez-Mondragon
Original Assignee
Hernandez-Mondragon Edwin A
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=56151532&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO2016106360(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Hernandez-Mondragon Edwin A filed Critical Hernandez-Mondragon Edwin A
Priority to PL15874332.8T priority Critical patent/PL3238457T3/en
Priority to US15/538,911 priority patent/US10123074B2/en
Priority to DK15874332.8T priority patent/DK3238457T3/en
Priority to EP15874332.8A priority patent/EP3238457B1/en
Priority to FIEP15874332.8T priority patent/FI3238457T3/en
Publication of WO2016106360A1 publication Critical patent/WO2016106360A1/en
Priority to US16/152,606 priority patent/US10524002B2/en
Priority to US16/729,986 priority patent/US11140441B2/en
Priority to US17/493,490 priority patent/US20220030306A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440218Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234309Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4 or from Quicktime to Realvideo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/238Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
    • H04N21/2387Stream processing in response to a playback request from an end-user, e.g. for trick-play
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6118Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving cable transmission, e.g. using a cable modem
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6143Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via a satellite
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • H04N21/8133Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/64Addressing
    • H04N21/6405Multicasting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/643Communication protocols
    • H04N21/64322IP

Definitions

  • the present disclosure provides a mechanism for delivering multimedia content from a cloud-based system to cable or satellite content providers.
  • the disclosure describes a novel method and a system for continuous media delivery and secured synchronization of multimedia files that are broadcasted to set top boxes or subscriber units in the field that are connected to IPTV or Cable Systems.
  • the present technology can securely connect to a cloud-based platform (e.g. OpenStack, Amazon EC2, Mediamplify Cloud, etc.), creates a set of multimedia files (e.g. MPEG4 movies) into a secured repository (e.g. Secured File System, Linux/Windows, etc.), modifies those files according to certain rules as defined by the method herein, and reproduces those signals via a multicast video stream (e.g.
  • UDP User Datagram Protocol
  • TS Transport Stream
  • TS Transport Stream
  • MSO Multiple System Operators
  • DSS Digital Satellite Systems
  • RF Radio Frequency
  • IRD Integrated Receiver and Decoder
  • ASI Asynchronous Serial Interface
  • MSOs and DSSs use dedicated hardware encoders and transcoding systems if the source video/audio is incompatible with the set top boxes of the cable or satellite operator.
  • the video/audio content delivered via the satellite to the cable or satellite operator arrives via wireless link or receiver, which is processed by an IRD that receives the stream from the satellite and can ultimately broadcast that signal to the cable operator head end in several formats.
  • some systems may include multicast units connected with IRDs that tune to the right transponder system.
  • Satellite systems are used to deliver data (e.g. TV, radio, Video-on-Demand, multimedia streams) from a particular source (e.g. third-party content provider) to cable head end systems or other satellite systems. Satellite delivery is performed via a broadcast signal containing digital pictures and audio inside a transponder frequency. Transponders are radio frequency (RF) space allocated in a particular satellite.
  • RF radio frequency
  • a third-party content provider can maintain multimedia files in CD/DVDs, prerecorded storage, or as pre-recorded files that are reproduced in a particular audio or video encoding format using a video or audio encoder (e.g. MP3, H.264, AC3, and others).
  • the audio and video is then made part of an RF signal that is broadcasted to the satellite system.
  • This signal is then received by the cable or satellite operator using an antenna receptor and is decoded to generate either an analog signal (e.g. RCA Cables or any other analog interface) or digital signal (e.g. ASI, Asynchronous Serial Interface).
  • an analog signal e.g. RCA Cables or any other analog interface
  • digital signal e.g. ASI, Asynchronous Serial Interface
  • DVB Digital Video Broadcasting
  • Cable and Satellite systems will encode this signal to their particular video/audio encoding requirements such that it can be broadcasted in digital packets via a multicast MPEG Transport Stream.
  • a receiver decodes the signal in ASI.
  • ASI Serial Bus Interface
  • this signal is converted from ASI to Ethernet, or simply redirected to a multicast MPEG Transport stream if arriving in a compatible encoding format for the receiving head end cable system.
  • IPTV systems can provide mechanisms to broadcast a signal using the multicast protocol in MPEG transport streams to a customer premises with compatible set top boxes (same encoder, same streams, encryption keys, etc).
  • a compatible set top box is such that when connected to the cable TV system it can properly decode video/audio from the cable operator' s head end system.
  • Some of these IPTV systems rely on CDN (Content Delivery Networks) that enable distribution of multimedia files to IP-enabled devices.
  • CDN Content Delivery Networks
  • a mechanism can be provided whereby a head end content (MSO) that is being broadcasted to a cable subscriber is also available to an IP-enabled device such as a smartphone, PC, or tablet computer.
  • MSO head end content
  • CDN Content Delivery Network
  • MSO Multiple System Operators
  • head-ends that include cable and satellite delivery mechanisms.
  • the present technology unifies cloud-based delivery (Mediamplify cloud) with the cable-based mechanism (e.g. Comcast, Verizon FiOS).
  • An example computer-implemented method can receive, from a content provider such as a cable or satellite operator, a request for at least one media stream for playback on a broadcast media channel.
  • the request for the at least one media stream can include a request for a plurality of multimedia items.
  • Content corresponding to the plurality of multimedia items can be obtained from at least one website or cloud service that offers the content in at least one first format.
  • the plurality of multimedia files can be generated in a second format that is compatible with the content provider.
  • the media stream can assembled using each of the multimedia items generated in the second format.
  • the media stream can be provided to the content provider for broadcast on the broadcast media channel.
  • obtaining the content can include retrieving at least one audio file corresponding to an audio component of the multimedia item and retrieving a plurality of images as screen captures that can correspond to a video component of the multimedia item.
  • Generating the multimedia files in the second format can include combining the plurality of images and the at least one audio file to create each of the multimedia items.
  • FIG. 1 shows an example block diagram illustrating an exemplary head end (MSO);
  • FIG. 2 shows an example block diagram of the apparatus located at the MSO facility (mediaplug);
  • FIG. 3 shows an example method for initialization of Mediaplug components
  • FIG. 4 shows an example method to obtain list of files to multicast to a cable operator MSO
  • FIG. 5 shows an example method for encoding video and audio files
  • FIG. 6 shows an example method for providing multimedia content to a cable operator
  • FIG. 7 shows an example block diagram of example software components of the Mediaplug
  • FIG. 8 shows an example method to multicast from a component in the Mediaplug
  • FIG. 9 shows an example method to insert metadata into in an MPEG Transport Stream
  • FIG. 10 shows an example block diagram for a Virtualized Mediaplug
  • FIG. 11 shows an example method to initialize virtual mediaplug components
  • FIG. 12 shows an example possible system embodiment for implementing various embodiments of the present technology
  • FIG. 1 illustrates a block diagram that includes aMediamplify Cloud 101.
  • the cloud 101 can refer to a set of virtualized instances or non-virtualized instances of one or more servers that provide access to archives and digital files with media content stored in them.
  • the cloud 101 can be implemented using VMWave or Xen and OpenStack, or simply reside in Amazon Web Services or Rackspace.
  • special provisioning can be provided in cloud 101 for secured access.
  • the cloud 101 service files can be accessible via the internet 105 and can be stored in a cloud-based file system or cluster file system (e.g. GlusterFS or Hadoop, Amazon S3). These files can be multimedia files that include, as non-limiting examples, MPEG-4 Video files with H.264 encoded video and AAC encoded audio or MPEG-layer 3 audio as well (MPEG2 Video could also be used). These files can be accessed using a secured transport protocol such as, but not limited to, SSL or a synchronization tool as "rsync," commonly found in Linux/MacOS environments.
  • the communication link 106 can be a DSL, Cable Modem, Wireless WAN, Fiber, Metro Ethernet, etc. and can be linked with the MediaPlug 110 unit.
  • the Mediaplug 110 can reside at the "MSO" facility and can be collocated with all the cable or satellite operator equipment.
  • the MSO is a facility used by the Cable Operator to receive all TV and Radio Programming that the cable operator will then multiplex, re-encode, and distribute from the MSO to a set of subscriber units with set top boxes, or modulate the signal in TV analog channels (2-60+) .
  • One function of the MediaPlug 110 is to provide multimedia content, e.g., video and audio, to an MPEG multiplexer at the headend 120.
  • the content can be pre-formatted in MPEG Transport Streams that the multiplexer 120 (e.g. Arris TMX-2010) will then be able to capture, reformat in any way the operator requires, and deliver to the cable operator's subscribers via Coaxial Cable 125 or any other means suitable for a multicast or unicast delivery to its customers with "Set Top boxes" 130, 131, 132.
  • a "Set Top Box” is customer premise equipment that will be able to decode and display in a TV set or other screen display, via a signal received by its HDMI or analog input, the contents of the TV programming the cable operator offers.
  • the Mediaplug 110 can be a computer or server or set of computers of servers that can communicate with the Cloud 101 using a standard Local Area Network (e.g. Gigabit Ethernet).
  • the Mediaplug 110 can store, encode, and transmit multimedia files in the format that the cable operator at its MSO requires. This cable format may include audio, video, and metadata (subtitles, control data, and any additional component that can be packaged in an MPEG Packet that is part of the MPEG Transport Stream).
  • Mediaplug 110 can include components such as a caching unit, and one or more multicasting servers.
  • FIG. 2 illustrates an example configuration of the hardware components of Mediaplug 110 which includes three servers.
  • a server is a computer device, usually rack mountable unit with volatile memory such as RAM, one or more micro-processors, and non-volatile memory such as a hard disk.
  • RAM random access memory
  • micro-processors such as RAM
  • non-volatile memory such as a hard disk.
  • FIG. 2 illustrates Mediamplify Cloud 202 which is connected to the Internet backbone 203.
  • Link 204 e.g. DSL, Cable Modem, Fiber
  • a Network Address Translator unit 205 e.g. Cisco RVS4000, or any other NAT device
  • the lines 206, 207, and 208 show Gigabit Ethernet links, however one skilled in the art will understand that Gigabit Ethernet or a Wireless LAN using 802.1 lg/n would be also a suitable interface to be used in a Mediaplug configuration.
  • the Caching unit 210 can be the main storage component in the system and it can contain all of the multimedia files in the format required by the cable operators, such as the Transport Stream format.
  • the caching unit can communicate with Multicast-01 215 and Multicast-02 220.
  • the two Multicast units 215, 220 can pull or obtain information from the caching unit via TCP/IP or UDP, using protocols such as HTTP (Hypertext Transport Protocol) or RTSP (Real-time Streaming Transport).
  • HTTP Hypertext Transport Protocol
  • RTSP Real-time Streaming Transport
  • the caching unit can also encode and transcode the multimedia files it downloads from the Cloud 202.
  • the caching unit can also keep updated information of all the metadata and create customizable screens that can be converted into video formats compatible with the Transport Stream content that will be provided to the cable operator at its MSO.
  • MPEG MUX MSO 230 can correspond to a program residing in a computer system that processes and handles the MPEG streams from multiple sources.
  • the MUX 230 can also function as the bridge between the source of the signal and the customer's set top box units.
  • a non-limiting example program would be the CherryPicker DM6400 by MOTOROLA.
  • Each of the components or modules of the Mediaplug can include software for performing functions and/or methods according to the present technology.
  • FIG. 3 illustrates an example method that can be used to configure and provision the units at boot-time.
  • Each unit can initialize keys for SSH/IP Sec access at 305.
  • the units can use IPSec and SSH Tunnel keys that can be stored in the administrator user account.
  • the keys can be generated by the Mediamplify Cloud that are private keys or certificates, shared secrets or similar components by which the Mediaplug can connect securely with the Mediamplify cloud.
  • Each cable operator can provide custom-based IP addressing at 310 for the Gigabit Ethernet ports that connect to the operators' network. Changing the IP addresses of the Multicast-01 and/or Multicast-02 refers to the IP address by which the links 221 and 216 in FIG. 2 are linked to the MPEG MUX MSO unit.
  • a change of the IP address can refer to the 'ifconfig' command which is used in Linux-based systems and that can be changed in an Ubuntu box at /etc/network/interfaces file.
  • the network routing can be changed if the multicast address cannot be routed to the default gateway which points to the 205 NAT, but to 221 and 216 respectively for Multicast-02 and Multicast-01.
  • the 239.*.*.* network address may be routed to those interfaces.
  • the Mediaplug can configure playlists per channel and multicast address.
  • the playlist and channel configuration can be obtained from the cable/satellite operator and can include any number of channels for broadcasting multimedia content to end customers.
  • the channels may include music channels corresponding to different music genres, e.g. rock, alternative, classical, jazz, hip hop, etc. Based on the particular genre, a corresponding playlist can be selected and/or configured.
  • the Mediaplug can also determine the multicast address for each of the channels.
  • a customizable user interface can be added to the output video MPEG Transport stream.
  • UI user interface
  • the custom video interface can be created and fetched as needed or it can be created in real-time for a particular channel or for all the channels.
  • the custom user interface variable can be uniquely configured per channel in the Mediaplug.
  • the mediaplug can also monitor 340 all of the processes to ensure stability and continuous, uninterrupted operation.
  • the initialization of the Mediaplug can also include setting the RSYNC profile 350.
  • the RSYNC profile can be used to update and synchronize media content with the cloud when the Mediaplug determines that the content has changed or been modified.
  • the Mediaplug can broadcast the multimedia MPEG stream using the programmed IP Address in multicast-01 and multicast-02 according to the method illustrated in FIG. 3.
  • FIG. 4 an example method is shown for obtaining the list of files to playback or broadcast to the MSO system.
  • the file type or format requested by the multicasting embodiment e.g. multicast-01 or multicast-02, can include Transport Stream (e.g. TS) and/or MPEG-4 Part 14 (e.g. MP4) values 405.
  • This request is executed from any of the multicasting embodiments, usually making use of the HTTP GET Request.
  • the multicasting unit can connect to an HTTP Server and retrieve the list of files based on the type identified by the request.
  • each channel can have a unique identifier corresponding to the name of the playlist.
  • channel name "Cool” or "Rock” 410 are names that can identify the channel to be retrieved from the caching server.
  • the activity log in the request can also be updated keeping track what step of the process is being completed 415.
  • the caching server can generate a "playlist file” of mime type "audio/x-mpegurl” which will be interpreted as a multimedia playlist file when requested.
  • the playlist can be kept in the local memory caching service 425.
  • the "Caching” unit can be configured to perform parallel encoding for all the different channels available from the unit.
  • FIG. 5 illustrates an example method that can be implemented by the Mediaplug, and in particular, by the "Caching" unit.
  • encoding flags or input parameters can be received which may include GOP (Group of Pictures) which is used by certain video encoders, resolution (e.g. 640x480) in pixels, bit rate in Kilobits per second, any encoding type of video (e.g. H.264), encoding type of audio (e.g. AC-3), duration of the video feed, etc.
  • GOP Group of Pictures
  • a "format" parameter can be received which refers to the output format or encapsulation, MPEG TS or MPEG4.
  • the caching unit encodes audio and video, and can use the disk space in the Mediaplug.
  • the method checks for sufficient space left in the memory or hard drive. This task can be run on a thread together with all the other tasks that are working in parallel in order to speed up the processes of encoding and transcoding.
  • indexes or identifiers by the name "ChID,” (these integers preferably no more than an integer value from 1 to the maximum number of channels) begin (e.g. MAX could be 50) to be generated by the caching unit 510.
  • This method can be performed for all the channels and all media linked to that particular channel identified by "MedialD” 515, in this particular case the "MedialD.”
  • the variable “MAX” in this context can refer to the maximum number of media files read (e.g. 500).
  • the data captured from the media could be "Artist,” “Author,” or any other variable from an audio file. If a video file is used this could be "program name,” “movie name,” or any associated metadata information tied to the multimedia file and the Channel Identifier (ChID) associated with it "MedialD” file.
  • the caching unit can add a "Custom UI," or a customizable User Interface that is designed using Hypertext Markup Language (e.g. HTML) or a web service.
  • a web service that matches a REST interface (Representation State Transfer) and uses Javascript Object Notation (JSON) or XML (Extensible Markup Language) to respond to a custom UI request can be used.
  • this custom UI can include JavaScript, CSS, and HTML content files that can be rendered using a browser using WebKit (e.g. PhantomJS or Safari) or any other browser-rendering engine compatible with HTML4/5 or any future HTML version.
  • WebKit e.g. PhantomJS or Safari
  • rendered images from a web service can be used to create overlapping video or replace any video in the video feed file.
  • this video or customized screen can be used when no video at all is defined in the multimedia file, e.g. Audio-only file.
  • the web-service to be contacted may use the "PlayListName" to be used for the Channel identified by the channel ID "ChID, " retrieved using the "PlaylistID” function 520.
  • Each new PlayListName may contain different identifies to retrieve new images. Therefore one or more screens will be retrieved from the rendering engine as such 522 can be used to retrieve a set of screens from the web-service being contacted.
  • Capturing screens 526 can be done using a tool similar to PhantomJS (Webkit-based headless engine for rendering images) or QtCapture, or simply capturing screen frame buffers from a machine running any software tool.
  • the method can capture a video file from another rendering function, which can create a sequence of screens that can be used to generate a video file. Once all screens and video are captured, 530 the resulting output created can be a video file that then is either replaced or merged (multiplexed) with the original file.
  • an audio content file such as an MP3 file
  • this function can create a video file from all the screens captured by 526. Those screens are generated by calling a URL where the web-service is associated and creating a video file with the input audio and the captured screens.
  • the output format can be an MPEG Transport Stream file that can be retrievable or streamed to the multicasting embodiment.
  • the audio can be multiplexed and added to one of the encoding channels of the resulting video output and saved into the caching unit storage disk 550.
  • the multiplexing can be done using tools such as FFMPEG or VLC or any proprietary tool that can multiplex audio and video files into a resulting MPEG Transport Stream (or MPEG-4 output, or any similar output).
  • FIG. 6 illustrates an example method 600 for providing multimedia content to a cable operator.
  • a device such as the Mediaplug device receives a request from a cable/satellite operator for a media stream.
  • a media stream can be formatted according to HTTP Live Streaming (HLS), an HTTP/RTSP playlist, or an RTSP stream.
  • the media stream can include any number of multimedia items such as audio/video multimedia items.
  • the request can include a channel identifier that is associated with a particular category of multimedia items, such as a music genre corresponding to rock, classical, jazz, etc.
  • contents corresponding to the multimedia items can be retrieved, downloaded, or otherwise obtained from one or more websites or cloud based servers.
  • the contents can include data corresponding to either or both of the audio and video components of the multimedia items.
  • the data corresponding to the audio portion can include a music file in an MP3 format.
  • the data corresponding to the video portion can include a video that is obtained directly from the cloud.
  • the video component can be assembled by using one or more screen captures from a website.
  • the screen captures can include captures of still images presented by the website as well as captures of images from videos that are available on the website. As illustrated in FIG. 5, the screen capture process can be repeated any number of times.
  • the Mediaplug can then generate 606 multimedia items in a second format based on the content obtained from the web server or cloud. That is, the Mediaplug can convert/encode the audio file into a format that is suitable for the cable operator and it can also assemble the various screen captures to generate the video component of the media item.
  • the audio and video files can be combined, multiplexed, or encoded to a format that can be utilized by the multimedia stream. For example, the audio can be multiplexed and added to one of the encoding channels of the video output.
  • the various multimedia items can be used 608 to assemble the media stream in a format corresponding with the request (e.g. HLS, HTTP/RTSP, RTSP stream).
  • the media stream can be provided 610 to the content provider for broadcast on the broadcast media channel.
  • FIG. 7 illustrates a block diagram of an example architecture of the Caching Unit components.
  • the caching unit can contain two types of operating system services HTTP 710 and RTSP 712.
  • the RTSP service provides access so the media saved or created in the repository of the caching unit file system 730.
  • the RTSP server can be implemented using the Live555 RTSP server or similar.
  • the HTTP Server can be implemented using "nginx" or "apache” web browsers, and the RTSP Servers can be implemented using Live555 services. In the case of the HTTP Server, this can also be used to get the Playlist of files generated from the method on FIG-5 715.
  • the GetPlayList 715 command can be implemented as follows:
  • HTTP header as M3U or PLS (e.g. a text file with a list of files or URLs to play), • Return a list of HTTP or RTSP URLs to access each video generated or stored in the caching unit.
  • M3U or PLS e.g. a text file with a list of files or URLs to play
  • the HTTP Server can also be used to clean up, update, and delete operations.
  • the block unit called “CleanUp” 722 is in charge of keeping up to date caching information data and/or deleting files already used or that are not required to be used anymore by the multicast unit(s).
  • the HTTP Server functions can include servicing M3U8 files corresponding to the HTTP Live Streaming format and all the segments created by the HTTP Live Streaming function.
  • a HTTP Live Streaming (HLS) generation program will produce the files and segment them as needed by the encoder and can then be serviced by the HTTP Server 710.
  • the Mediamplify Parallel Encoder 705 corresponds to a block that can implement the method shown in FIG 5.
  • the resulting outputs from the Media Parallel Encoder 605 are stored in the file system 730.
  • the Mediamplify Parallel encoder 605 can be implemented in Python and can rely on web services (such as Twisted or Django), PhantomJS, and FFMPEG.
  • Monit 750 correspnds to a monitoring block that can ensure that the process running the "Media Parallel encoder" keeps running at all times and does not stop upon failure. Additional blocks that can be included in Mediaplug are RSYNC 742 and SSH 744. Thes blocks can be used to ensure synchronization and updates of new media files coming from the cloud.
  • the RSYNC block uses the "rsync” protocol and program over ssh (Secured Socket Layer Shell) to download media from the "Mediamplify Cloud” and synchronize multimedia files that are newly updated and could be part of particular Channel ID and a Playlist Name.
  • ssh Silicon Socket Layer Shell
  • Each new file synchronized with the Mediamplify cloud has a Media ID value which the cloud can assign, and include additional metadata information such as "Artist” and "Author.”
  • the "Mediamplify Parallel Encoder” also may receive requests from other components such as getStatus or StartEncoding, indicating what file is being encoded, its progress, and what other files are in the queue for encoding.
  • the information from getStatus or StartEncoding can be generated using a JSON message retrieved upon a request via HTTP GET to the Mediamplify Parallel Encoder.
  • a graphical user interface can then be used to plot the status of each song being encoded or provide previews of the videos being generated.
  • Port 80 (as an example), or the standard HTTP port, used to retrieve content, view video content generated, update/delete cached files, and access the playlists
  • Port 9553 (as an example), or the standard Mediamplify Parallel encoder port, used to monitor and review status of the songs being encoded, the data being collected, and all other related activities with encoding. Via this port, encoding/transcoding can be started or stopped.
  • the "Mediaplug" components labeled “Multicast-01” and “Multicast-02” can also provide an HTTP interface to access the current song being played or being multicast via the HTTP Server 770.
  • the HTTP Server may run a PHP or Python script that will return the current song being played by the streamer.
  • the streamer block 768 can connect to the caching server via HTTP or RTSP protocols. The implementation of this block can be done using a combination of FIG-8 instructions and using multicasting services available in VLC or FFMPEG (e.g. multicast, or tstools / tsplay).
  • One function of the streamer 768 is to read the files encoded by the caching unit, process those for multicasting and creating the MPEG Transport Stream multicast packets sent via the Ethernet port. Additionally, the multicast-01 and multicast-02 blocks contain a "Recording" function 790, which is in charge or recording the multicasting streams and recording for later use in case of failure on the system. This operates by listening into the UDP (User Datagram Protocol) multicast address of each individual channels being broadcasted, and recording the output into a file. This file is then stored and made accessible for future use. Monit 764 can maintain the streamer process running at all times throughout a regular broadcasting cycle. [0052] FIG.
  • the streamer can perform functions that include: connect to the caching unit; retrieve generated videos per playlist; determine if unit is working in fault-tolerant mode; playback pre-recorded audio if a failure is detected; and record units own output or any other streamer output if connected to switch or hub.
  • the streamer reads the multicast address associated to the channel being broadcasted, in order to generate IGMP (Internet Group Management Protocol) messages and establish a multicast group (e.g. 239.0.0.1:8001).
  • the channel name is also initialized at 810. This value can be used to retrieve the proper content from the Caching server.
  • a configuration table can be created with tuples: ⁇ ChID>, ⁇ Channel Name>, ⁇ Multicast Address>, ⁇ Multicast Port>, such that no duplicates are used or created. Automation software can create these configuration tables and automate the process of creating all streamer calls.
  • a streamer method can be called (e.g.
  • the streamer will check 850 if there are already multicasting transport streams generated by any other multicast component and avoid collisions. If such multicast data is detected, a delay can be used to delay a new attempt to check on that particular multicast group 860.
  • this block indicates "Wait ⁇ Time>" or stop all processing for the amount of ⁇ Time> recommended by the specification.
  • the time for delay can be the "Main source" of input less the amount of time for receiving input from a backup server.
  • the streamer will check if the Caching server is "alive" 815. This can be performed in multiple ways, for example an HTTP Request must not generate a 404 or 403 error, but instead a successful 202 value. Likewise, ping requests can also be used before making an HTTP GET call to identify if the caching server is present and available. In the case that there are no ping responses, a timeout will trigger a failure and a pre-recorded content 820 will be played or sent to the multicast address and port of record. Recordings can be scheduled daily for an extended period of time (e.g. 8 hours or more). These recordings can store the recorded file in a temporary folder and then move that file over to the playback folder where streamer can read it.
  • an HTTP Request must not generate a 404 or 403 error, but instead a successful 202 value.
  • ping requests can also be used before making an HTTP GET call to identify if the caching server is present and available.
  • a timeout will trigger a failure and a pre-
  • HTTP Live Streaming a multi-segment stream usually defined by extension M3U8, containing a set number of segments for a real-time stream source (Video/audio)
  • HTTP/RTSP file playlist this means that one or many files are read from the caching server, usually defined by extension M3U or PLS
  • RTSP stream this could be constant stream similar to HTTP Live Streaming (HLS) but using an RTSP protocol instead of HLS stream.
  • HLS HTTP Live Streaming
  • the bit-rate may vary between playlist files. Different bit-rates among files in a playlist could cause problems in non-CBR (Constant Bit Rate) systems.
  • the Mediaplug can maintain the stream at the bit rate of the previous stream by adding null MPEG TS packets at the same rate as the previous media file was being sent over the multicast port. This will maintain appropriate bit-rate for play lists.
  • RTSP service streaming occurs by writing the media content to a named pipe or a FIFO file and allowing the RTSP service to read from that file and generate an RTSP stream.
  • the caching server knows the filename of the file being streamed.
  • both the caching and multicast unit know the name of the file being streamed, as the multicast server can then request from the caching server the metadata associated with the file name being streamed.
  • the present technology can accommodate both cases and retrieve the current file name being streamed as a JSON response.
  • the response can identify the content displayed or played back at the end-user's set top box.
  • HLS HLS
  • the streamer will connect to the proper M3U8 file 820 to pull all the segments corresponding to the live playlist file on the M3U8 stream.
  • Programs and libraries such as FFMPEG or VLC can then be used to encapsulate the stream feed to an MPEG TS and multicast it to the pre-programmed multicast address and port.
  • an RTSP stream 825 is used, the same process takes place at 830 using the RTSP as a protocol instead of HTTP.
  • a list of files in a playlist 832 is in use "M3U,” the list can be obtained using an HTTP GET request made to the caching server. Each file can then be streamed to the multicast address and port of record 838 until all elements of the playlist are completed and played back 835.
  • Metadata can be added to the stream as an additional PID (Program Identifier) in the MPEG Transport Stream. If metadata is needed 841 then it will be inserted 842 here in accordance with FIG. 9.
  • PID Program Identifier
  • the metadata added could be as simple as a string containing Artist, Name, or other information, or as complex as DCII Text (DigiCipher version 2, as specified by Motorola), MPEG TS Control messages (Two-way), or any other metadata or PID that can be inserted into the stream being sent to the multicast address and port.
  • DCII Text DigiCipher version 2, as specified by Motorola
  • MPEG TS Control messages Two-way
  • FIG. 9 illustrates an example method used to insert a special data packet into the stream being written to the multicast address, without making use of timing information. Therefore, the metadata can be inserted such that it does not have timing dependency. Alternatively, certain type of metadata such as subtitles or lyrics may need to be inserted only at certain times so that it is synchronized with the video data.
  • the sample MPEG TS packet shown in 902 and 905 corresponds to the metadata inside the packet. This matches standard MPEG TS packet structure and may vary depending on the type of metadata being added to the stream.
  • the method to insert the packet starts by reading the MPEG TS file retrieved from the caching server 910. If metadata corresponds to only song information, this information can be added at uniform times. Non-uniform timing inclusions can also be done by adding a table that includes the time and the metadata content added to the stream 915. This metadata can be retrieved from the Caching unit using an HTTP Get request directed to the caching server 918 and may include artist name, title of song, or any other information 920. [0069] Creation of the MPEG Packet(s) 925 to be inserted at the appropriate times, which will set the places with those packets, will be inserted in the MPEG TS. Modify the PMT (Program Management Table) of the generated file and include the PID being inserted into the MPEG TS 930. At step 932, an audio PID may include a metadata PID of type DCII Text, which is added to the stream.
  • PMT Program Management Table
  • the method can insert those packets and generate a buffer that can be sent to the multicast address as in illustrated in FIG. 8.
  • the disclosed technology can be implemented using a set of modules included in the Mediaplug apparatus, which can be implemented using one or more servers accordingly.
  • virtualization environments such as VMWare or Xen (FIG-9) can include a system and architecture that can be used to host all the Mediaplug functionality in a virtualized environment using a single server with multiple GbE ports.
  • FIG. 10 shows a configuration of three virtual machines used for implementing the Mediaplug.
  • the multicast Ethernet ports 1024 and 1014 are bridged with ethl 1004 and eth2 1003 respectively via bridge 1009 and bridge 1008, respectively. That is, the same content generated at 1024 can be forwarded via the bridge 1009 to the physical interface ethl 1004 in the hardware.
  • the vSwitch 1045 or the virtual switch between the two bridged interfaces 1008 and 1009 permits the two virtual machines of Multicast-01 1010 and Multicast-02 1020 to record each other and monitor their activity.
  • the Mediaplug can use the example method shown in FIG. 11. Once a create unit command is issued 1105, provisioning of the metadata information and proper IP Address is provided for that particular instance 1110. Once an IP Address and metadata has been chosen for the unit, the server can be allocated in the cloud and the instance can be created using CloudStack or OpenStack tools and services 1115. The instance is then allocated and runs in the proper node in the cloud 1120. Configuration of the assigned IP, Netmask, and VPN credentials are loaded at 1125. Finally, certificate information is uploaded 1130 as well as all services run to enable the unit for execution of its caching functions, multicasting functions, or both.
  • FIG. 12 illustrates a conventional system bus computing system architecture 1200 such as can be used in Mediaplug 110.
  • Those of ordinary skill in the art will appreciate that other system architectures are also possible.
  • Example system 1200 includes a processing unit (CPU or processor) 1210 and a system bus 1205 that couples various system components including the system memory 1215, such as read only memory (ROM) 1220 and random access memory (RAM) 1225, to the processor 1210.
  • the system 1200 can include a cache of high-speed memory connected directly with, in close proximity to, or integrated as part of the processor 1210.
  • the system 1200 can copy data from the memory 1215 and/or the storage device 1230 to the cache 1212 for quick access by the processor 1210. In this way, the cache can provide a performance boost that avoids processor 1210 delays while waiting for data.
  • the processor 1210 can include any general purpose processor and a hardware module or software module, such as module 1 1232, module 2 1234, and module 3 1236 stored in storage device 1230, configured to control the processor 1210 as well as a special-purpose processor where software instructions are incorporated into the actual processor design.
  • the processor 1210 may essentially be a completely self-contained computing system, containing multiple cores or processors, a bus, memory controller, cache, etc.
  • a multi-core processor may be symmetric or asymmetric.
  • an input device 1245 can represent any number of input mechanisms, such as a microphone for speech, a touch- sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth.
  • An output device 1235 can also be one or more of a number of output mechanisms known to those of skill in the art.
  • multimodal systems can enable a user to provide multiple types of input to communicate with the computing device 1200.
  • the communications interface 1240 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
  • Storage device 1230 is a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 1225, read only memory (ROM) 1220, and hybrids thereof.
  • RAMs random access memories
  • ROM read only memory
  • the storage device 1230 can include software modules 1232, 1234, 1236 for controlling the processor 1210. Other hardware or software modules are contemplated.
  • the storage device 1230 can be connected to the system bus 1205.
  • a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 1210, bus 1205, display 1235, and so forth, to carry out the function.
  • example system 1200 can have more than one processor 1210 or be part of a group or cluster of computing devices networked together to provide greater processing capability.
  • a software module can be software that resides in memory of a client device and/or one or more servers of a content management system and perform one or more functions when a processor executes the software associated with the module.
  • the memory can be a non-transitory computer-readable medium.
  • the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like.
  • non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
  • Methods according to the above-described examples can be implemented using computer- executable instructions that are stored or otherwise available from computer readable media.
  • Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network.
  • the computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
  • Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
  • the instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.

Abstract

Systems, methods, and computer-readable media for delivering multimedia content from the cloud to cable operators are disclosed. A device located at the cable headend or implemented in the cloud can receive a request for at least one media stream for playback on a broadcast media channel. Content corresponding to a plurality of multimedia files in the media stream can be obtained from the internet or a cloud based service. The content can be used to generate the multimedia files in a format that is compatible with the cable operator. The multimedia files can be used to assemble the at least one media stream which can be provided to the cable operator for broadcast on the broadcast media channel.

Description

METHOD, SYSTEM, AND APPARATUS FOR MULTIMEDIA CONTENT DELIVERY TO CABLE TV AND SATELLITE OPERATORS
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application claims priority to U.S. Provisional Application 62/095,504, filed December 22, 2014, the complete disclosure of which is incorporated herein by reference in its entirety.
FIELD OF THE DISCLOSURE
[0002] The present disclosure provides a mechanism for delivering multimedia content from a cloud-based system to cable or satellite content providers.
[0003] The disclosure describes a novel method and a system for continuous media delivery and secured synchronization of multimedia files that are broadcasted to set top boxes or subscriber units in the field that are connected to IPTV or Cable Systems. The present technology can securely connect to a cloud-based platform (e.g. OpenStack, Amazon EC2, Mediamplify Cloud, etc.), creates a set of multimedia files (e.g. MPEG4 movies) into a secured repository (e.g. Secured File System, Linux/Windows, etc.), modifies those files according to certain rules as defined by the method herein, and reproduces those signals via a multicast video stream (e.g. User Datagram Protocol (UDP) multicast, etc.) encapsulated in a Transport Stream (TS) frame that can be a multiple transport stream (several transports in one TS frame) or a single transport stream (one transport stream according to the MPEG-2 Part 1 Specification or similar multimedia streaming that is compatible with MPEG-TS at a predetermined rate or at a variable bit rate).
BACKGROUND
[0004] Cable operators, Multiple System Operators (MSO), and Digital Satellite Systems (DSS) deliver media content by utilizing satellite, Radio Frequency(RF) and Integrated Receiver and Decoder (IRD) technologies, which provide outputs that can include Asynchronous Serial Interface (ASI) signal formats. Additionally, MSOs and DSSs use dedicated hardware encoders and transcoding systems if the source video/audio is incompatible with the set top boxes of the cable or satellite operator. In general, the video/audio content delivered via the satellite to the cable or satellite operator arrives via wireless link or receiver, which is processed by an IRD that receives the stream from the satellite and can ultimately broadcast that signal to the cable operator head end in several formats. In addition, some systems may include multicast units connected with IRDs that tune to the right transponder system.
[0005] Satellite systems are used to deliver data (e.g. TV, radio, Video-on-Demand, multimedia streams) from a particular source (e.g. third-party content provider) to cable head end systems or other satellite systems. Satellite delivery is performed via a broadcast signal containing digital pictures and audio inside a transponder frequency. Transponders are radio frequency (RF) space allocated in a particular satellite.
[0006] For example, a third-party content provider can maintain multimedia files in CD/DVDs, prerecorded storage, or as pre-recorded files that are reproduced in a particular audio or video encoding format using a video or audio encoder (e.g. MP3, H.264, AC3, and others). The audio and video is then made part of an RF signal that is broadcasted to the satellite system. This signal is then received by the cable or satellite operator using an antenna receptor and is decoded to generate either an analog signal (e.g. RCA Cables or any other analog interface) or digital signal (e.g. ASI, Asynchronous Serial Interface). If an analog signal is received, modern DVB (Digital Video Broadcasting) Cable and Satellite systems will encode this signal to their particular video/audio encoding requirements such that it can be broadcasted in digital packets via a multicast MPEG Transport Stream. The same occurs if a receiver decodes the signal in ASI. As such, this signal is converted from ASI to Ethernet, or simply redirected to a multicast MPEG Transport stream if arriving in a compatible encoding format for the receiving head end cable system.
[0007] When no satellite is being used, video, pictures and music, are read out from the DVD or a hard-drive containing MPEG-layer 3 or MPEG-layer 2 files that were encoded at a pre-determined bit rate according. Those video feeds can be reproduced at a synchronized timing using a multicast MPEG Transport Stream using RTP (Real-time Transport Protocol)-based multicast communication and RTCP (RTP Control Protocol)-based mechanisms, as well as standard UDP (User Datagram Protocol) multicast streams with proprietary timing.
[0008] Additionally, IPTV systems can provide mechanisms to broadcast a signal using the multicast protocol in MPEG transport streams to a customer premises with compatible set top boxes (same encoder, same streams, encryption keys, etc). A compatible set top box is such that when connected to the cable TV system it can properly decode video/audio from the cable operator' s head end system. Some of these IPTV systems rely on CDN (Content Delivery Networks) that enable distribution of multimedia files to IP-enabled devices. For example, a mechanism can be provided whereby a head end content (MSO) that is being broadcasted to a cable subscriber is also available to an IP-enabled device such as a smartphone, PC, or tablet computer.
SUMMARY
[0009] Disclosed are systems, devices, methods, and non-transitory computer-readable media for media delivery from a cloud or IP-based distribution network, or CDN (Content Delivery Network) to MSO (Multiple System Operators) or head-ends that include cable and satellite delivery mechanisms. The present technology unifies cloud-based delivery (Mediamplify cloud) with the cable-based mechanism (e.g. Comcast, Verizon FiOS).
[0010] An example computer-implemented method can receive, from a content provider such as a cable or satellite operator, a request for at least one media stream for playback on a broadcast media channel. The request for the at least one media stream can include a request for a plurality of multimedia items. Content corresponding to the plurality of multimedia items can be obtained from at least one website or cloud service that offers the content in at least one first format. Based on the obtained content, the plurality of multimedia files can be generated in a second format that is compatible with the content provider. The media stream can assembled using each of the multimedia items generated in the second format. The media stream can be provided to the content provider for broadcast on the broadcast media channel.
[0011] In some aspects of the present technology, obtaining the content can include retrieving at least one audio file corresponding to an audio component of the multimedia item and retrieving a plurality of images as screen captures that can correspond to a video component of the multimedia item. Generating the multimedia files in the second format can include combining the plurality of images and the at least one audio file to create each of the multimedia items.
BRIEF DESCRIPTION OF THE DRAWING
[0012] There are shown in the drawings, embodiments which are presently preferred, it being understood, however, that the disclosure is not limited to the precise arrangements and instrumentalities shown [0013] FIG. 1 shows an example block diagram illustrating an exemplary head end (MSO);
[0014] FIG. 2 shows an example block diagram of the apparatus located at the MSO facility (mediaplug);
[0015] FIG. 3 shows an example method for initialization of Mediaplug components;
[0016] FIG. 4 shows an example method to obtain list of files to multicast to a cable operator MSO;
[0017] FIG. 5 shows an example method for encoding video and audio files;
[0018] FIG. 6 shows an example method for providing multimedia content to a cable operator;
[0019] FIG. 7 shows an example block diagram of example software components of the Mediaplug;
[0020] FIG. 8 shows an example method to multicast from a component in the Mediaplug;
[0021] FIG. 9 shows an example method to insert metadata into in an MPEG Transport Stream;
[0022] FIG. 10 shows an example block diagram for a Virtualized Mediaplug
[0023] FIG. 11 shows an example method to initialize virtual mediaplug components; and
[0024] FIG. 12 shows an example possible system embodiment for implementing various embodiments of the present technology
[0025] While the present technology is amenable to various modifications and alternative forms, specifics thereof have been shown by way of example in the drawings and will be described in detail. It should be understood, however, that the intention is not to limit the present technology to the particular embodiments described. On the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the present technology.
DETAILED DESCRIPTION
[0026] The following description should be read with reference of the drawings, in which like elements are numbered in similar fashion. The drawings are not necessarily at scale and are not intended to limit the scope of the disclosure. Those skilled in the art understand that there are alternatives that may be used as well which are also considered to be within the scope of the disclosure.
[0027] Generally, the disclosed technology pertains to a delivery mechanism for providing multimedia content to cable systems without the use of a satellite system in a fault-tolerant and reliable fashion. FIG. 1 illustrates a block diagram that includes aMediamplify Cloud 101. The cloud 101 can refer to a set of virtualized instances or non-virtualized instances of one or more servers that provide access to archives and digital files with media content stored in them. For example, the cloud 101 can be implemented using VMWave or Xen and OpenStack, or simply reside in Amazon Web Services or Rackspace. In addition, special provisioning can be provided in cloud 101 for secured access.
[0028] The cloud 101 service files can be accessible via the internet 105 and can be stored in a cloud-based file system or cluster file system (e.g. GlusterFS or Hadoop, Amazon S3). These files can be multimedia files that include, as non-limiting examples, MPEG-4 Video files with H.264 encoded video and AAC encoded audio or MPEG-layer 3 audio as well (MPEG2 Video could also be used). These files can be accessed using a secured transport protocol such as, but not limited to, SSL or a synchronization tool as "rsync," commonly found in Linux/MacOS environments. The communication link 106 can be a DSL, Cable Modem, Wireless WAN, Fiber, Metro Ethernet, etc. and can be linked with the MediaPlug 110 unit. The Mediaplug 110 can reside at the "MSO" facility and can be collocated with all the cable or satellite operator equipment. The MSO is a facility used by the Cable Operator to receive all TV and Radio Programming that the cable operator will then multiplex, re-encode, and distribute from the MSO to a set of subscriber units with set top boxes, or modulate the signal in TV analog channels (2-60+) .
[0029] One function of the MediaPlug 110 is to provide multimedia content, e.g., video and audio, to an MPEG multiplexer at the headend 120. The content can be pre-formatted in MPEG Transport Streams that the multiplexer 120 (e.g. Arris TMX-2010) will then be able to capture, reformat in any way the operator requires, and deliver to the cable operator's subscribers via Coaxial Cable 125 or any other means suitable for a multicast or unicast delivery to its customers with "Set Top boxes" 130, 131, 132. A "Set Top Box" is customer premise equipment that will be able to decode and display in a TV set or other screen display, via a signal received by its HDMI or analog input, the contents of the TV programming the cable operator offers.
[0030] The Mediaplug 110 can be a computer or server or set of computers of servers that can communicate with the Cloud 101 using a standard Local Area Network (e.g. Gigabit Ethernet). The Mediaplug 110 can store, encode, and transmit multimedia files in the format that the cable operator at its MSO requires. This cable format may include audio, video, and metadata (subtitles, control data, and any additional component that can be packaged in an MPEG Packet that is part of the MPEG Transport Stream). In one embodiment, Mediaplug 110 can include components such as a caching unit, and one or more multicasting servers.
[0031] FIG. 2 illustrates an example configuration of the hardware components of Mediaplug 110 which includes three servers. A server is a computer device, usually rack mountable unit with volatile memory such as RAM, one or more micro-processors, and non-volatile memory such as a hard disk. However, one skilled in the art will recognize that the present technology is not limited to a particular configuration and can be varied according to different specifications. FIG. 2 illustrates Mediamplify Cloud 202 which is connected to the Internet backbone 203. Link 204 (e.g. DSL, Cable Modem, Fiber) provides connectivity between cloud 202 and a Network Address Translator unit 205 (e.g. Cisco RVS4000, or any other NAT device) which can provide Gigabit Ethernet switching capabilities. The lines 206, 207, and 208 show Gigabit Ethernet links, however one skilled in the art will understand that Gigabit Ethernet or a Wireless LAN using 802.1 lg/n would be also a suitable interface to be used in a Mediaplug configuration. The Caching unit 210 can be the main storage component in the system and it can contain all of the multimedia files in the format required by the cable operators, such as the Transport Stream format. The caching unit can communicate with Multicast-01 215 and Multicast-02 220. The two Multicast units 215, 220 can pull or obtain information from the caching unit via TCP/IP or UDP, using protocols such as HTTP (Hypertext Transport Protocol) or RTSP (Real-time Streaming Transport). The caching unit can also encode and transcode the multimedia files it downloads from the Cloud 202. The caching unit can also keep updated information of all the metadata and create customizable screens that can be converted into video formats compatible with the Transport Stream content that will be provided to the cable operator at its MSO. MPEG MUX MSO 230 can correspond to a program residing in a computer system that processes and handles the MPEG streams from multiple sources. The MUX 230 can also function as the bridge between the source of the signal and the customer's set top box units. A non-limiting example program would be the CherryPicker DM6400 by MOTOROLA. Each of the components or modules of the Mediaplug can include software for performing functions and/or methods according to the present technology.
[0032] FIG. 3 illustrates an example method that can be used to configure and provision the units at boot-time. Each unit can initialize keys for SSH/IP Sec access at 305. For example, the units can use IPSec and SSH Tunnel keys that can be stored in the administrator user account. The keys can be generated by the Mediamplify Cloud that are private keys or certificates, shared secrets or similar components by which the Mediaplug can connect securely with the Mediamplify cloud. Each cable operator can provide custom-based IP addressing at 310 for the Gigabit Ethernet ports that connect to the operators' network. Changing the IP addresses of the Multicast-01 and/or Multicast-02 refers to the IP address by which the links 221 and 216 in FIG. 2 are linked to the MPEG MUX MSO unit. A change of the IP address can refer to the 'ifconfig' command which is used in Linux-based systems and that can be changed in an Ubuntu box at /etc/network/interfaces file. Similarly, the network routing can be changed if the multicast address cannot be routed to the default gateway which points to the 205 NAT, but to 221 and 216 respectively for Multicast-02 and Multicast-01. In general, the 239.*.*.* network address may be routed to those interfaces.
[0033] At step 320, the Mediaplug can configure playlists per channel and multicast address. The playlist and channel configuration can be obtained from the cable/satellite operator and can include any number of channels for broadcasting multimedia content to end customers. As an example, the channels may include music channels corresponding to different music genres, e.g. rock, alternative, classical, jazz, hip hop, etc. Based on the particular genre, a corresponding playlist can be selected and/or configured. The Mediaplug can also determine the multicast address for each of the channels.
[0034] In some configurations a customizable user interface (UI) can be added to the output video MPEG Transport stream. As such, if the "Custom UI" variable is then true and valid, 325, a link or URL 330 is loaded into the Mediaplug configuration that can be used when the video interface is created. The custom video interface can be created and fetched as needed or it can be created in real-time for a particular channel or for all the channels. The custom user interface variable can be uniquely configured per channel in the Mediaplug. The mediaplug can also monitor 340 all of the processes to ensure stability and continuous, uninterrupted operation. The initialization of the Mediaplug can also include setting the RSYNC profile 350. The RSYNC profile can be used to update and synchronize media content with the cloud when the Mediaplug determines that the content has changed or been modified.
[0035] As described above, the Mediaplug can broadcast the multimedia MPEG stream using the programmed IP Address in multicast-01 and multicast-02 according to the method illustrated in FIG. 3. In FIG. 4, an example method is shown for obtaining the list of files to playback or broadcast to the MSO system. The file type or format requested by the multicasting embodiment, e.g. multicast-01 or multicast-02, can include Transport Stream (e.g. TS) and/or MPEG-4 Part 14 (e.g. MP4) values 405. This request is executed from any of the multicasting embodiments, usually making use of the HTTP GET Request. The multicasting unit can connect to an HTTP Server and retrieve the list of files based on the type identified by the request. As the caching unit may handle multiple channels of different types and name, each channel can have a unique identifier corresponding to the name of the playlist. For example channel name "Cool" or "Rock," 410 are names that can identify the channel to be retrieved from the caching server. The activity log in the request can also be updated keeping track what step of the process is being completed 415. At 420, the caching server can generate a "playlist file" of mime type "audio/x-mpegurl" which will be interpreted as a multimedia playlist file when requested. Finally, as the playlist usually may not vary for a period of time until a new request is received, the playlist can be kept in the local memory caching service 425.
[0036] As the requests arrive at the "Caching" unit from the multicasting units, the "Caching" unit can be configured to perform parallel encoding for all the different channels available from the unit. FIG. 5 illustrates an example method that can be implemented by the Mediaplug, and in particular, by the "Caching" unit. At step 505, encoding flags or input parameters can be received which may include GOP (Group of Pictures) which is used by certain video encoders, resolution (e.g. 640x480) in pixels, bit rate in Kilobits per second, any encoding type of video (e.g. H.264), encoding type of audio (e.g. AC-3), duration of the video feed, etc. In addition, a "format" parameter can be received which refers to the output format or encapsulation, MPEG TS or MPEG4. The caching unit encodes audio and video, and can use the disk space in the Mediaplug. Hence in 508 the method checks for sufficient space left in the memory or hard drive. This task can be run on a thread together with all the other tasks that are working in parallel in order to speed up the processes of encoding and transcoding.
[0037] Several indexes or identifiers by the name "ChID," (these integers preferably no more than an integer value from 1 to the maximum number of channels) begin (e.g. MAX could be 50) to be generated by the caching unit 510. This method can be performed for all the channels and all media linked to that particular channel identified by "MedialD" 515, in this particular case the "MedialD." The variable "MAX" in this context can refer to the maximum number of media files read (e.g. 500). The data captured from the media could be "Artist," "Author," or any other variable from an audio file. If a video file is used this could be "program name," "movie name," or any associated metadata information tied to the multimedia file and the Channel Identifier (ChID) associated with it "MedialD" file.
[0038] The caching unit can add a "Custom UI," or a customizable User Interface that is designed using Hypertext Markup Language (e.g. HTML) or a web service. A web service that matches a REST interface (Representation State Transfer) and uses Javascript Object Notation (JSON) or XML (Extensible Markup Language) to respond to a custom UI request can be used. Moreover, this custom UI can include JavaScript, CSS, and HTML content files that can be rendered using a browser using WebKit (e.g. PhantomJS or Safari) or any other browser-rendering engine compatible with HTML4/5 or any future HTML version. If a customizable UI is part of the configuration 518, rendered images from a web service can be used to create overlapping video or replace any video in the video feed file. Likewise, this video or customized screen can be used when no video at all is defined in the multimedia file, e.g. Audio-only file. The web-service to be contacted may use the "PlayListName" to be used for the Channel identified by the channel ID "ChID, " retrieved using the "PlaylistID" function 520. Each new PlayListName may contain different identifies to retrieve new images. Therefore one or more screens will be retrieved from the rendering engine as such 522 can be used to retrieve a set of screens from the web-service being contacted.
[0039] Capturing screens 526 can be done using a tool similar to PhantomJS (Webkit-based headless engine for rendering images) or QtCapture, or simply capturing screen frame buffers from a machine running any software tool. Alternatively, the method can capture a video file from another rendering function, which can create a sequence of screens that can be used to generate a video file. Once all screens and video are captured, 530 the resulting output created can be a video file that then is either replaced or merged (multiplexed) with the original file. As an example, if an audio content file is used such as an MP3 file, this function can create a video file from all the screens captured by 526. Those screens are generated by calling a URL where the web-service is associated and creating a video file with the input audio and the captured screens. The output format can be an MPEG Transport Stream file that can be retrievable or streamed to the multicasting embodiment.
[0040] Once the output file containing the screen shoot sequence of videos encoded in H.264 (for example, or MPEG2Video) is created, the audio can be multiplexed and added to one of the encoding channels of the resulting video output and saved into the caching unit storage disk 550. The multiplexing can be done using tools such as FFMPEG or VLC or any proprietary tool that can multiplex audio and video files into a resulting MPEG Transport Stream (or MPEG-4 output, or any similar output).
[0041] FIG. 6 illustrates an example method 600 for providing multimedia content to a cable operator. At 602, a device such as the Mediaplug device receives a request from a cable/satellite operator for a media stream. A media stream can be formatted according to HTTP Live Streaming (HLS), an HTTP/RTSP playlist, or an RTSP stream. The media stream can include any number of multimedia items such as audio/video multimedia items. In some instances, the request can include a channel identifier that is associated with a particular category of multimedia items, such as a music genre corresponding to rock, classical, jazz, etc.
[0042] At 604, contents corresponding to the multimedia items can be retrieved, downloaded, or otherwise obtained from one or more websites or cloud based servers. The contents can include data corresponding to either or both of the audio and video components of the multimedia items. For example, the data corresponding to the audio portion can include a music file in an MP3 format. Similarly, the data corresponding to the video portion can include a video that is obtained directly from the cloud. Alternatively, the video component can be assembled by using one or more screen captures from a website. The screen captures can include captures of still images presented by the website as well as captures of images from videos that are available on the website. As illustrated in FIG. 5, the screen capture process can be repeated any number of times.
[0043] The Mediaplug can then generate 606 multimedia items in a second format based on the content obtained from the web server or cloud. That is, the Mediaplug can convert/encode the audio file into a format that is suitable for the cable operator and it can also assemble the various screen captures to generate the video component of the media item. Once the audio and video files are created, they can be combined, multiplexed, or encoded to a format that can be utilized by the multimedia stream. For example, the audio can be multiplexed and added to one of the encoding channels of the video output. The various multimedia items can be used 608 to assemble the media stream in a format corresponding with the request (e.g. HLS, HTTP/RTSP, RTSP stream). The media stream can be provided 610 to the content provider for broadcast on the broadcast media channel.
[0044] FIG. 7 illustrates a block diagram of an example architecture of the Caching Unit components. The figure shows 720 as the main operating system, which can be Linux, Windows, or any other operating system in use. The caching unit can contain two types of operating system services HTTP 710 and RTSP 712. The RTSP service provides access so the media saved or created in the repository of the caching unit file system 730. The RTSP server can be implemented using the Live555 RTSP server or similar. Alternatively, the HTTP Server can be implemented using "nginx" or "apache" web browsers, and the RTSP Servers can be implemented using Live555 services. In the case of the HTTP Server, this can also be used to get the Playlist of files generated from the method on FIG-5 715.
[0045] The GetPlayList 715 command can be implemented as follows:
• Receive and process a request for the name of the channel or playlist to be used,
• Retrieve all the filenames ordered by date of creation of all the files in a particular directory,
• Cache all the files into cache memory (e.g. Memcache or similar tool),
• Add type of HTTP header as M3U or PLS (e.g. a text file with a list of files or URLs to play), • Return a list of HTTP or RTSP URLs to access each video generated or stored in the caching unit.
[0046] In some embodiments, the HTTP Server can also be used to clean up, update, and delete operations. The block unit called "CleanUp" 722 is in charge of keeping up to date caching information data and/or deleting files already used or that are not required to be used anymore by the multicast unit(s). Also the HTTP Server functions can include servicing M3U8 files corresponding to the HTTP Live Streaming format and all the segments created by the HTTP Live Streaming function. As such, a HTTP Live Streaming (HLS) generation program will produce the files and segment them as needed by the encoder and can then be serviced by the HTTP Server 710. The Mediamplify Parallel Encoder 705 corresponds to a block that can implement the method shown in FIG 5. The resulting outputs from the Media Parallel Encoder 605 are stored in the file system 730. The Mediamplify Parallel encoder 605 can be implemented in Python and can rely on web services (such as Twisted or Django), PhantomJS, and FFMPEG. Monit 750 correspnds to a monitoring block that can ensure that the process running the "Media Parallel encoder" keeps running at all times and does not stop upon failure. Additional blocks that can be included in Mediaplug are RSYNC 742 and SSH 744. Thes blocks can be used to ensure synchronization and updates of new media files coming from the cloud. The RSYNC block uses the "rsync" protocol and program over ssh (Secured Socket Layer Shell) to download media from the "Mediamplify Cloud" and synchronize multimedia files that are newly updated and could be part of particular Channel ID and a Playlist Name. Each new file synchronized with the Mediamplify cloud has a Media ID value which the cloud can assign, and include additional metadata information such as "Artist" and "Author." The "Mediamplify Parallel Encoder" also may receive requests from other components such as getStatus or StartEncoding, indicating what file is being encoded, its progress, and what other files are in the queue for encoding.
[0047] In some embodiments, the information from getStatus or StartEncoding can be generated using a JSON message retrieved upon a request via HTTP GET to the Mediamplify Parallel Encoder. A graphical user interface can then be used to plot the status of each song being encoded or provide previews of the videos being generated. Below is one non-limiting example of a JSON response made by the "Mediamplify Parallel Encoder" service: {"song_encoded" : "ts", "artist": "Madonna", "duration": 0, "exists_on_disk": True, "genre": "Pop", "title": "Like a Virgin", "music_id": " 1082863" }
[0048] Response regarding the tatus for a song being encoded or a list of songs being encoded, as follows:
{"music_id": "109281", "encoding": "78%", "output_file": "/var/w ww/eglacast/pop/ 109281- abcdef-uuuid.ts", "thumbnail" : http://mediamplify.com/images/abcdef-uuid.png" }
[0049] In some embodiments, there are two possible TCP/IP sockets open tor listening used by the Caching component of the Mediaplug. A) Port 80 (as an example), or the standard HTTP port, used to retrieve content, view video content generated, update/delete cached files, and access the playlists, and B) Port 9553 (as an example), or the standard Mediamplify Parallel encoder port, used to monitor and review status of the songs being encoded, the data being collected, and all other related activities with encoding. Via this port, encoding/transcoding can be started or stopped.
[0050] Similarly, the "Mediaplug" components labeled "Multicast-01" and "Multicast-02" can also provide an HTTP interface to access the current song being played or being multicast via the HTTP Server 770. For example, the HTTP Server may run a PHP or Python script that will return the current song being played by the streamer. The streamer block 768 can connect to the caching server via HTTP or RTSP protocols. The implementation of this block can be done using a combination of FIG-8 instructions and using multicasting services available in VLC or FFMPEG (e.g. multicast, or tstools / tsplay).
[0051] One function of the streamer 768 is to read the files encoded by the caching unit, process those for multicasting and creating the MPEG Transport Stream multicast packets sent via the Ethernet port. Additionally, the multicast-01 and multicast-02 blocks contain a "Recording" function 790, which is in charge or recording the multicasting streams and recording for later use in case of failure on the system. This operates by listening into the UDP (User Datagram Protocol) multicast address of each individual channels being broadcasted, and recording the output into a file. This file is then stored and made accessible for future use. Monit 764 can maintain the streamer process running at all times throughout a regular broadcasting cycle. [0052] FIG. 8 illustrates an example method that can be performed by the streamer module in the Mediaplug. In one embodiment, the streamer can perform functions that include: connect to the caching unit; retrieve generated videos per playlist; determine if unit is working in fault-tolerant mode; playback pre-recorded audio if a failure is detected; and record units own output or any other streamer output if connected to switch or hub.
[0053] Turning to the method from FIG. 8, at step 805, the streamer reads the multicast address associated to the channel being broadcasted, in order to generate IGMP (Internet Group Management Protocol) messages and establish a multicast group (e.g. 239.0.0.1:8001). The channel name is also initialized at 810. This value can be used to retrieve the proper content from the Caching server. In one embodiment, a configuration table can be created with tuples: <ChID>, <Channel Name>, <Multicast Address>, <Multicast Port>, such that no duplicates are used or created. Automation software can create these configuration tables and automate the process of creating all streamer calls. A streamer method can be called (e.g. from /etc/init.d/) at boot time and monitored with a tool such as Monit. Once the configuration is known, the streamer will check 850 if there are already multicasting transport streams generated by any other multicast component and avoid collisions. If such multicast data is detected, a delay can be used to delay a new attempt to check on that particular multicast group 860. Hence this block indicates "Wait <Time>" or stop all processing for the amount of <Time> recommended by the specification. For example, the time for delay can be the "Main source" of input less the amount of time for receiving input from a backup server. During initialization, if the same multicast address is used to broadcast main and backup, the one with less amount of time delay will be first, followed by the "Backup." In cases were a hot spare is used, or two different multicast addresses can be in use, one for main and the other one for backup, then this step may be optional.
[0054] Once it is determined that there are no elements multicasting at the same address, the streamer will check if the Caching server is "alive" 815. This can be performed in multiple ways, for example an HTTP Request must not generate a 404 or 403 error, but instead a successful 202 value. Likewise, ping requests can also be used before making an HTTP GET call to identify if the caching server is present and available. In the case that there are no ping responses, a timeout will trigger a failure and a pre-recorded content 820 will be played or sent to the multicast address and port of record. Recordings can be scheduled daily for an extended period of time (e.g. 8 hours or more). These recordings can store the recorded file in a temporary folder and then move that file over to the playback folder where streamer can read it.
[0055] If it is determined that the caching server is functioning properly, then there are three types of streams that can be sent to the multicast address and port:
[0056] HTTP Live Streaming (HLS) a multi-segment stream usually defined by extension M3U8, containing a set number of segments for a real-time stream source (Video/audio)
[0057] HTTP/RTSP file playlist, this means that one or many files are read from the caching server, usually defined by extension M3U or PLS
[0058] RTSP stream, this could be constant stream similar to HTTP Live Streaming (HLS) but using an RTSP protocol instead of HLS stream.
[0059] When multiple files are part of a playlist, the bit-rate may vary between playlist files. Different bit-rates among files in a playlist could cause problems in non-CBR (Constant Bit Rate) systems. To avoid such problems, the Mediaplug can maintain the stream at the bit rate of the previous stream by adding null MPEG TS packets at the same rate as the previous media file was being sent over the multicast port. This will maintain appropriate bit-rate for play lists.
[0060] In RTSP service, streaming occurs by writing the media content to a named pipe or a FIFO file and allowing the RTSP service to read from that file and generate an RTSP stream. As such, in HLS and RTSP Streaming, the caching server knows the filename of the file being streamed. When the file is streamed from via playlist, both the caching and multicast unit know the name of the file being streamed, as the multicast server can then request from the caching server the metadata associated with the file name being streamed.
[0061] Accordingly the present technology can accommodate both cases and retrieve the current file name being streamed as a JSON response. The response can identify the content displayed or played back at the end-user's set top box.
[0062] If HLS is used for streaming 818, the streamer will connect to the proper M3U8 file 820 to pull all the segments corresponding to the live playlist file on the M3U8 stream. Programs and libraries such as FFMPEG or VLC can then be used to encapsulate the stream feed to an MPEG TS and multicast it to the pre-programmed multicast address and port. Similarly, if an RTSP stream 825 is used, the same process takes place at 830 using the RTSP as a protocol instead of HTTP.
[0063] If a list of files in a playlist 832 is in use "M3U," the list can be obtained using an HTTP GET request made to the caching server. Each file can then be streamed to the multicast address and port of record 838 until all elements of the playlist are completed and played back 835.
[0064] In some instances, metadata can be added to the stream as an additional PID (Program Identifier) in the MPEG Transport Stream. If metadata is needed 841 then it will be inserted 842 here in accordance with FIG. 9.
[0065] For example, the metadata added could be as simple as a string containing Artist, Name, or other information, or as complex as DCII Text (DigiCipher version 2, as specified by Motorola), MPEG TS Control messages (Two-way), or any other metadata or PID that can be inserted into the stream being sent to the multicast address and port.
[0066] FIG. 9 illustrates an example method used to insert a special data packet into the stream being written to the multicast address, without making use of timing information. Therefore, the metadata can be inserted such that it does not have timing dependency. Alternatively, certain type of metadata such as subtitles or lyrics may need to be inserted only at certain times so that it is synchronized with the video data.
[0067] The sample MPEG TS packet shown in 902 and 905 corresponds to the metadata inside the packet. This matches standard MPEG TS packet structure and may vary depending on the type of metadata being added to the stream.
[0068] The method to insert the packet starts by reading the MPEG TS file retrieved from the caching server 910. If metadata corresponds to only song information, this information can be added at uniform times. Non-uniform timing inclusions can also be done by adding a table that includes the time and the metadata content added to the stream 915. This metadata can be retrieved from the Caching unit using an HTTP Get request directed to the caching server 918 and may include artist name, title of song, or any other information 920. [0069] Creation of the MPEG Packet(s) 925 to be inserted at the appropriate times, which will set the places with those packets, will be inserted in the MPEG TS. Modify the PMT (Program Management Table) of the generated file and include the PID being inserted into the MPEG TS 930. At step 932, an audio PID may include a metadata PID of type DCII Text, which is added to the stream.
[0070] For each timing or location where the PID needs to be inserted 935, the method can insert those packets and generate a buffer that can be sent to the multicast address as in illustrated in FIG. 8.
[0071] As described above, the disclosed technology can be implemented using a set of modules included in the Mediaplug apparatus, which can be implemented using one or more servers accordingly. Alternatively, virtualization environments such as VMWare or Xen (FIG-9) can include a system and architecture that can be used to host all the Mediaplug functionality in a virtualized environment using a single server with multiple GbE ports.
[0072] The illustration in FIG. 10 shows a configuration of three virtual machines used for implementing the Mediaplug. Caching 1030 and its respective Ethernet port 1032, Multicast-01 1020 with two virtual Ethernet ports 1022 and 1024, and Multicast-02 1010, with two Ethernet ports 1012 and 1014. There are three bridges that include 1007, 1008, and 1009. The multicast Ethernet ports 1024 and 1014 are bridged with ethl 1004 and eth2 1003 respectively via bridge 1009 and bridge 1008, respectively. That is, the same content generated at 1024 can be forwarded via the bridge 1009 to the physical interface ethl 1004 in the hardware. The vSwitch 1045 or the virtual switch between the two bridged interfaces 1008 and 1009 permits the two virtual machines of Multicast-01 1010 and Multicast-02 1020 to record each other and monitor their activity.
[0073] To achieve concurrency and proper execution in a cloud environment, the Mediaplug can use the example method shown in FIG. 11. Once a create unit command is issued 1105, provisioning of the metadata information and proper IP Address is provided for that particular instance 1110. Once an IP Address and metadata has been chosen for the unit, the server can be allocated in the cloud and the instance can be created using CloudStack or OpenStack tools and services 1115. The instance is then allocated and runs in the proper node in the cloud 1120. Configuration of the assigned IP, Netmask, and VPN credentials are loaded at 1125. Finally, certificate information is uploaded 1130 as well as all services run to enable the unit for execution of its caching functions, multicasting functions, or both.
[0074] FIG. 12 illustrates a conventional system bus computing system architecture 1200 such as can be used in Mediaplug 110. Those of ordinary skill in the art will appreciate that other system architectures are also possible.
[0075] In computing system architecture 1200 the components of the system are in electrical communication with each other using a bus 1205. Example system 1200 includes a processing unit (CPU or processor) 1210 and a system bus 1205 that couples various system components including the system memory 1215, such as read only memory (ROM) 1220 and random access memory (RAM) 1225, to the processor 1210. The system 1200 can include a cache of high-speed memory connected directly with, in close proximity to, or integrated as part of the processor 1210. The system 1200 can copy data from the memory 1215 and/or the storage device 1230 to the cache 1212 for quick access by the processor 1210. In this way, the cache can provide a performance boost that avoids processor 1210 delays while waiting for data. These and other modules can control or be configured to control the processor 1210 to perform various actions. Other system memory 1215 may be available for use as well. The memory 1215 can include multiple different types of memory with different performance characteristics. The processor 1210 can include any general purpose processor and a hardware module or software module, such as module 1 1232, module 2 1234, and module 3 1236 stored in storage device 1230, configured to control the processor 1210 as well as a special-purpose processor where software instructions are incorporated into the actual processor design. The processor 1210 may essentially be a completely self-contained computing system, containing multiple cores or processors, a bus, memory controller, cache, etc. A multi-core processor may be symmetric or asymmetric.
[0076] To enable user interaction with the computing device 1200, an input device 1245 can represent any number of input mechanisms, such as a microphone for speech, a touch- sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 1235 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with the computing device 1200. The communications interface 1240 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
[0077] Storage device 1230 is a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 1225, read only memory (ROM) 1220, and hybrids thereof.
[0078] The storage device 1230 can include software modules 1232, 1234, 1236 for controlling the processor 1210. Other hardware or software modules are contemplated. The storage device 1230 can be connected to the system bus 1205. In one aspect, a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 1210, bus 1205, display 1235, and so forth, to carry out the function.
[0079] It can be appreciated that example system 1200 can have more than one processor 1210 or be part of a group or cluster of computing devices networked together to provide greater processing capability.
[0080] For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.
[0081] Any of the steps, operations, functions, or processes described herein may be performed or implemented by a combination of hardware and software modules, alone or in combination with other devices. In an embodiment, a software module can be software that resides in memory of a client device and/or one or more servers of a content management system and perform one or more functions when a processor executes the software associated with the module. The memory can be a non-transitory computer-readable medium.
[0082] In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
[0083] Methods according to the above-described examples can be implemented using computer- executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
[0084] Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
[0085] The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.
[0086] Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims.

Claims

CLAIMS What is claimed is:
1. A computer- implemented method comprising:
receiving, from a content provider, a request for at least one media stream for playback on a broadcast media channel, wherein the at least one media stream includes a plurality of multimedia items;
obtaining content corresponding to the plurality of multimedia items from at least one website offering the content in at least one first format;
generating each of the plurality of multimedia items in a second format based on the content obtained from the at least one website, wherein the second format is compatible with the at least one content provider;
assembling the at least one media stream using each of the plurality of multimedia items generated; and
providing the at least one media stream to the content provider for broadcast on the broadcast media channel.
2. The method of claim 1, wherein the obtaining comprises:
retrieving, for each of the plurality of multimedia items, at least one audio file corresponding to an audio component of the multimedia item and a plurality of screen captures corresponding to a video component of the multimedia item.
3. The method of claim 2, wherein the generating comprises:
combining the plurality of screen captures and the at least one audio file to create each of the plurality of multimedia items.
4. The method of claim 2, wherein the retrieving of the plurality of screen captures comprises obtaining the plurality of screen captures from a playback of a video on the at least one website.
5. The method of claim 4, wherein the combining of the plurality of screen captures comprises metadata corresponding to at least one of a song title, an artist, and a music genre.
6. The method of claim 1, further comprising:
identifying the at least one website based on a channel identifier associated with the broadcast media channel.
7. The method of claim 1, further comprising:
querying the content provider to determine the second format, wherein the second format corresponds to a Moving Picture Experts Group (MPEG) format.
8. The method of claim 3, further comprising:
detecting a change at the at least one website corresponding to at least one of the plurality of multimedia items;
in response to detecting the change, retrieving a new plurality of screen captures from the at least one website for the at least one of the plurality of multimedia items; and
reassembling the at least one of the plurality of multimedia items using the new plurality of screen captures to yield an updated multimedia item.
9. The method of claim 1, wherein the content provider is a cable television operator or a satellite television operator.
10. The method of claim 1, wherein providing the at least one media stream to the content provider comprises streaming the at least one media stream to an MPEG multiplexer associated with the content provider.
11. The method of claim 1, further comprising:
provisioning, on at least one server, at least one virtual machine for performing the receiving, obtaining, generating, assembling, and providing.
12. A system comprising:
at least one processor;
a computer-readable storage medium having stored therein instructions which, when executed by the processor, cause the at least one processor to perform operations comprising: receive, from a content provider, a request for a multimedia content for playback on a broadcast media channel;
retrieve, from a cloud based service offering the multimedia content in a first format that is not compatible with a required format associated with the content provider, a plurality of screen captures corresponding to a video component of the multimedia content and at least one audio file corresponding to an audio component of the multimedia content;
merge the plurality of screen captures with the at least one audio file to create the multimedia content in the required format; and
send the multimedia content to the content provider.
13. The system of claim 12, further comprising:
retrieve a video from the cloud based service, wherein the plurality of screen captures correspond to a plurality of static images obtained from the video.
14. The system of claim 12, further comprising:
select the multimedia content from the cloud based service based on a music genre associated with the broadcast media channel.
15. The system of claim 12, wherein the request for the multimedia content comprises a request for a plurality of multimedia streams for playback on a plurality of broadcast media channels.
16. The system of claim 12, wherein the multimedia content is sent to a broadcast internet protocol (IP) address associated with the content provider.
17. A non-transitory computer-readable storage medium having stored therein instructions which, when executed by a processor, cause the processor to perform operations comprising:
receiving, from a content provider, a request for at least one media stream for playback on a broadcast media channel, wherein the at least one media stream includes a plurality of multimedia items; obtaining content corresponding to the plurality of multimedia items from at least one website offering the content in at least one first format;
generating each of the plurality of multimedia items in a second format based on the content obtained from the at least one website, wherein the second format is compatible with the at least one content provider;
assembling the at least one media stream using each of the plurality of multimedia items generated; and
providing the at least one media stream to the content provider for broadcast on the broadcast media channel.
18. The non-transitory computer-readable storage medium of claim 17,
wherein the obtaining comprises retrieving, for each of the plurality of multimedia items, at least one audio file corresponding to an audio component of the multimedia item and a plurality of screen captures corresponding to a video component of the multimedia item; and
wherein the generating comprises combining the plurality of screen captures and the at least one audio file to create each of the plurality of multimedia items.
19. The non-transitory computer-readable storage medium of claim 17, wherein the at least one media stream corresponds to at least one of an HTTP live stream (HLS), an HTTP play list, and a Real-time Streaming Transport (RTSP) stream.
20. The non-transitory computer-readable storage medium of claim 17, storing additional instructions which, when executed by the processor, result in operations comprising:
detect a network failure affecting connectivity with the at least one website; and
in response, provide a previously recorded media stream to the content provider.
21. The non-transitory computer-readable storage medium of claim 17, wherein assembling the at least one media stream comprises inserting at least one MPEG packet including metadata corresponding to the multimedia item.
22. A computer-implemented method comprising:
provisioning, on at least one server, a first virtual machine configured to obtain and cache multimedia content from at least one cloud-based service;
provisioning, on the at least one server, a second virtual machine communicatively coupled to the first virtual machine and to a cable operator, wherein the second virtual machine is configured to multicast the multimedia content from the first virtual machine to the cable operator; and
provisioning, on the at least one server, a third virtual machine communicatively coupled to the first virtual machine, the second virtual machine, and the cable operator, wherein the third virtual machine is configured to monitor an output from the second virtual machine to determine if a failure condition has occurred.
23. The method of claim 22, wherein provisioning the first virtual machine, the second virtual machine, and the third virtual machine comprises:
configuring at least one digital certificate to provide access to the multimedia content; and configuring at least one internet protocol (IP) address corresponding to a multicast address for the cable operator.
PCT/US2015/067464 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators WO2016106360A1 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
PL15874332.8T PL3238457T3 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators
US15/538,911 US10123074B2 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators
DK15874332.8T DK3238457T3 (en) 2014-12-22 2015-12-22 Method system and apparatus for delivering multimedia content to cable television and satellite operators
EP15874332.8A EP3238457B1 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators
FIEP15874332.8T FI3238457T3 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators
US16/152,606 US10524002B2 (en) 2014-12-22 2018-10-05 Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators
US16/729,986 US11140441B2 (en) 2014-12-22 2019-12-30 Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators
US17/493,490 US20220030306A1 (en) 2014-12-22 2021-10-04 Method, system and apparatus for multimedia content delivery to cable tv and satellite operators

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201462095504P 2014-12-22 2014-12-22
US62/095,504 2014-12-22

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US15/538,911 A-371-Of-International US10123074B2 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators
US16/152,606 Continuation US10524002B2 (en) 2014-12-22 2018-10-05 Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators

Publications (1)

Publication Number Publication Date
WO2016106360A1 true WO2016106360A1 (en) 2016-06-30

Family

ID=56151532

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2015/067464 WO2016106360A1 (en) 2014-12-22 2015-12-22 Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators

Country Status (8)

Country Link
US (4) US10123074B2 (en)
EP (1) EP3238457B1 (en)
DK (1) DK3238457T3 (en)
FI (1) FI3238457T3 (en)
HU (1) HUE063857T2 (en)
PL (1) PL3238457T3 (en)
PT (1) PT3238457T (en)
WO (1) WO2016106360A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107819863A (en) * 2017-11-16 2018-03-20 郑州云海信息技术有限公司 A kind of Explore of Unified Management Ideas and device of cloud platform user

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3238457B1 (en) * 2014-12-22 2023-07-26 Edwin A. Hernandez-Mondragon Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators
KR102381335B1 (en) * 2016-10-18 2022-03-31 이엑스피웨이 How to deliver content to mobile user devices
US10271077B2 (en) 2017-07-03 2019-04-23 At&T Intellectual Property I, L.P. Synchronizing and dynamic chaining of a transport layer network service for live content broadcasting
US11108840B2 (en) * 2017-07-03 2021-08-31 At&T Intellectual Property I, L.P. Transport layer network service for live content broadcasting
FR3073112B1 (en) * 2017-11-02 2021-04-02 Ateme METHOD AND SYSTEM FOR PROCESSING MULTIMEDIA CONTENT IN A METROPOLITAN AREA NETWORK
US11431773B2 (en) * 2019-06-13 2022-08-30 Caffeine Inc. Multicast broadcast network architcture
GB2597334A (en) * 2020-07-17 2022-01-26 Playrcart Ltd A media player
US11914555B2 (en) * 2021-09-10 2024-02-27 Arris Enterprises Llc Data processing for networking systems

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020138852A1 (en) * 2000-11-27 2002-09-26 Steven Reynolds System and method for providing an omnimedia package
US20030115612A1 (en) * 1998-04-01 2003-06-19 Morecom, Inc. Digital TV system with synchronized World Wide Web content
US20080040354A1 (en) * 2006-08-10 2008-02-14 Qualcomm Incorporated System and method for media content delivery
US20090178003A1 (en) * 2001-06-20 2009-07-09 Recent Memory Incorporated Method for internet distribution of music and other streaming content
US20140068690A1 (en) * 2012-08-31 2014-03-06 General Instrument Corporation Broadcast content to http client conversion
US20140143437A1 (en) * 2012-11-16 2014-05-22 Adobe Systems Incorporated Converting Live Streaming Content to Video-on-Demand Streaming Content

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7320025B1 (en) 2002-03-18 2008-01-15 Music Choice Systems and methods for providing a broadcast entertainment service and an on-demand entertainment service
US7275256B1 (en) 2001-08-28 2007-09-25 Music Choice System and method for providing an interactive, visual complement to an audio program
US8077263B2 (en) * 2006-10-23 2011-12-13 Sony Corporation Decoding multiple remote control code sets
US8763038B2 (en) * 2009-01-26 2014-06-24 Sony Corporation Capture of stylized TV table data via OCR
US9106800B2 (en) * 2007-08-31 2015-08-11 At&T Intellectual Property I, L.P. System and method of monitoring video data packet delivery
JP5669277B2 (en) * 2009-09-15 2015-02-12 トムソン ライセンシングThomson Licensing Method and apparatus for providing supplemental information
EP2609749A4 (en) * 2010-08-27 2015-04-15 Intel Corp Peer to peer streaming of dvr buffered program data
US9418388B1 (en) * 2010-09-08 2016-08-16 Imdb.Com, Inc. Techniques for displaying third party content
WO2012051528A2 (en) * 2010-10-14 2012-04-19 Activevideo Networks, Inc. Streaming digital video between video devices using a cable television system
KR20130048094A (en) * 2011-11-01 2013-05-09 한국전자통신연구원 Node apparatus for contents streaming relay and method using the same
US10231019B2 (en) * 2012-03-15 2019-03-12 Black Wave Adventures, Llc Digital parental controls interface
BR122015005194A2 (en) * 2012-06-28 2019-08-20 Ericsson Ab METHOD AND SYSTEM FOR INSERTING ADVERTISING IN LIVE MEDIA OVER THE TOP
US8953618B2 (en) * 2012-10-10 2015-02-10 Telefonaktiebolaget L M Ericsson (Publ) IP multicast service leave process for MPLS-based virtual private cloud networking
US9497489B2 (en) * 2013-03-12 2016-11-15 Google Technology Holdings LLC System and method for stream fault tolerance through usage based duplication and shadow sessions
US9143840B2 (en) * 2013-05-20 2015-09-22 Veenome, Inc. Systems and methods for evaluating online videos
US20140351871A1 (en) 2013-05-22 2014-11-27 Microsoft Corporation Live media processing and streaming service
US20150382081A1 (en) * 2014-06-27 2015-12-31 BlurbIQ, Inc. Generation of video content
EP3238457B1 (en) * 2014-12-22 2023-07-26 Edwin A. Hernandez-Mondragon Method, system, and apparatus for multimedia content delivery to cable tv and satellite operators

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030115612A1 (en) * 1998-04-01 2003-06-19 Morecom, Inc. Digital TV system with synchronized World Wide Web content
US20020138852A1 (en) * 2000-11-27 2002-09-26 Steven Reynolds System and method for providing an omnimedia package
US20090178003A1 (en) * 2001-06-20 2009-07-09 Recent Memory Incorporated Method for internet distribution of music and other streaming content
US20080040354A1 (en) * 2006-08-10 2008-02-14 Qualcomm Incorporated System and method for media content delivery
US20140068690A1 (en) * 2012-08-31 2014-03-06 General Instrument Corporation Broadcast content to http client conversion
US20140143437A1 (en) * 2012-11-16 2014-05-22 Adobe Systems Incorporated Converting Live Streaming Content to Video-on-Demand Streaming Content

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107819863A (en) * 2017-11-16 2018-03-20 郑州云海信息技术有限公司 A kind of Explore of Unified Management Ideas and device of cloud platform user

Also Published As

Publication number Publication date
HUE063857T2 (en) 2024-02-28
US20220030306A1 (en) 2022-01-27
US20190037271A1 (en) 2019-01-31
US10123074B2 (en) 2018-11-06
US20200322679A1 (en) 2020-10-08
EP3238457B1 (en) 2023-07-26
PT3238457T (en) 2023-11-16
US10524002B2 (en) 2019-12-31
PL3238457T3 (en) 2024-02-05
US11140441B2 (en) 2021-10-05
EP3238457A4 (en) 2018-09-19
DK3238457T3 (en) 2023-10-30
US20170374417A1 (en) 2017-12-28
FI3238457T3 (en) 2023-11-13
EP3238457A1 (en) 2017-11-01

Similar Documents

Publication Publication Date Title
US11140441B2 (en) Method, system, and apparatus for multimedia content delivery to cable TV and satellite operators
US11483598B2 (en) Method and system for real-time transcoding of MPEG-DASH on-demand media segments while in transit from content host to dash client
US10715843B2 (en) Forming one or more tile streams on the basis of one or more video streams
US11005903B2 (en) Processing of streamed multimedia data
US9954717B2 (en) Dynamic adaptive streaming over hypertext transfer protocol as hybrid multirate media description, delivery, and storage format
US9113186B2 (en) Providing syndication feed content on a television set-top box with limited decoder capability
CN108702527A (en) System and method for using the media of general interlayer distribution formats to transmit
WO2016058411A1 (en) Splicing method and splicing system for http live streaming media fragmentation
EP2521369A2 (en) Media file storage format and adaptive delivery system
JP2018530210A (en) Forming tiled video based on media streams
KR102303582B1 (en) Processing media data using file tracks for web content
US20170180821A1 (en) Video File Universal Identifier for Metadata Resolution
EP3001690A1 (en) Content supply device, content supply method, program, and content supply system
CN105340280B (en) Content supply device, Content supply method, storage medium, terminal installation and contents providing system
RU2656093C2 (en) Content supply device, content supply method, program, terminal device and content supply system
JP2019071680A (en) Terminal device and receiving device
US9843615B2 (en) Signaling and handling of forensic marking for adaptive streaming
US20220182690A1 (en) Method for media stream processing and apparatus for implementing the same
US11563792B2 (en) Managing repositioning requests in HLS and dash
EP4009650A1 (en) Method for media stream processing and apparatus for implementing the same
Thompson Maximizing content with cloud, virtualization and AI
CN115174998A (en) Future network-based broadcast system of real 4K home theater broadband metropolitan area network
WEIWEI An experimental study of video uploading from mobile devices with HTTP streaming

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15874332

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 15538911

Country of ref document: US

REEP Request for entry into the european phase

Ref document number: 2015874332

Country of ref document: EP