US20100269147A1 - Video stream index generation at a video content transmitter - Google Patents

Video stream index generation at a video content transmitter Download PDF

Info

Publication number
US20100269147A1
US20100269147A1 US12/424,399 US42439909A US2010269147A1 US 20100269147 A1 US20100269147 A1 US 20100269147A1 US 42439909 A US42439909 A US 42439909A US 2010269147 A1 US2010269147 A1 US 2010269147A1
Authority
US
United States
Prior art keywords
video
stream
index
frames
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/424,399
Inventor
Charlie William Zetterower
David A. Kummer
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
DISH Technologies LLC
Original Assignee
EchoStar Technologies LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by EchoStar Technologies LLC filed Critical EchoStar Technologies LLC
Priority to US12/424,399 priority Critical patent/US20100269147A1/en
Assigned to ECHOSTAR TECHNOLOGIES LLC reassignment ECHOSTAR TECHNOLOGIES LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KUMMER, DAVID A., ZETTEROWER, CHARLIE WILLIAM
Priority to TW99110000A priority patent/TWI420910B/en
Priority to PCT/US2010/030902 priority patent/WO2010120787A1/en
Publication of US20100269147A1 publication Critical patent/US20100269147A1/en
Priority to US13/487,101 priority patent/US8855468B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/4147PVR [Personal Video Recorder]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/78Television signal recording using magnetic recording
    • H04N5/782Television signal recording using magnetic recording on tape

Definitions

  • a popular and effective way of accessing television programming is via a satellite or cable television broadcast receiver or “set-top box” designed to receive television video and audio data from an orbiting satellite or a cable television headend.
  • a subscriber to a satellite or cable television service provider may receive any of several hundred programming channels providing news, sports, entertainment, and the like for live program viewing.
  • the set-top box may include a digital video recorder (DVR) for recording programs for subsequent playback at a time more convenient for the subscriber.
  • DVR digital video recorder
  • a DVR often incorporates a magnetic disk drive or similar data storage technology for storing the recorded programs.
  • Incorporating a DVR in a set-top box provides the further advantage of allowing the subscriber to implement “trick modes”, which are video playback modes other than the familiar normal-speed, forward playback mode.
  • Trick modes may include forward and reverse scanning of video at faster-than-normal speeds, forward and reverse slow-motion playback, forward and reverse frame-by-frame review, and pause, each of which is controlled by the subscriber, typically by way of a remote control device.
  • DVR normally records programs specifically indicated by the subscriber for later viewing, as well as programs currently being delivered or broadcast to the set-top box, the subscriber may use the trick modes on essentially any program being viewed.
  • the DVR is often designed to display to the subscriber disjoint frames of the video program, each for a short period of time in rapid succession, so that the subscriber or user may discern the progress of the scan.
  • the DVR or associated set-top box may generate indexing information indicating various access points into the video program being viewed at which complete video frames may be retrieved.
  • Such information is often necessary, as video programs are typically encoded in a compressed format in which many of the video frames are encoded presuming known information about the previous or subsequent frame.
  • Such “predicted” frames are thus typically poor candidates for presentation during a fast-scanning trick mode due to the lack of complete information being provided for that frame. At least some of the other trick modes may benefit from the use of this indexing information as well.
  • FIG. 1 is a simplified block diagram of a video communication network according to an embodiment of the invention.
  • FIG. 2 is a flow diagram of a method according to an embodiment of the invention for generating and using indexing information for video content in the video communication system of FIG. 1 .
  • FIG. 3 is a block diagram of an audio/video communication network according to an embodiment of the invention.
  • FIG. 4 is a block diagram of an audio/video content transmitter of FIG. 3 according to an embodiment of the invention.
  • FIG. 5 is a block diagram of an audio/video content receiver of FIG. 3 according to an embodiment of the invention.
  • FIG. 6 is a graphical representation of an example video stream employed in an embodiment of the invention.
  • FIG. 7 is a graphical representation of an example audio/video file employed in an embodiment of the invention.
  • FIG. 8A is a graphical representation of an index transport packet according to an embodiment of the invention.
  • FIG. 8B is a graphical representation of an index section within the index metadata transport packet of FIG. 8A according to an embodiment of the invention.
  • FIG. 8C is a graphical representation of an index within the index section of FIG. 8B according to an embodiment of the invention.
  • FIG. 9 is a flow diagram of a method according to an embodiment of the invention for processing an audio/video file and associated index file in an audio/video content receiver.
  • FIG. 1 is a simplified block diagram of a video communication network 100 according to an embodiment of the invention.
  • the video communication network 100 includes a video content transmitter 102 and a video content receiver 104 .
  • the video communication network 100 may be a satellite television broadcast network, with the video content transmitter 102 existing as part of a satellite uplink center incorporating a satellite headend, and the video content receiver 104 being embodied as a satellite television receiver or set-top box.
  • the video communication network 100 may be a cable television broadcast network, with the video content transmitter 102 incorporated in a cable headend, while the video content receiver 104 is a cable television receiver or set-top box.
  • Other video communication networks such as terrestrial (over-the-air) television broadcast networks, computer networks (such as the Internet), or any other wired or wireless communication network, may serve as the video communication network 100 of FIG. 1 .
  • FIG. 1 indicates that the video content transmitter 102 is in direct communication with the video content receiver 104
  • other intermediate devices such as routers, repeaters, splitters, and other communication circuitry, may help form the communication path between the transmitter 102 and the receiver 104 while remaining with the scope of the invention.
  • the communication path from the video content transmitter 102 to the video content receiver 104 may include an orbiting satellite employing a number of transponders for receiving data from the video content transmitter 102 and relaying that data to the video content receiver 104 , as well as other video content receivers not depicted in FIG. 1 .
  • FIG. 2 presents a flow diagram of a method 200 of generating and using indexing information for video content in the video communication network 100 of FIG. 1 according to an embodiment of the invention.
  • other video communication systems other than the specific system 100 of FIG. 1 may employ the method 200 of FIG. 2 to similar effect.
  • an index stream 112 for a video stream 110 is generated at the video content transmitter 102 (operation 202 ).
  • the video stream 110 includes video frames.
  • a video frame is data representing one still video image of a video presentation, such as a movie, sports presentation, or other program.
  • the index stream 112 includes an index for each of at least some of the video frames within the video stream 110 .
  • the video stream 110 and the index stream 112 are both transmitted from the video content transmitter 102 to the video content receiver 104 (operation 204 ).
  • the video content receiver 104 After receiving and storing the video stream 110 and the index stream 112 (operation 206 ), the video content receiver 104 processes the index stream 112 to locate the video frames within the video stream 110 that are associated with at least some of the indexes of the index stream 112 (operation 208 ). At the video content receiver 104 , at least some of the located video frames of the video stream 110 are presented to an output device (not shown in FIG. 1 ) under control 114 of a user of the video content receiver 104 (operation 210 ).
  • FIG. 2 While the operations of FIG. 2 are depicted as being executed in a serial fashion, other orders of execution, including concurrent execution of two or more operations, are also possible.
  • the operations 202 , 204 associated with the transmitter 102 may proceed concurrently with each other, as well as with the operations 206 - 210 associated with the receiver 104 .
  • any of the operations 206 - 210 of the receiver 104 may be executed concurrently.
  • either or both of the video stream 110 and the index stream 112 maybe partially or completely stored prior to the index stream 112 is being processed and used to present video frames to the output device independently of the receiving and storing process.
  • a computer-readable storage medium may have encoded thereon instructions for a processor or other control circuitry to direct the video content transmitter 102 and/or the video content receiver 104 to implement the method 200 .
  • the video content transmitter 102 which is normally supplied or operated by a video service provider, such as a satellite or cable broadcast television service provider, assumes the responsibility of generating indexing data for its audio/video programs to allow the use of at least some playback trick modes, thus relieving the video content receiver 104 of this duty. Therefore, processing power within the video content receiver 104 may be conserved for other tasks, such as responding to user trick mode commands.
  • a video service provider such as a satellite or cable broadcast television service provider
  • FIG. 3 is a block diagram of an audio/video communication network 300 according to another embodiment of the invention.
  • the network 300 includes an audio/video content transmitter 302 and an audio/video content receiver 304 .
  • the network 300 of FIG. 3 may be a satellite or cable television broadcast network, or any other type of communication network capable of transmitting video data and related indexing data from the transmitter 302 to the receiver 304 .
  • the transmitter 302 and receiver 304 may employ multiple transmitters 302 and receivers 304 within the network 300 .
  • the audio/video content transmitter 302 generates and transmits a transport data stream 310 including audio and video streams, along with an index stream and possibly other metadata associated with the audio and video streams.
  • a transport stream is a transport stream (TS) associated with one of the Motion Picture Experts Group (MPEG) audio/video data formats, such as MPEG-2 or MPEG-4, although other data stream formats may be utilized in other implementations.
  • MPEG Motion Picture Experts Group
  • Each of the audio, video, index, and other metadata streams may be associated with a particular audio/video service, such as a specific television channel or network.
  • the transmitted transport stream 310 may include audio, video, index, and other streams associated with other audio/video services as well.
  • the audio/video content receiver 304 of FIG. 3 receives the transmitted transport stream 310 , and then processes the included index stream to locate various video frames within an associated video stream carried in the transmitted transport stream 310 . At least some of the located video frames are then presented under user control 314 of the receiver 304 , which may include the use of the video trick modes discussed above.
  • Each facet of the generation and use of the index stream in associated with the video stream is described in greater detail hereinafter.
  • FIG. 4 illustrates the audio/video content transmitter 302 of FIG. 3 according to one implementation.
  • the audio/video content transmitter 302 includes an indexing circuit 410 , a multiplexer 412 , and a transmitter circuit 414 .
  • Other components for performing other functions within the transmitter 302 not specifically described herein may also be included in other examples.
  • the indexing circuit 410 receives an audio/video stream 408 as input.
  • the audio/video stream 408 may be MPEG-2 or MPEG-4 audio/video data formatted in an MPEG-TS (transport stream) format, an MPEG-PS (program stream) format, or other data formatting means.
  • the audio/video stream 408 may be delivered to the audio/video content transmitter 302 over an Internet Protocol (IP) connection, an Asynchronous Serial Interface (ASI) connection, or other connection capable of carrying the audio/video stream 408 .
  • IP Internet Protocol
  • ASI Asynchronous Serial Interface
  • the audio/video stream 408 is received from an external source, such as an audio/video programming network, and supplied directly to the indexing circuit 410 .
  • the audio/video stream 408 may be decoded, altered, and re-encoded prior to being presented to the indexing circuit 410 .
  • the audio/video stream 408 may be generated within the audio/video content transmitter 302 .
  • the audio/video stream 408 may include audio and video streams for multiple audio/video services, such as multiple television stations or programming outlets.
  • an audio/video stream 408 carrying multiple services ultimately may be transmitted from the audio/video content transmitter 302 to the audio/video content receiver 304 by way of a single transponder of an orbiting satellite.
  • the indexing circuit 410 processes the incoming audio/video data stream 408 to generate an index stream 422 describing the audio/video data stream 408 .
  • the index stream 422 includes a number of indexes, each of which describes the location of a separate video frame within the audio/video data stream 408 .
  • Each index may describe either or both of an absolute location of the associated video frame within the audio/video stream 408 , and a relative location of the associated video frame compared to another indexed video frame of the stream 408 . This location information may be presented by way of a timestamp, an amount of data, or some other format.
  • the index stream 422 is generated in a format compatible with the audio/video stream 408 , such as an MPEG-2 or MPEG-4 format stream. A specific example of an index and related data is discussed further below in conjunction with FIGS. 8A-8C .
  • the indexing circuit 410 may be implemented in a number of ways.
  • the indexing circuit 410 may be a hardware circuit embodied in one or more application-specific integrated circuits (ASICs) or other types of logic circuitry.
  • the indexing circuit 410 may be a more general-purpose computer processing system programmed with software configured to perform the indexing task.
  • a combination of specialized hardware and general-purpose algorithmic processing circuitry, such as a microprocessor, microcontroller, digital signal processor (DSP) or the like, driven by software configured to implement the indexing process, may serve as another indexing circuit 410 arrangement.
  • the indexing circuit 410 may represent its own circuit, or may be included in within circuitry performing other functions, such as MPEG compression.
  • the indexing circuit 410 may be configured to detect the various services and generate a separate index stream 422 for each service.
  • multiple indexing circuits 410 may be employed, one for each of the separate audio/video services represented in the audio/video stream 408 .
  • one or more additional indexing circuits 410 may be utilized as redundant circuits to be employed in the event a primary indexing circuit 410 becomes inoperative.
  • the indexing circuit 410 may be configured to selectively generate portions of an audio/video stream 408 at one time or another based upon a received command from an operator or other person or system associated with the audio/video content transmitter 302 .
  • the indexing circuit 410 may also be programmed or commanded to generate index data for only select audio/video services represented in the audio/video stream 408 .
  • Other options regarding the generation of indexing information for at least some portion of the received audio/video stream 408 may be provided in further implementations.
  • the multiplexer 412 combines the index stream 422 and associated audio/video stream 408 into a single audio/video transport stream 424 , such as an MPEG-TS or MPEG-PS stream, as mentioned earlier.
  • the index stream 422 and the audio/video stream 408 are multiplexed by way of time-division multiplexing (TDM) of packets of both streams 408 , 422 .
  • TDM time-division multiplexing
  • different types of multiplexing may be employed to combine the streams 408 , 422 .
  • the multiplexer 412 may also generate and insert any formatting consistent with the transport stream 424 format.
  • the multiplexer 412 may perform any other processing required of the transport stream 424 or the individual streams 408 , 422 , such as data scrambling to prevent unauthorized access to the streams 408 , 422 .
  • the resulting transport stream 424 then may be carried over an IP connection, ASI connection, or other suitable connection to the transmitter circuit 414 .
  • the transmitter circuit 414 receives the transport stream 424 and transmits the transport stream 424 according to a format or technology commensurate with the network 300 in which the audio/video content transmitter 302 operates, resulting in the transmitted transport stream 310 .
  • the transmitter circuit 414 generates and amplifies a signal of a particular frequency carrying the transport stream 424 which may be forwarded to an uplink antenna for transmission to a particular satellite transponder.
  • Other types of audio/video communication networks, such as cable television networks, may employ modulation, amplification, and other signal conditioning techniques appropriate for that type of network.
  • the audio/video content receiver 304 receives the transmitted transport stream 310 transmitted from the audio/video content transmitter 302 .
  • the audio/video content receiver 304 includes a receiver circuit 510 , data storage 512 , an output interface 514 , a user interface 516 , and control circuitry 518 .
  • Other components not explicitly depicted in FIG. 5 may also be included in other versions of the receiver 304 , but are not discussed herein to focus on the components most closely involved with the inventive concepts discussed below.
  • the receiver circuit 510 is configured to receive the transmitted transport stream 310 from the audio/video content transmitter 302 and convert the stream 310 to a form more usable for processing within the audio/video content receiver 304 . Such conversion may include frequency down-conversion, amplification, reformatting, and other functions.
  • the receiver circuit 510 may be coupled with or include a hyperboloid antenna combined with a low-noise block-converter/feedhorn (LNBF), which collects and amplifies the incoming signals carrying the transmitted transport stream 302 , and down-converts the signals from microwave frequencies to intermediate frequencies. Concurrently, the receiver circuit 510 may also receive and process other transport streams not specifically indicated in FIG.
  • LNBF low-noise block-converter/feedhorn
  • the receiver circuit 510 may also include tuner circuitry for selecting one or more of the audio/video data services being carried by the transmitted transport stream 310 or other transport streams, descrambler circuitry for descrambling streams, and other circuitry.
  • the resulting streams are then forwarded to the data storage 512 for recording thereon.
  • the data storage 512 under the direction of the control circuitry 518 , provides the DVR functionality for the audio/video content receiver 512 , as well as provides the memory by which the user may employ trick mode playback.
  • the control circuitry 518 may condition or alter the streams received from the receiver circuit 510 before causing those streams to be stored in the data storage 512 .
  • the audio/video streams may be separated from their associated index streams, processed, and stored separately on the data storage 512 .
  • the data storage 512 may be include any type of data storage medium, including solid state memory, such as static and/or dynamic random access memory (RAM) or flash memory; and direct access memory, such as magnetic and optical disk drives.
  • the output interface 518 is configured to decode and/or convert the audio/video streams stored in the data storage 512 and deliver them, or portions thereof, to an output device 550 , such as a television, video monitor, audio/video receiver, and so on, in a format usable by the device 550 for presentation to the user or subscriber.
  • the video portion of the audio/video streams may be delivered by way of a modulated video cable connection, a composite or component video RCA-style (Radio Corporation of America) connection, and a Digital Video Interface (DVI) or High-Definition Multimedia Interface (HDMI) connection.
  • the audio portion may be transported over a monaural or stereo audio RCA-style connection, or over an HDMI connection.
  • Other audio/video formats and related connections may be employed in other embodiments.
  • the user interface 516 provides means by which a user may issue the commands 314 to control the operation of the audio/video content receiver 305 , including the playback of the audio/video streams stored in the data storage 512 , such as normal-speed playback and trick mode playback.
  • the user interface 516 may provide a number of different speeds of playback in both the forward and reverse directions, such as 4 ⁇ (i.e., four times normal playback speed), 15 ⁇ , 60 ⁇ , and 300 ⁇ , although other speeds may be implemented.
  • the user interface 516 allows control of other trick modes, such as pause, forward or reverse slow-motion playback, forward or reverse frame-by-frame playback, and skip forward or reverse, in like fashion.
  • the user interface 516 may provide either or both of a control panel connection located directly on a surface of the receiver 506 , and a remote control interface.
  • the remote control interface may receive commands from a remote control device 552 by way of commands transmitted over a radio frequency (RF) or infrared (IR) frequency band.
  • RF radio frequency
  • IR infrared
  • the control circuitry 512 is configured to control and interact with the receiver circuit 510 , the data storage 512 , the output interface 514 , and the user interface 516 according to the commands and other input received by way of the user interface 516 .
  • the control circuitry 518 processes the stored index streams in conjunction with their associated audio/video streams to locate “access point” frames in the audio/video stream for performing the trick modes that do not require playback of every video frame in a section of the audio/video stream. A more detailed discussion regarding this process is provided below in conjunction with FIGS. 6-9 .
  • the control circuitry 518 may include one or more processors, such as a microprocessor, microcontroller, or digital signal processor (DSP), configured to execute instructions directing the processor to perform the functions discussed in greater detail below.
  • the control circuitry 518 may also include memory or data storage adapted to contain such instructions.
  • the control circuitry 518 may be strictly hardware-based logic, or may include a combination of hardware, firmware, and/or software elements.
  • the video stream 600 includes a series of compressed video frames 602 .
  • Each intra-coded frame (“I-frame”, denoted by an “I” in FIG. 6 ) contains all of the information required to produce its own frame for presentation on the output device 550 .
  • non-intra-coded frames depend on information contained in adjacent frames in addition to its own data in order to be reconstructed for presentation to the user. More specifically, P-frames rely on data from the most recent I-frame or P-frame, while B-frames depend on data from the most recent or immediately following I-frame or P-frame to be reconstructed.
  • the process typically begins at an I-frame, as I-frames are the only frames not dependent upon a nearby frame for reconstruction, and thus may serve as initial access points for video playback within the video stream 600 .
  • the location of the I-frames within the video stream 600 are indexed in an index stream generated in the audio/video content transmitter 302 .
  • P-frames and/or B-frames of the video stream 600 may also be indexed to provide greater access to various points within the video stream 600 .
  • P-frames and B-frames provided greater compression than I-frames due to their dependence on other video frames, so they generally outnumber I-frames by a significant margin.
  • I-frames are periodically employed to allow for playback access points throughout the video stream 600 .
  • the video stream 600 typically includes an I-frame followed by multiple P-frames and B-frames, in a repetitive fashion, as illustrated in FIG. 6 .
  • Each set of frames beginning with an I-frame is denoted a “group-of-pictures” (GOP) 604 .
  • GOP group-of-pictures
  • the number of frames in each GOP 604 may be fixed or variable.
  • each GOP 604 begins with and includes only a single I-frame
  • the amount of data in each GOP 604 may be recorded in the index stream to acquire a subsequent or previous I-frame based on a known I-frame in the video stream 600 during trick mode operation.
  • the audio/video content receiver 304 separates and stores various audio/video streams and associated index streams from the transmitted transport stream 310 for recording on the data storage 512 , typically as files.
  • a packet identifier PID
  • PID packet identifier
  • One type of such data is a Program Map Table (PMT), which is associated with a particular video program or service.
  • PMT Program Map Table
  • the PMT indicates which PIDs of the transmitted transport stream 310 correspond with the video stream for that service, as well as the one or more audio streams, the index stream, and other data associated with that video stream.
  • the video stream, the audio streams, and any separate metadata streams are each referred to as a packetized elementary stream (PES).
  • PES packetized elementary stream
  • the audio/video content receiver 304 may separate at least some of the incoming MPEG-TS packets into their appropriate PESes, and stores at least some of those streams as files on the data storage 512 .
  • FIG. 7 illustrates an audio/video file 700 , which includes video transport packets 702 and associated audio transport packets 704 for a particular program or service as received at the audio/video content receiver 304 .
  • other metadata may be stored in the audio/video file 700 .
  • the audio transport packets 704 may be stored in a separate file.
  • the video transport packets 702 include the I-frame, P-frame, and B-frame data discussed above in connection with FIG. 6 .
  • fewer audio transport packets 704 are required, and are interspersed among the video transport packets 702 so that their location within the file is proximate to the video transport packets 702 with which they are to be presented to the user by way of the output device 550 .
  • FIG. 8A graphically illustrates an index transport packet 800 provided within the transmitted transport stream 310 of FIG. 3 according to one embodiment.
  • Such packets 800 appear sparingly within the transport stream 310 , as the amount of data required to sufficiently index an associated audio/video file 700 is much less than that of the audio/video file 700 itself.
  • each index transport packet 800 is placed in the transport stream 310 near the corresponding audio/video stream data in some implementations to minimize latency between the two.
  • the index transport packet 800 begins with a transport packet header 802 of four bytes, a pointer byte 804 , an index section 806 of 52 bytes, and transport packet pad bytes 808 (such as all bytes of FF in hexadecimal notation).
  • the packet header 802 includes a PID indicating that the packet 802 is an index packet.
  • the pointer byte 804 may be employed as a pointer to other data, but may also be reserved or not used in other implementations.
  • the index section 806 includes the index information and related data, as shown more specifically in FIG. 8B .
  • the pad bytes 808 fill out the remainder of the packet 800 so that only a single index is stored within the packet 800 , thus allowing each index to be placed close to its associated audio/video data in the transmitted transport stream 310 .
  • the index transport packet 800 is 188 bytes in length to match the length of each of the video transport packets 702 and the audio transport packets 704 of the transmitted transport stream 310 .
  • FIG. 8B provides a more detailed representation of the index section 806 of the index transport packet 800 in one implementation.
  • Many of the fields shown therein align with corresponding fields of MPEG sections provided in the video transport packets 702 and audio transport packets 702 to engender processing compatibility of the index sections 806 with audio sections, video sections, and other metadata sections.
  • the index section 806 begins with a one-byte table ID 810 , which in one embodiment is set to EE in hexadecimal notation to distinguish the index section from other types of sections.
  • the one-bit section syntax indicator 812 may be set to one, while the one-bit private indicator 814 may be set to either zero or one.
  • the two-bit reserved areas 816 , 824 may be set to 11 in binary notation.
  • the twelve-bit section length field 818 , 820 stores the value of 49 (i.e., 31 in hexadecimal notation).
  • the two-byte table ID extension 822 may be set to any value.
  • the five-bit version number 826 may be set to zero, while the single-bit current/next indicator 828 may hold a value of one.
  • the one-byte section number field 830 and the one-byte last section number field 832 may both contain a zero value.
  • the index 834 which includes all of the relevant index information for a particular video frame of the audio/video data file 700 , is forty bytes in length in this particular implementation, and is described in greater detail with respect to FIG. 8C .
  • the 32-bit cyclic redundancy check (CRC) field 836 provides error detection capability of the entire index section 806 , thus allowing the audio/video content receiver 304 to determine if the received index information has been corrupted during transmission from the audio/video content transmitter 302 . While specific example values for most of these fields 810 - 836 are noted above, the fields 810 - 836 may include different while remaining with in the scope of the invention.
  • FIG. 8C presents a more detailed view of the index area 834 of the index section 806 shown in FIG. 8B according to one example of the invention.
  • the index 834 includes an eight-byte sequence number 840 , a four-byte I-frame transport packet header 842 , a four-byte presentation time stamp (PTS) field 844 , an eight-byte byte count 846 , a four-byte last I-frame size field 848 , a four-byte last GOP size field 850 , and an eight-byte elapsed time value 852 , for a total of forty bytes.
  • PTS presentation time stamp
  • Each of these fields may aid in locating the particular I-frame associated with the index 834 .
  • other or different fields may be utilized in the index 834 in other implementations to facilitate the same functionality.
  • the sequence number 840 indicates the order in which the indexes 834 are generated for a service.
  • the first index 834 generated for a service may have a sequence number 840 of zero or one, with each index 834 for the same service having a sequence number 840 one greater than the previous sequence number 840 .
  • the control circuitry 518 of the audio/video content receiver 304 may then use the sequence numbers 840 to determine whether any of the indexes 834 received in the transport stream 310 are missing.
  • the I-frame transport packet header 842 is a copy of the header for the video transport packet 702 that contains the beginning of the I-frame associated with this index 834 . As such, the I-frame transport packet header 842 may be compared against the header of the video transport packet 702 containing the start of the associated I-frame as verification that the association between the I-frame and the index 834 is proper.
  • the byte count 846 holds a running count of the number of audio and video bytes for the audio/video service in question at the point the I-frame associated with the index 834 is generated. In one implementation, the byte count 846 is reset to zero whenever the MPEG encoder employed to generate the video stream 600 is reset. The byte count 846 may facilitate finding the video transport packet 702 data associated with this index, as the difference in byte count 846 between two consecutive indexes 834 should be identical or similar to the distance in bytes between their associated I-frames in the audio/video file 700 .
  • the PTS 844 represents a portion of the PTS stored in a metadata field of at least one of the video transport packets 702 containing the I-frame associated with the index 834 .
  • the PTS of the transport packet 702 indicates the time by which the associated video frame is to be decoded and presented for display to the user. In the particular example of FIG. 8C , the PTS 844 holds the most significant 32 bits of the PTS stored with the I-frame.
  • the PTS field 844 of the index 834 thus provides the control circuitry 518 of the audio/video content receiver 304 with a way of initially matching the index 834 with its I-frame packet 702 , thus providing initial synchronization between the indexes 834 and the audio/video file 700 located in the data storage 512 .
  • the last I-frame size 848 is the size in bytes of the transport packets 702 , 704 of the audio/video file 700 that span the video I-frame immediately prior to the I-frame associated with the index 834 , including any audio transport packets 704 interspersed throughout. As each transport packet in this specific implementation is 188 bytes, the last I-frame size 848 is a multiple of 188 bytes.
  • the last I-frame size 848 may be useful from the standpoint that the last GOP size 850 (described immediately below) minus the last I-frame size 848 provides the distance in bytes from the end of the last I-frame to the start of the current I-frame associated with the index 834 , thus facilitating the transition from one I-frame to the next when presenting the I-frames in a trick mode. Such functionality is useful if the indexes 834 and the audio/video file 700 are synchronized, as discussed above.
  • the last GOP size 850 is the size in bytes of the transport packets 702 , 704 of the audio/video file 700 that span the previous GOP (i.e., group of pictures) immediately prior to the I-frame associated with this index 834 , including any audio transport packets 704 interspersed throughout.
  • each transport packet in the present implementation is 188 bytes, so the last GOP size 850 is also a multiple of 188 bytes.
  • the last GOP size 850 facilitates jumping or spanning from the start of the previous I-frame to the start of the current I-frame in the audio/video file 700 .
  • the elapsed time field 852 contains the number of milliseconds elapsed since the byte count field 846 equaled zero. In some cases, the elapsed time 852 may provide another way of determining whether any of the video transport packets 702 of the audio/file 700 are missing, thus possibly causing a lack of synchronization or alignment between the audio/video file 700 and the indexes 834 .
  • the audio/video file 700 shown in the embodiment of FIG. 7 includes video transport packets 702 stored in the data storage 512 in chronological order, with corresponding audio transport packets 704 interspersed throughout in one example, although other storage arrangement are also possible as long as the index information for that service reflects those arrangements.
  • the index information may be stored in a number of ways in the data storage 512 .
  • the entire transport stream for the indexes 834 including each of the index transport packets 800 (depicted in FIG. 8A ) for the associated A/V file 700 , may be recorded in the data storage 512 as a single file, thus reducing the amount of preprocessing required before storing the indexes 834 .
  • each index transport packet 800 may be stored therein, thus reducing the amount of space consumed by the index information in the data storage 512 .
  • FIG. 9 provides a simplified flow diagram of the operation of the audio/video content receiver 304 regarding the use of trick mode playback.
  • the receiver 304 processes the indexes 834 to navigate the audio/video file 700 to find specific I-frames which may be read from the data storage 512 , which are then decoded and presented to the output device 550 by way of the output interface 514 .
  • the receiver 304 Under the direction of its control circuitry 518 , synchronizes the index file with the audio/video file 700 by matching an index 834 with its corresponding I-frame in the audio/video file 700 (operation 902 ).
  • This operation is often initiated upon a user command 314 , such as a command to initiate playback of the audio/video file 700 .
  • the location of the first I-frame of the audio/video file 700 is known with relative certainty, so matching the first I-frame with the first index 834 is a preferred method for synchronizing the audio/video file 700 .
  • the PTS field 844 of the index 834 is matched with the PTS of the video transport packet 702 to ensure proper matching.
  • the I-frame transport packet header field 842 of the index 834 may be searched within the audio/video file 700 to identify candidate video transport packets 702 of the audio/video file 700 before performing a more detailed search of those packets 702 using the PTS field 844 .
  • a marker such as a bookmark, may be generated to indicate both the exact location of the start of the I-frame and the associated index 834 .
  • a bookmark may be stored in a separate table or other data structure for access by the control circuitry 518 when a subsequent synchronizing operation 902 is performed.
  • the control circuitry 518 may then navigate the audio/video file 700 to perform any of several trick modes for playback under user control 314 (operation 904 ). For example, for a forward scan, the control circuitry 518 may search ahead any number of I-frames in the audio/video file 700 by accessing the next index 834 , using one or more values in the index 834 , such as the last GOP size 850 , the last I-frame size 848 , the byte count 846 , or some combination thereof, to find the next I-frame in the file 700 . This process may then be repeated a number of times commensurate with the number of I-frames indicated by the user control 314 .
  • control circuitry 518 may also utilize this same basic process to perform reverse-direction trick modes as well. As with the synchronization operation 902 , the control circuitry 518 may bookmark I-frames found during navigation of the audio/video file 700 for later use. In other implementations, other types of frames, such as P-frames and/or B-frames, may be indexed and employed for navigation as discussed above.
  • the control circuitry 518 may attempt to match another index 834 with its corresponding I-frame in the audio/video file 700 to again attain synchronization (operation 902 ).
  • bookmarks may be employed to mark off areas of the audio/video file 700 or associated index information to prevent a subsequent occurrence of the failed navigation.
  • other frames indexed or not in the vicinity of the I-frame or other indexed frame may also be decoded and presented to the user. For example, in 4 ⁇ fast-forward mode, both I-frames and nearby P-frames may be displayed. In forward slow-motion mode, I-frames, P-frames, and B-frames may all be decoded and presented. Variations for other types of trick modes may also be employed in other embodiments.
  • control circuitry 518 may read and decode the I-frame and present the I-frame to the output device 550 by way of the output interface 514 (operation 908 ).
  • the control circuitry 518 may also control how long (i.e., the number of frame times) that the I-frame is presented to the user. This amount of time may vary depending on the speed of the trick mode involved, with faster trick modes possibly involving longer presentation times due to their more disjointed appearance to the user.
  • the user may designate an end to the current trick mode, such as by pressing a stop, play, pause, or channel change button, or by initiating some other command by way of the remote control device 552 (operation 910 ).
  • the control circuitry 518 may stop the processing of the index information to allow standard forward playback mode, static display of an I-frame in pause mode, and other display operations.
  • the control circuitry 518 may then continue the method 900 using newer parameters commensurate with the new mode.
  • At least some embodiments as described herein for generating index information in a video content transmitter, and using that information in a video content receiver, result in more processing bandwidth in the receiver being available to perform trick modes and other functions since the receiver need not generate the index information.
  • This advantage is experienced in all of the receivers associated with a video content communication system, such as the thousands or millions of receivers associated with satellite or cable television broadcast networks. Instead, relatively few processing systems in direct control of the video service provider need provide the processing power required to generate the index information. Further, the communication bandwidth needed to transmit the index transport packets from the content transmitter to the content receiver is small in comparison to the associated audio and video packets of the provided service.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Television Signal Processing For Recording (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A method of generating and using indexing information for video content is presented. In the method, an index stream for a video stream is generated at a video content transmitter. The video stream includes video frames, and the index stream includes an index for each of at least some of the video frames within the video stream. The video stream and the index stream are transmitted from the video content transmitter to a video content receiver, which receives and stores the video and index streams. Further, at the video content receiver, the index stream is processed to locate the video frames within the video stream that are associated with at least some of the indexes of the index stream, and at least some of the located video frames of the video stream are presented to an output device under control of a user of the video content receiver.

Description

    BACKGROUND
  • A popular and effective way of accessing television programming is via a satellite or cable television broadcast receiver or “set-top box” designed to receive television video and audio data from an orbiting satellite or a cable television headend. By employing such a set-top box, a subscriber to a satellite or cable television service provider may receive any of several hundred programming channels providing news, sports, entertainment, and the like for live program viewing. In addition, the set-top box may include a digital video recorder (DVR) for recording programs for subsequent playback at a time more convenient for the subscriber. A DVR often incorporates a magnetic disk drive or similar data storage technology for storing the recorded programs.
  • Incorporating a DVR in a set-top box provides the further advantage of allowing the subscriber to implement “trick modes”, which are video playback modes other than the familiar normal-speed, forward playback mode. Trick modes may include forward and reverse scanning of video at faster-than-normal speeds, forward and reverse slow-motion playback, forward and reverse frame-by-frame review, and pause, each of which is controlled by the subscriber, typically by way of a remote control device. Further, as a set-top box DVR normally records programs specifically indicated by the subscriber for later viewing, as well as programs currently being delivered or broadcast to the set-top box, the subscriber may use the trick modes on essentially any program being viewed.
  • To implement the forward and reverse fast-scanning modes, the DVR is often designed to display to the subscriber disjoint frames of the video program, each for a short period of time in rapid succession, so that the subscriber or user may discern the progress of the scan. To that end, the DVR or associated set-top box may generate indexing information indicating various access points into the video program being viewed at which complete video frames may be retrieved. Such information is often necessary, as video programs are typically encoded in a compressed format in which many of the video frames are encoded presuming known information about the previous or subsequent frame. Such “predicted” frames are thus typically poor candidates for presentation during a fast-scanning trick mode due to the lack of complete information being provided for that frame. At least some of the other trick modes may benefit from the use of this indexing information as well.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Many aspects of the present disclosure may be better understood with reference to the following drawings. The components in the drawings are not necessarily depicted to scale, as emphasis is instead placed upon clear illustration of the principles of the disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views. Also, while several embodiments are described in connection with these drawings, the disclosure is not limited to the embodiments disclosed herein. On the contrary, the intent is to cover all alternatives, modifications, and equivalents.
  • FIG. 1 is a simplified block diagram of a video communication network according to an embodiment of the invention.
  • FIG. 2 is a flow diagram of a method according to an embodiment of the invention for generating and using indexing information for video content in the video communication system of FIG. 1.
  • FIG. 3 is a block diagram of an audio/video communication network according to an embodiment of the invention.
  • FIG. 4 is a block diagram of an audio/video content transmitter of FIG. 3 according to an embodiment of the invention.
  • FIG. 5 is a block diagram of an audio/video content receiver of FIG. 3 according to an embodiment of the invention.
  • FIG. 6 is a graphical representation of an example video stream employed in an embodiment of the invention.
  • FIG. 7 is a graphical representation of an example audio/video file employed in an embodiment of the invention.
  • FIG. 8A is a graphical representation of an index transport packet according to an embodiment of the invention.
  • FIG. 8B is a graphical representation of an index section within the index metadata transport packet of FIG. 8A according to an embodiment of the invention.
  • FIG. 8C is a graphical representation of an index within the index section of FIG. 8B according to an embodiment of the invention.
  • FIG. 9 is a flow diagram of a method according to an embodiment of the invention for processing an audio/video file and associated index file in an audio/video content receiver.
  • DETAILED DESCRIPTION
  • The enclosed drawings and the following description depict specific embodiments of the invention to teach those skilled in the art how to make and use the best mode of the invention. For the purpose of teaching inventive principles, some conventional aspects have been simplified or omitted. Those skilled in the art will appreciate variations of these embodiments that fall within the scope of the invention. Those skilled in the art will also appreciate that the features described below can be combined in various ways to form multiple embodiments of the invention. As a result, the invention is not limited to the specific embodiments described below, but only by the claims and their equivalents.
  • FIG. 1 is a simplified block diagram of a video communication network 100 according to an embodiment of the invention. The video communication network 100 includes a video content transmitter 102 and a video content receiver 104. In one embodiment, the video communication network 100 may be a satellite television broadcast network, with the video content transmitter 102 existing as part of a satellite uplink center incorporating a satellite headend, and the video content receiver 104 being embodied as a satellite television receiver or set-top box. In another example, the video communication network 100 may be a cable television broadcast network, with the video content transmitter 102 incorporated in a cable headend, while the video content receiver 104 is a cable television receiver or set-top box. Other video communication networks, such as terrestrial (over-the-air) television broadcast networks, computer networks (such as the Internet), or any other wired or wireless communication network, may serve as the video communication network 100 of FIG. 1.
  • Further, while FIG. 1 indicates that the video content transmitter 102 is in direct communication with the video content receiver 104, other intermediate devices, such as routers, repeaters, splitters, and other communication circuitry, may help form the communication path between the transmitter 102 and the receiver 104 while remaining with the scope of the invention. For example, in the case of a satellite television broadcast network, the communication path from the video content transmitter 102 to the video content receiver 104 may include an orbiting satellite employing a number of transponders for receiving data from the video content transmitter 102 and relaying that data to the video content receiver 104, as well as other video content receivers not depicted in FIG. 1.
  • FIG. 2 presents a flow diagram of a method 200 of generating and using indexing information for video content in the video communication network 100 of FIG. 1 according to an embodiment of the invention. However, other video communication systems other than the specific system 100 of FIG. 1 may employ the method 200 of FIG. 2 to similar effect.
  • In the method 200, an index stream 112 for a video stream 110 is generated at the video content transmitter 102 (operation 202). The video stream 110 includes video frames. For example, a video frame is data representing one still video image of a video presentation, such as a movie, sports presentation, or other program. The index stream 112 includes an index for each of at least some of the video frames within the video stream 110. The video stream 110 and the index stream 112 are both transmitted from the video content transmitter 102 to the video content receiver 104 (operation 204).
  • After receiving and storing the video stream 110 and the index stream 112 (operation 206), the video content receiver 104 processes the index stream 112 to locate the video frames within the video stream 110 that are associated with at least some of the indexes of the index stream 112 (operation 208). At the video content receiver 104, at least some of the located video frames of the video stream 110 are presented to an output device (not shown in FIG. 1) under control 114 of a user of the video content receiver 104 (operation 210).
  • While the operations of FIG. 2 are depicted as being executed in a serial fashion, other orders of execution, including concurrent execution of two or more operations, are also possible. For example, the operations 202, 204 associated with the transmitter 102 may proceed concurrently with each other, as well as with the operations 206-210 associated with the receiver 104. Similarly, any of the operations 206-210 of the receiver 104 may be executed concurrently. Also, either or both of the video stream 110 and the index stream 112 maybe partially or completely stored prior to the index stream 112 is being processed and used to present video frames to the output device independently of the receiving and storing process. In another embodiment, a computer-readable storage medium may have encoded thereon instructions for a processor or other control circuitry to direct the video content transmitter 102 and/or the video content receiver 104 to implement the method 200.
  • As a result, the video content transmitter 102, which is normally supplied or operated by a video service provider, such as a satellite or cable broadcast television service provider, assumes the responsibility of generating indexing data for its audio/video programs to allow the use of at least some playback trick modes, thus relieving the video content receiver 104 of this duty. Therefore, processing power within the video content receiver 104 may be conserved for other tasks, such as responding to user trick mode commands. Other advantages may be recognized from the various implementations of the invention discussed in greater detail below.
  • FIG. 3 is a block diagram of an audio/video communication network 300 according to another embodiment of the invention. The network 300 includes an audio/video content transmitter 302 and an audio/video content receiver 304. As with the video communication network 100 of FIG. 1, the network 300 of FIG. 3 may be a satellite or cable television broadcast network, or any other type of communication network capable of transmitting video data and related indexing data from the transmitter 302 to the receiver 304. Further, while only one transmitter 302 and receiver 304 are shown, other implementations may employ multiple transmitters 302 and receivers 304 within the network 300.
  • As shown in FIG. 3, the audio/video content transmitter 302 generates and transmits a transport data stream 310 including audio and video streams, along with an index stream and possibly other metadata associated with the audio and video streams. An example of such a data stream is a transport stream (TS) associated with one of the Motion Picture Experts Group (MPEG) audio/video data formats, such as MPEG-2 or MPEG-4, although other data stream formats may be utilized in other implementations. Each of the audio, video, index, and other metadata streams may be associated with a particular audio/video service, such as a specific television channel or network. Further, the transmitted transport stream 310 may include audio, video, index, and other streams associated with other audio/video services as well.
  • The audio/video content receiver 304 of FIG. 3 receives the transmitted transport stream 310, and then processes the included index stream to locate various video frames within an associated video stream carried in the transmitted transport stream 310. At least some of the located video frames are then presented under user control 314 of the receiver 304, which may include the use of the video trick modes discussed above. Each facet of the generation and use of the index stream in associated with the video stream is described in greater detail hereinafter.
  • FIG. 4 illustrates the audio/video content transmitter 302 of FIG. 3 according to one implementation. The audio/video content transmitter 302 includes an indexing circuit 410, a multiplexer 412, and a transmitter circuit 414. Other components for performing other functions within the transmitter 302 not specifically described herein may also be included in other examples.
  • As shown in FIG. 4, the indexing circuit 410 receives an audio/video stream 408 as input. In one embodiment, the audio/video stream 408 may be MPEG-2 or MPEG-4 audio/video data formatted in an MPEG-TS (transport stream) format, an MPEG-PS (program stream) format, or other data formatting means. Further, the audio/video stream 408 may be delivered to the audio/video content transmitter 302 over an Internet Protocol (IP) connection, an Asynchronous Serial Interface (ASI) connection, or other connection capable of carrying the audio/video stream 408. As depicted in FIG. 4, the audio/video stream 408 is received from an external source, such as an audio/video programming network, and supplied directly to the indexing circuit 410. In another example, the audio/video stream 408 may be decoded, altered, and re-encoded prior to being presented to the indexing circuit 410. Alternatively, the audio/video stream 408 may be generated within the audio/video content transmitter 302.
  • Additionally, the audio/video stream 408 may include audio and video streams for multiple audio/video services, such as multiple television stations or programming outlets. In the case of a satellite broadcast network, an audio/video stream 408 carrying multiple services ultimately may be transmitted from the audio/video content transmitter 302 to the audio/video content receiver 304 by way of a single transponder of an orbiting satellite.
  • The indexing circuit 410 processes the incoming audio/video data stream 408 to generate an index stream 422 describing the audio/video data stream 408. More specifically, the index stream 422 includes a number of indexes, each of which describes the location of a separate video frame within the audio/video data stream 408. Each index may describe either or both of an absolute location of the associated video frame within the audio/video stream 408, and a relative location of the associated video frame compared to another indexed video frame of the stream 408. This location information may be presented by way of a timestamp, an amount of data, or some other format. In one example, the index stream 422 is generated in a format compatible with the audio/video stream 408, such as an MPEG-2 or MPEG-4 format stream. A specific example of an index and related data is discussed further below in conjunction with FIGS. 8A-8C.
  • The indexing circuit 410 may be implemented in a number of ways. For example, the indexing circuit 410 may be a hardware circuit embodied in one or more application-specific integrated circuits (ASICs) or other types of logic circuitry. In another implementation, the indexing circuit 410 may be a more general-purpose computer processing system programmed with software configured to perform the indexing task. A combination of specialized hardware and general-purpose algorithmic processing circuitry, such as a microprocessor, microcontroller, digital signal processor (DSP) or the like, driven by software configured to implement the indexing process, may serve as another indexing circuit 410 arrangement. Further, the indexing circuit 410 may represent its own circuit, or may be included in within circuitry performing other functions, such as MPEG compression.
  • As the audio/video stream 408 may contain streams for multiple audio/video services, as mentioned above, the indexing circuit 410 may be configured to detect the various services and generate a separate index stream 422 for each service. In another example, multiple indexing circuits 410 may be employed, one for each of the separate audio/video services represented in the audio/video stream 408. Furthermore, one or more additional indexing circuits 410 may be utilized as redundant circuits to be employed in the event a primary indexing circuit 410 becomes inoperative.
  • The indexing circuit 410 may be configured to selectively generate portions of an audio/video stream 408 at one time or another based upon a received command from an operator or other person or system associated with the audio/video content transmitter 302. The indexing circuit 410 may also be programmed or commanded to generate index data for only select audio/video services represented in the audio/video stream 408. Other options regarding the generation of indexing information for at least some portion of the received audio/video stream 408 may be provided in further implementations.
  • As the indexing circuit 410 generates the index stream 422, the multiplexer 412 combines the index stream 422 and associated audio/video stream 408 into a single audio/video transport stream 424, such as an MPEG-TS or MPEG-PS stream, as mentioned earlier. In this environment, the index stream 422 and the audio/video stream 408 are multiplexed by way of time-division multiplexing (TDM) of packets of both streams 408, 422. However, under other transport stream formats, different types of multiplexing may be employed to combine the streams 408, 422. The multiplexer 412 may also generate and insert any formatting consistent with the transport stream 424 format. In addition, the multiplexer 412 may perform any other processing required of the transport stream 424 or the individual streams 408, 422, such as data scrambling to prevent unauthorized access to the streams 408, 422. The resulting transport stream 424 then may be carried over an IP connection, ASI connection, or other suitable connection to the transmitter circuit 414.
  • The transmitter circuit 414 receives the transport stream 424 and transmits the transport stream 424 according to a format or technology commensurate with the network 300 in which the audio/video content transmitter 302 operates, resulting in the transmitted transport stream 310. For example, in a satellite television broadcast network, the transmitter circuit 414 generates and amplifies a signal of a particular frequency carrying the transport stream 424 which may be forwarded to an uplink antenna for transmission to a particular satellite transponder. Other types of audio/video communication networks, such as cable television networks, may employ modulation, amplification, and other signal conditioning techniques appropriate for that type of network.
  • The audio/video content receiver 304, an example of which is depicted in FIG. 5, receives the transmitted transport stream 310 transmitted from the audio/video content transmitter 302. The audio/video content receiver 304 includes a receiver circuit 510, data storage 512, an output interface 514, a user interface 516, and control circuitry 518. Other components not explicitly depicted in FIG. 5 may also be included in other versions of the receiver 304, but are not discussed herein to focus on the components most closely involved with the inventive concepts discussed below.
  • The receiver circuit 510 is configured to receive the transmitted transport stream 310 from the audio/video content transmitter 302 and convert the stream 310 to a form more usable for processing within the audio/video content receiver 304. Such conversion may include frequency down-conversion, amplification, reformatting, and other functions. In the case of a satellite audio/video communication network, the receiver circuit 510 may be coupled with or include a hyperboloid antenna combined with a low-noise block-converter/feedhorn (LNBF), which collects and amplifies the incoming signals carrying the transmitted transport stream 302, and down-converts the signals from microwave frequencies to intermediate frequencies. Concurrently, the receiver circuit 510 may also receive and process other transport streams not specifically indicated in FIG. 5 from other satellite transponders in the same manner. As presented in FIG. 5, the receiver circuit 510 may also include tuner circuitry for selecting one or more of the audio/video data services being carried by the transmitted transport stream 310 or other transport streams, descrambler circuitry for descrambling streams, and other circuitry.
  • The resulting streams are then forwarded to the data storage 512 for recording thereon. The data storage 512, under the direction of the control circuitry 518, provides the DVR functionality for the audio/video content receiver 512, as well as provides the memory by which the user may employ trick mode playback. The control circuitry 518, described more fully below, may condition or alter the streams received from the receiver circuit 510 before causing those streams to be stored in the data storage 512. For example, the audio/video streams may be separated from their associated index streams, processed, and stored separately on the data storage 512. The data storage 512 may be include any type of data storage medium, including solid state memory, such as static and/or dynamic random access memory (RAM) or flash memory; and direct access memory, such as magnetic and optical disk drives.
  • The output interface 518 is configured to decode and/or convert the audio/video streams stored in the data storage 512 and deliver them, or portions thereof, to an output device 550, such as a television, video monitor, audio/video receiver, and so on, in a format usable by the device 550 for presentation to the user or subscriber. For example, the video portion of the audio/video streams may be delivered by way of a modulated video cable connection, a composite or component video RCA-style (Radio Corporation of America) connection, and a Digital Video Interface (DVI) or High-Definition Multimedia Interface (HDMI) connection. The audio portion may be transported over a monaural or stereo audio RCA-style connection, or over an HDMI connection. Other audio/video formats and related connections may be employed in other embodiments.
  • The user interface 516 provides means by which a user may issue the commands 314 to control the operation of the audio/video content receiver 305, including the playback of the audio/video streams stored in the data storage 512, such as normal-speed playback and trick mode playback. As to the trick modes, the user interface 516 may provide a number of different speeds of playback in both the forward and reverse directions, such as 4× (i.e., four times normal playback speed), 15×, 60×, and 300×, although other speeds may be implemented. The user interface 516 allows control of other trick modes, such as pause, forward or reverse slow-motion playback, forward or reverse frame-by-frame playback, and skip forward or reverse, in like fashion.
  • The user interface 516 may provide either or both of a control panel connection located directly on a surface of the receiver 506, and a remote control interface. The remote control interface may receive commands from a remote control device 552 by way of commands transmitted over a radio frequency (RF) or infrared (IR) frequency band. Different communication methods, such as those employing optical or acoustic transmission of remote commands, may be used in other implementations.
  • The control circuitry 512 is configured to control and interact with the receiver circuit 510, the data storage 512, the output interface 514, and the user interface 516 according to the commands and other input received by way of the user interface 516. With respect to audio/video playback, the control circuitry 518 processes the stored index streams in conjunction with their associated audio/video streams to locate “access point” frames in the audio/video stream for performing the trick modes that do not require playback of every video frame in a section of the audio/video stream. A more detailed discussion regarding this process is provided below in conjunction with FIGS. 6-9.
  • The control circuitry 518 may include one or more processors, such as a microprocessor, microcontroller, or digital signal processor (DSP), configured to execute instructions directing the processor to perform the functions discussed in greater detail below. The control circuitry 518 may also include memory or data storage adapted to contain such instructions. In another implementation, the control circuitry 518 may be strictly hardware-based logic, or may include a combination of hardware, firmware, and/or software elements.
  • To preface the explanation of the indexing process provided below, an example of a portion of an MPEG-2 or MPEG-4 video stream 600 is illustrated in FIG. 6. The video stream 600 includes a series of compressed video frames 602. Each intra-coded frame (“I-frame”, denoted by an “I” in FIG. 6) contains all of the information required to produce its own frame for presentation on the output device 550. Oppositely, predictive-coded frames (“P-frames”, indicated by a “P” in FIG. 6) and bidirectionally-predictive-coded frames (“B-frames”, signified by a “B” in FIG. 6) (collectively, non-intra-coded frames) depend on information contained in adjacent frames in addition to its own data in order to be reconstructed for presentation to the user. More specifically, P-frames rely on data from the most recent I-frame or P-frame, while B-frames depend on data from the most recent or immediately following I-frame or P-frame to be reconstructed. As a result, to begin decompressing and decoding video at a point within the video stream 600, the process typically begins at an I-frame, as I-frames are the only frames not dependent upon a nearby frame for reconstruction, and thus may serve as initial access points for video playback within the video stream 600.
  • Given the importance of the I-frames as access points, the location of the I-frames within the video stream 600 are indexed in an index stream generated in the audio/video content transmitter 302. In other embodiments, P-frames and/or B-frames of the video stream 600 may also be indexed to provide greater access to various points within the video stream 600.
  • Typically, P-frames and B-frames provided greater compression than I-frames due to their dependence on other video frames, so they generally outnumber I-frames by a significant margin. However, I-frames are periodically employed to allow for playback access points throughout the video stream 600. As a result, the video stream 600 typically includes an I-frame followed by multiple P-frames and B-frames, in a repetitive fashion, as illustrated in FIG. 6. Each set of frames beginning with an I-frame is denoted a “group-of-pictures” (GOP) 604. Depending on the particular strategy used, the number of frames in each GOP 604 may be fixed or variable. As each GOP 604 begins with and includes only a single I-frame, the amount of data in each GOP 604, as well as related information, may be recorded in the index stream to acquire a subsequent or previous I-frame based on a known I-frame in the video stream 600 during trick mode operation.
  • As mentioned above, the audio/video content receiver 304 separates and stores various audio/video streams and associated index streams from the transmitted transport stream 310 for recording on the data storage 512, typically as files. For example, presuming the use of MPEG-TS packets for the transmitted transport stream 310, a packet identifier (PID) provided in a header area of each TS packet signifies the type of data held in each packet. One type of such data is a Program Map Table (PMT), which is associated with a particular video program or service. The PMT, in turn, indicates which PIDs of the transmitted transport stream 310 correspond with the video stream for that service, as well as the one or more audio streams, the index stream, and other data associated with that video stream. The video stream, the audio streams, and any separate metadata streams are each referred to as a packetized elementary stream (PES). Using the PMTs and associated PIDs, the audio/video content receiver 304 may separate at least some of the incoming MPEG-TS packets into their appropriate PESes, and stores at least some of those streams as files on the data storage 512.
  • FIG. 7 illustrates an audio/video file 700, which includes video transport packets 702 and associated audio transport packets 704 for a particular program or service as received at the audio/video content receiver 304. In other embodiments, other metadata may be stored in the audio/video file 700. In another implementation, the audio transport packets 704 may be stored in a separate file. The video transport packets 702 include the I-frame, P-frame, and B-frame data discussed above in connection with FIG. 6. Typically, fewer audio transport packets 704 are required, and are interspersed among the video transport packets 702 so that their location within the file is proximate to the video transport packets 702 with which they are to be presented to the user by way of the output device 550. Given the variable length of the various frames stored in the packets 702, as well as the existence of the audio transport packets 704 in the file 700, locating the I-frames within the file 700 would be time-consuming without the associated indexing information provided by the audio/video content transmitter 302.
  • FIG. 8A graphically illustrates an index transport packet 800 provided within the transmitted transport stream 310 of FIG. 3 according to one embodiment. Such packets 800 appear sparingly within the transport stream 310, as the amount of data required to sufficiently index an associated audio/video file 700 is much less than that of the audio/video file 700 itself. Also, each index transport packet 800 is placed in the transport stream 310 near the corresponding audio/video stream data in some implementations to minimize latency between the two.
  • In FIG. 8A, the index transport packet 800 begins with a transport packet header 802 of four bytes, a pointer byte 804, an index section 806 of 52 bytes, and transport packet pad bytes 808 (such as all bytes of FF in hexadecimal notation). The packet header 802 includes a PID indicating that the packet 802 is an index packet. The pointer byte 804 may be employed as a pointer to other data, but may also be reserved or not used in other implementations. The index section 806 includes the index information and related data, as shown more specifically in FIG. 8B. The pad bytes 808 fill out the remainder of the packet 800 so that only a single index is stored within the packet 800, thus allowing each index to be placed close to its associated audio/video data in the transmitted transport stream 310. In the specific embodiment of FIG. 8A, the index transport packet 800 is 188 bytes in length to match the length of each of the video transport packets 702 and the audio transport packets 704 of the transmitted transport stream 310.
  • FIG. 8B provides a more detailed representation of the index section 806 of the index transport packet 800 in one implementation. Many of the fields shown therein align with corresponding fields of MPEG sections provided in the video transport packets 702 and audio transport packets 702 to engender processing compatibility of the index sections 806 with audio sections, video sections, and other metadata sections. The index section 806 begins with a one-byte table ID 810, which in one embodiment is set to EE in hexadecimal notation to distinguish the index section from other types of sections. In one implementation, the one-bit section syntax indicator 812 may be set to one, while the one-bit private indicator 814 may be set to either zero or one. The two-bit reserved areas 816, 824 may be set to 11 in binary notation. The twelve-bit section length field 818, 820 stores the value of 49 (i.e., 31 in hexadecimal notation). The two-byte table ID extension 822 may be set to any value. Further, the five-bit version number 826 may be set to zero, while the single-bit current/next indicator 828 may hold a value of one. The one-byte section number field 830 and the one-byte last section number field 832 may both contain a zero value. The index 834, which includes all of the relevant index information for a particular video frame of the audio/video data file 700, is forty bytes in length in this particular implementation, and is described in greater detail with respect to FIG. 8C. The 32-bit cyclic redundancy check (CRC) field 836 provides error detection capability of the entire index section 806, thus allowing the audio/video content receiver 304 to determine if the received index information has been corrupted during transmission from the audio/video content transmitter 302. While specific example values for most of these fields 810-836 are noted above, the fields 810-836 may include different while remaining with in the scope of the invention.
  • FIG. 8C presents a more detailed view of the index area 834 of the index section 806 shown in FIG. 8B according to one example of the invention. The index 834 includes an eight-byte sequence number 840, a four-byte I-frame transport packet header 842, a four-byte presentation time stamp (PTS) field 844, an eight-byte byte count 846, a four-byte last I-frame size field 848, a four-byte last GOP size field 850, and an eight-byte elapsed time value 852, for a total of forty bytes. Each of these fields may aid in locating the particular I-frame associated with the index 834. However, other or different fields may be utilized in the index 834 in other implementations to facilitate the same functionality.
  • The sequence number 840 indicates the order in which the indexes 834 are generated for a service. In one example, the first index 834 generated for a service may have a sequence number 840 of zero or one, with each index 834 for the same service having a sequence number 840 one greater than the previous sequence number 840. The control circuitry 518 of the audio/video content receiver 304 may then use the sequence numbers 840 to determine whether any of the indexes 834 received in the transport stream 310 are missing.
  • The I-frame transport packet header 842 is a copy of the header for the video transport packet 702 that contains the beginning of the I-frame associated with this index 834. As such, the I-frame transport packet header 842 may be compared against the header of the video transport packet 702 containing the start of the associated I-frame as verification that the association between the I-frame and the index 834 is proper.
  • The byte count 846 holds a running count of the number of audio and video bytes for the audio/video service in question at the point the I-frame associated with the index 834 is generated. In one implementation, the byte count 846 is reset to zero whenever the MPEG encoder employed to generate the video stream 600 is reset. The byte count 846 may facilitate finding the video transport packet 702 data associated with this index, as the difference in byte count 846 between two consecutive indexes 834 should be identical or similar to the distance in bytes between their associated I-frames in the audio/video file 700.
  • The PTS 844 represents a portion of the PTS stored in a metadata field of at least one of the video transport packets 702 containing the I-frame associated with the index 834. The PTS of the transport packet 702 indicates the time by which the associated video frame is to be decoded and presented for display to the user. In the particular example of FIG. 8C, the PTS 844 holds the most significant 32 bits of the PTS stored with the I-frame. The PTS field 844 of the index 834 thus provides the control circuitry 518 of the audio/video content receiver 304 with a way of initially matching the index 834 with its I-frame packet 702, thus providing initial synchronization between the indexes 834 and the audio/video file 700 located in the data storage 512.
  • The last I-frame size 848 is the size in bytes of the transport packets 702, 704 of the audio/video file 700 that span the video I-frame immediately prior to the I-frame associated with the index 834, including any audio transport packets 704 interspersed throughout. As each transport packet in this specific implementation is 188 bytes, the last I-frame size 848 is a multiple of 188 bytes. The last I-frame size 848 may be useful from the standpoint that the last GOP size 850 (described immediately below) minus the last I-frame size 848 provides the distance in bytes from the end of the last I-frame to the start of the current I-frame associated with the index 834, thus facilitating the transition from one I-frame to the next when presenting the I-frames in a trick mode. Such functionality is useful if the indexes 834 and the audio/video file 700 are synchronized, as discussed above.
  • The last GOP size 850 is the size in bytes of the transport packets 702, 704 of the audio/video file 700 that span the previous GOP (i.e., group of pictures) immediately prior to the I-frame associated with this index 834, including any audio transport packets 704 interspersed throughout. As with the last I-frame size 848, each transport packet in the present implementation is 188 bytes, so the last GOP size 850 is also a multiple of 188 bytes. The last GOP size 850 facilitates jumping or spanning from the start of the previous I-frame to the start of the current I-frame in the audio/video file 700.
  • Lastly, the elapsed time field 852 contains the number of milliseconds elapsed since the byte count field 846 equaled zero. In some cases, the elapsed time 852 may provide another way of determining whether any of the video transport packets 702 of the audio/file 700 are missing, thus possibly causing a lack of synchronization or alignment between the audio/video file 700 and the indexes 834.
  • As discussed above, the audio/video file 700 shown in the embodiment of FIG. 7 includes video transport packets 702 stored in the data storage 512 in chronological order, with corresponding audio transport packets 704 interspersed throughout in one example, although other storage arrangement are also possible as long as the index information for that service reflects those arrangements. Similarly, the index information may be stored in a number of ways in the data storage 512. For example, the entire transport stream for the indexes 834, including each of the index transport packets 800 (depicted in FIG. 8A) for the associated A/V file 700, may be recorded in the data storage 512 as a single file, thus reducing the amount of preprocessing required before storing the indexes 834. In other embodiments, only the index section 806 (shown in FIG. 8B), or only the index 834 (FIG. 8C), of each index transport packet 800 may be stored therein, thus reducing the amount of space consumed by the index information in the data storage 512.
  • FIG. 9 provides a simplified flow diagram of the operation of the audio/video content receiver 304 regarding the use of trick mode playback. Generally, the receiver 304 processes the indexes 834 to navigate the audio/video file 700 to find specific I-frames which may be read from the data storage 512, which are then decoded and presented to the output device 550 by way of the output interface 514.
  • To allow the use of the trick modes, the receiver 304, under the direction of its control circuitry 518, synchronizes the index file with the audio/video file 700 by matching an index 834 with its corresponding I-frame in the audio/video file 700 (operation 902). This operation is often initiated upon a user command 314, such as a command to initiate playback of the audio/video file 700. In many cases, the location of the first I-frame of the audio/video file 700 is known with relative certainty, so matching the first I-frame with the first index 834 is a preferred method for synchronizing the audio/video file 700. In one implementation, the PTS field 844 of the index 834 is matched with the PTS of the video transport packet 702 to ensure proper matching. In cases in which the location of the I-frame is more difficult, such as an I-frame associated with an index 834 referencing the middle of the audio/video file 700, the I-frame transport packet header field 842 of the index 834 may be searched within the audio/video file 700 to identify candidate video transport packets 702 of the audio/video file 700 before performing a more detailed search of those packets 702 using the PTS field 844.
  • Once the match of the index 834 and the appropriate I-frame is made, a marker, such as a bookmark, may be generated to indicate both the exact location of the start of the I-frame and the associated index 834. Such a bookmark may be stored in a separate table or other data structure for access by the control circuitry 518 when a subsequent synchronizing operation 902 is performed.
  • Once the synchronization or alignment has occurred, the control circuitry 518 may then navigate the audio/video file 700 to perform any of several trick modes for playback under user control 314 (operation 904). For example, for a forward scan, the control circuitry 518 may search ahead any number of I-frames in the audio/video file 700 by accessing the next index 834, using one or more values in the index 834, such as the last GOP size 850, the last I-frame size 848, the byte count 846, or some combination thereof, to find the next I-frame in the file 700. This process may then be repeated a number of times commensurate with the number of I-frames indicated by the user control 314. For example, higher trick mode speeds are likely to require jumps involving more I-frames being skipped. Other data within the various indexes 834, such as the byte count 846, the elapsed time 852, the PTS 844, or other information, may also be employed to determine how many I-frames should be skipped for a particular trick mode. The control circuitry 518 may also utilize this same basic process to perform reverse-direction trick modes as well. As with the synchronization operation 902, the control circuitry 518 may bookmark I-frames found during navigation of the audio/video file 700 for later use. In other implementations, other types of frames, such as P-frames and/or B-frames, may be indexed and employed for navigation as discussed above.
  • If the navigation was not successful (operation 906), possibly as a result of one or more missing indexes 834 or transport packets 702, 704, the control circuitry 518 may attempt to match another index 834 with its corresponding I-frame in the audio/video file 700 to again attain synchronization (operation 902). In some implementations, bookmarks may be employed to mark off areas of the audio/video file 700 or associated index information to prevent a subsequent occurrence of the failed navigation. In some trick modes, other frames (indexed or not) in the vicinity of the I-frame or other indexed frame may also be decoded and presented to the user. For example, in 4× fast-forward mode, both I-frames and nearby P-frames may be displayed. In forward slow-motion mode, I-frames, P-frames, and B-frames may all be decoded and presented. Variations for other types of trick modes may also be employed in other embodiments.
  • Once control circuitry 518 has successfully navigated to the desired I-frame in the audio/video file 700, the control circuitry 518 may read and decode the I-frame and present the I-frame to the output device 550 by way of the output interface 514 (operation 908). The control circuitry 518 may also control how long (i.e., the number of frame times) that the I-frame is presented to the user. This amount of time may vary depending on the speed of the trick mode involved, with faster trick modes possibly involving longer presentation times due to their more disjointed appearance to the user.
  • At any point in the process, the user may designate an end to the current trick mode, such as by pressing a stop, play, pause, or channel change button, or by initiating some other command by way of the remote control device 552 (operation 910). As a result, the control circuitry 518 may stop the processing of the index information to allow standard forward playback mode, static display of an I-frame in pause mode, and other display operations. In the case the user merely alters the speed or direction of the desired high-speed trick mode, the control circuitry 518 may then continue the method 900 using newer parameters commensurate with the new mode.
  • At least some embodiments as described herein for generating index information in a video content transmitter, and using that information in a video content receiver, result in more processing bandwidth in the receiver being available to perform trick modes and other functions since the receiver need not generate the index information. This advantage is experienced in all of the receivers associated with a video content communication system, such as the thousands or millions of receivers associated with satellite or cable television broadcast networks. Instead, relatively few processing systems in direct control of the video service provider need provide the processing power required to generate the index information. Further, the communication bandwidth needed to transmit the index transport packets from the content transmitter to the content receiver is small in comparison to the associated audio and video packets of the provided service.
  • While several embodiments of the invention have been discussed herein, other implementations encompassed by the scope of the invention are possible. For example, while various embodiments have been described primarily within the context of satellite and cable television broadcast networks, other communication networks not strictly involving the distribution of audio/video programming, such as the Internet and other wide-area networks (WANs), local-area networks (LANs), wireless communication networks, and the like, may benefit from application of the various concepts presented herein. Further, while the various implementations described herein make extensive use of the MPEG audio/video format standards, other audio/video formats employing data compression techniques may also be utilized in the embodiments discussed above. In addition, aspects of one embodiment disclosed herein may be combined with those of alternative embodiments to create further implementations of the present invention. Thus, while the present invention has been described in the context of specific embodiments, such descriptions are provided for illustration and not limitation. Accordingly, the proper scope of the present invention is delimited only by the following claims and their equivalents.

Claims (20)

1. A method of generating and using indexing information for video content, the method comprising:
at the video content transmitter, generating an index stream for a video stream, wherein the video stream comprises video frames, and wherein the index stream comprises an index for each of at least some of the video frames within the video stream; and
at the video content transmitter, transmitting the video stream and the index stream to a video content receiver.
2. The method of claim 1, further comprising:
at the video content receiver, receiving and storing the video stream and the index stream;
at the video content receiver, processing the index stream to locate the video frames within the video stream that are associated with at least some of the indexes of the index stream; and
at the video content receiver, presenting at least some of the located video frames of the video stream to an output device under control of a user of the video content receiver.
3. The method of claim 1, wherein:
the at least some of the video frames comprise intra-coded video frames.
4. The method of claim 1, wherein:
the video stream is accompanied by an associated audio stream.
5. The method of claim 1, wherein:
the video content transmitter comprises one of a satellite television headend transmitter and a cable television headend transmitter.
6. The method of claim 1, wherein:
the video stream and the index stream are transmitted within a transport data stream, wherein the video stream and the index stream are packetized elementary streams of the transport data stream.
7. The method of claim 1, wherein:
each of the at least some of the video frames comprises a timestamp; and
each of the indexes of the index stream comprises at least a portion of the timestamp of the video frame associated with the index.
8. The method of claim 7, further comprising:
at the video content receiver, matching the timestamp of at least one of the indexes with the timestamp of at least one of the video frames of the video stream to synchronize the index stream with the video stream.
9. The method of claim 1, wherein:
each of the indexes of the index stream comprises a sequence number indicating a relative position of the index compared to other indexes of the index stream.
10. The method of claim 9, further comprising:
at the video content receiver, employing the sequence number of each of the indexes of the index stream to determine if any of the indexes of the index stream are missing.
11. The method of claim 1, wherein:
each of the indexes of the index stream comprises a data offset of the associated video frame relative to the video stream.
12. The method of claim 11, wherein:
presenting the at least some of the located video frames of the video stream to the output device is accomplished using at least one of the data offsets of the index stream.
13. The method of claim 1, wherein:
each of the indexes of the index stream comprises a data offset indicating an amount of data of the video stream from the beginning of the video frame associated with an immediately preceding indexed video frame up to, but not including, the video frame associated with the index.
14. The method of claim 13, wherein:
presenting the at least some of the located video frames of the video stream to the output device is accomplished using at least one of the data offsets of the index stream.
15. A video content transmitter, comprising:
an indexing circuit configured to generate an index stream from a video stream, wherein the video stream comprises video frames, and wherein the index stream comprises an index for each of at least some of the video frames within the video stream;
a multiplexer configured to multiplex the video stream and the index stream into a data stream; and
a transmitter circuit configured to transmit the data stream to a video content receiver.
16. The video content transmitter of claim 15, wherein:
the transmitting circuit is configured to transmit the data stream over one of a cable television network and a satellite television network to the video content receiver.
17. A video content receiver, comprising:
a receiver circuit configured to receive a data stream comprising a video stream and an index stream, wherein the video stream comprises video frames, and wherein the index stream comprises indexes, wherein each of the indexes indicates a location of one of the video frames within the video stream;
data storage for storing the data stream and the index stream;
an output interface configured to present the video stream to an output device;
a user interface; and
control circuitry configured to:
process the index stream to locate the video frames within the video stream that are associated with at least some of the indexes of the index stream; and
present at least some of the located video frames of the video stream to the output device by way of the output interface under user control by way of the user interface.
18. The video content receiver of claim 17, wherein, to process the index stream, the control circuitry is configured to:
synchronize the index stream with the video stream by matching a timestamp of at least one video frame with a timestamp of an associated index.
19. The video content receiver of claim 18, wherein, to process the index stream, the control circuitry is configured to:
locate another video frame associated with another index by using the other index to determine a location of the other video frame based on a location of the at least one video frame.
20. The video content receiver of claim 18, wherein, to process the index stream, the control circuitry is configured to:
attempt to locate another video frame associated with another index by using the other index to determine a location of the other video frame based on a location of the at least one video frame; and
if the attempt to locate the other video frame is unsuccessful, synchronize the index stream with the video stream by matching a timestamp of another one of the video frames with a timestamp of an associated index.
US12/424,399 2009-04-15 2009-04-15 Video stream index generation at a video content transmitter Abandoned US20100269147A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US12/424,399 US20100269147A1 (en) 2009-04-15 2009-04-15 Video stream index generation at a video content transmitter
TW99110000A TWI420910B (en) 2009-04-15 2010-03-31 Video stream index generation at a video content transmitter
PCT/US2010/030902 WO2010120787A1 (en) 2009-04-15 2010-04-13 Video stream index generation at a video content transmitter
US13/487,101 US8855468B2 (en) 2009-04-15 2012-06-01 Video stream index generation at a video content transmitter

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/424,399 US20100269147A1 (en) 2009-04-15 2009-04-15 Video stream index generation at a video content transmitter

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/487,101 Continuation US8855468B2 (en) 2009-04-15 2012-06-01 Video stream index generation at a video content transmitter

Publications (1)

Publication Number Publication Date
US20100269147A1 true US20100269147A1 (en) 2010-10-21

Family

ID=42211942

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/424,399 Abandoned US20100269147A1 (en) 2009-04-15 2009-04-15 Video stream index generation at a video content transmitter
US13/487,101 Active 2030-02-14 US8855468B2 (en) 2009-04-15 2012-06-01 Video stream index generation at a video content transmitter

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/487,101 Active 2030-02-14 US8855468B2 (en) 2009-04-15 2012-06-01 Video stream index generation at a video content transmitter

Country Status (3)

Country Link
US (2) US20100269147A1 (en)
TW (1) TWI420910B (en)
WO (1) WO2010120787A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100316351A1 (en) * 2009-06-16 2010-12-16 Samsung Electronics Co. Ltd. System and method for providing video file
US20120121023A1 (en) * 2010-04-14 2012-05-17 Yezdi Antia System and method for multi-carrier multiplexing
US20120254664A1 (en) * 2011-04-04 2012-10-04 Agustawestland S.P.A. Automatic test system for digital display systems
US20130051775A1 (en) * 2011-08-31 2013-02-28 Sensaburo Nakamura Image output apparatus, method of controlling image output apparatus, and program
US8843752B1 (en) 2011-01-24 2014-09-23 Prima Cimema, Inc. Multi-factor device authentication
US8990574B1 (en) 2010-10-06 2015-03-24 Prima Cinema, Inc. Secure device authentication protocol
US20150128200A1 (en) * 2013-11-06 2015-05-07 Broadcom Corporation Recovering channel bonded program streams
CN105493509A (en) * 2013-08-12 2016-04-13 索尼公司 Transmission apparatus, transmission method, reception apparatus, and reception method
US20190066735A1 (en) * 2012-12-17 2019-02-28 Intel Corporation Embedding Thumbnail Information Into Video Streams
WO2019121890A1 (en) * 2017-12-22 2019-06-27 Odass Gbr Method for reducing the computing time of a data processing unit
US10339278B2 (en) 2015-11-04 2019-07-02 Screening Room Media, Inc. Monitoring nearby mobile computing devices to prevent digital content misuse
US10452819B2 (en) 2017-03-20 2019-10-22 Screening Room Media, Inc. Digital credential system
US10805663B2 (en) * 2018-07-13 2020-10-13 Comcast Cable Communications, Llc Audio video synchronization
US11277672B1 (en) * 2020-10-07 2022-03-15 Dish Wireless L.L.C. 5G with ATSC 3.0 network extender

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI449430B (en) * 2011-08-01 2014-08-11 Hides Inc High - bandwidth audio and video transmission processing system
GB2501759B (en) * 2012-05-04 2019-06-26 Saturn Licensing Llc Receiving audio/video content
US10567765B2 (en) 2014-01-15 2020-02-18 Avigilon Corporation Streaming multiple encodings with virtual stream identifiers
US9489387B2 (en) 2014-01-15 2016-11-08 Avigilon Corporation Storage management of data streamed from a video source device
CN105338411A (en) * 2014-07-31 2016-02-17 宇龙计算机通信科技(深圳)有限公司 Video play processing method and system
CN105049954B (en) * 2015-07-02 2020-07-10 中兴通讯股份有限公司 Multicast transmission method, device and system for OTT media
TWI604323B (en) * 2016-11-10 2017-11-01 財團法人工業技術研究院 Method for video indexing and device using the same
US10835823B2 (en) 2018-12-27 2020-11-17 Electronic Arts Inc. Sensory-based dynamic game-state configuration
US10860345B2 (en) * 2019-03-27 2020-12-08 Electronic Arts Inc. System for user sentiment tracking
CN113127679A (en) 2019-12-30 2021-07-16 阿里巴巴集团控股有限公司 Video searching method and device and index construction method and device
CN113840110B (en) * 2021-11-23 2022-06-28 北京亮亮视野科技有限公司 Video processing method and device for workflow, electronic equipment and storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5594490A (en) * 1994-05-23 1997-01-14 Cable Services Technologies, Inc. System for distributing video/audio files from central location to a plurality of cable headends
US5864682A (en) * 1995-07-14 1999-01-26 Oracle Corporation Method and apparatus for frame accurate access of digital audio-visual information
US20020184637A1 (en) * 2001-05-30 2002-12-05 Perlman Stephen G. System and method for improved multi-stream multimedia transmission and processing
US6850284B2 (en) * 2002-08-27 2005-02-01 Motorola, Inc. Method and apparatus for decoding audio and video information
US6876705B2 (en) * 2001-03-05 2005-04-05 Intervideo, Inc. Systems and methods for decoding of partially corrupted reversible variable length code (RVLC) intra-coded macroblocks and partial block decoding of corrupted macroblocks in a video decoder
US20070300249A1 (en) * 2006-06-22 2007-12-27 Smith Kevin P In-band data recognition and synchronization system
US20090037965A1 (en) * 2007-07-31 2009-02-05 Leonard Tsai Portal based video indexing
US7522816B2 (en) * 2003-08-12 2009-04-21 The Directv Group, Inc. Method and apparatus for navigating content in a personal video recorder
US20100195974A1 (en) * 2009-02-04 2010-08-05 Google Inc. Server-side support for seamless rewind and playback of video streaming

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5949948A (en) * 1995-11-20 1999-09-07 Imedia Corporation Method and apparatus for implementing playback features for compressed video data
US8689265B2 (en) * 1999-03-30 2014-04-01 Tivo Inc. Multimedia mobile personalization system
IL132859A (en) * 1999-11-10 2008-07-08 Nds Ltd System for data stream processing
EP1286351B1 (en) * 2001-08-21 2012-08-08 Thomson Licensing File and content management
US20070050837A1 (en) * 2005-08-25 2007-03-01 Lopez-Estrada Alex A Method, apparatus and system for generating and distributing rich digital bookmarks for digital content navigation
US8214516B2 (en) * 2006-01-06 2012-07-03 Google Inc. Dynamic media serving infrastructure
US7826712B2 (en) * 2006-08-29 2010-11-02 Echostar Technologies Corporation Method and apparatus for receiving, storing, and presenting multimedia programming without indexing prior to storage
KR101443632B1 (en) * 2008-04-11 2014-11-03 엘지전자 주식회사 Device for Recording and Playing Content, sever for managing content location information, information recording media, method for managing content information and recording media for the method

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5594490A (en) * 1994-05-23 1997-01-14 Cable Services Technologies, Inc. System for distributing video/audio files from central location to a plurality of cable headends
US5864682A (en) * 1995-07-14 1999-01-26 Oracle Corporation Method and apparatus for frame accurate access of digital audio-visual information
US6876705B2 (en) * 2001-03-05 2005-04-05 Intervideo, Inc. Systems and methods for decoding of partially corrupted reversible variable length code (RVLC) intra-coded macroblocks and partial block decoding of corrupted macroblocks in a video decoder
US20020184637A1 (en) * 2001-05-30 2002-12-05 Perlman Stephen G. System and method for improved multi-stream multimedia transmission and processing
US7093277B2 (en) * 2001-05-30 2006-08-15 Digeo, Inc. System and method for improved multi-stream multimedia transmission and processing
US6850284B2 (en) * 2002-08-27 2005-02-01 Motorola, Inc. Method and apparatus for decoding audio and video information
US7522816B2 (en) * 2003-08-12 2009-04-21 The Directv Group, Inc. Method and apparatus for navigating content in a personal video recorder
US20070300249A1 (en) * 2006-06-22 2007-12-27 Smith Kevin P In-band data recognition and synchronization system
US20090037965A1 (en) * 2007-07-31 2009-02-05 Leonard Tsai Portal based video indexing
US20100195974A1 (en) * 2009-02-04 2010-08-05 Google Inc. Server-side support for seamless rewind and playback of video streaming

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100316351A1 (en) * 2009-06-16 2010-12-16 Samsung Electronics Co. Ltd. System and method for providing video file
US9264675B2 (en) * 2010-04-14 2016-02-16 Hughes Network Systems, Llc System and method for multi-carrier multiplexing
US20120121023A1 (en) * 2010-04-14 2012-05-17 Yezdi Antia System and method for multi-carrier multiplexing
US8990574B1 (en) 2010-10-06 2015-03-24 Prima Cinema, Inc. Secure device authentication protocol
US8843752B1 (en) 2011-01-24 2014-09-23 Prima Cimema, Inc. Multi-factor device authentication
US20120254664A1 (en) * 2011-04-04 2012-10-04 Agustawestland S.P.A. Automatic test system for digital display systems
US9332252B2 (en) * 2011-04-04 2016-05-03 Agusta Westland S.P.A. Automatic test system for digital display systems
CN102969012A (en) * 2011-08-31 2013-03-13 索尼公司 Image output apparatus, method of controlling image output apparatus, and program
US9117486B2 (en) * 2011-08-31 2015-08-25 Sony Corporation Image output apparatus, method of controlling image output apparatus, and program
US20130051775A1 (en) * 2011-08-31 2013-02-28 Sensaburo Nakamura Image output apparatus, method of controlling image output apparatus, and program
US20190066735A1 (en) * 2012-12-17 2019-02-28 Intel Corporation Embedding Thumbnail Information Into Video Streams
US10777231B2 (en) * 2012-12-17 2020-09-15 Intel Corporation Embedding thumbnail information into video streams
CN105493509A (en) * 2013-08-12 2016-04-13 索尼公司 Transmission apparatus, transmission method, reception apparatus, and reception method
KR20160041909A (en) * 2013-08-12 2016-04-18 소니 주식회사 Transmission apparatus, transmission method, reception apparatus, and reception method
US10542074B2 (en) * 2013-08-12 2020-01-21 Saturn Licensing Llc Stream transmission using MPEG media transport (MMT)
US20160149994A1 (en) * 2013-08-12 2016-05-26 Sony Corporation Transmitting apparatus, transmitting method, receiving apparatus and receiving method
KR102208235B1 (en) 2013-08-12 2021-01-27 소니 주식회사 Transmission apparatus, transmission method, reception apparatus, and reception method
US20150128200A1 (en) * 2013-11-06 2015-05-07 Broadcom Corporation Recovering channel bonded program streams
US9838729B2 (en) * 2013-11-06 2017-12-05 Avago Technologies General Ip (Singapore) Pte. Ltd. Recovering channel bonded program streams
US10395011B2 (en) 2015-11-04 2019-08-27 Screening Room Media, Inc. Monitoring location of a client-side digital content delivery device to prevent digital content misuse
US11853403B2 (en) 2015-11-04 2023-12-26 Sr Labs, Inc. Pairing devices to prevent digital content misuse
US10423762B2 (en) 2015-11-04 2019-09-24 Screening Room Media, Inc. Detecting digital content misuse based on know violator usage clusters
US10430560B2 (en) 2015-11-04 2019-10-01 Screening Room Media, Inc. Monitoring digital content usage history to prevent digital content misuse
US11941089B2 (en) 2015-11-04 2024-03-26 Sr Labs, Inc. Pairing devices to prevent digital content misuse
US10460083B2 (en) 2015-11-04 2019-10-29 Screening Room Media, Inc. Digital credential system
US10409964B2 (en) 2015-11-04 2019-09-10 Screening Room Media, Inc. Pairing devices to prevent digital content misuse
US10339278B2 (en) 2015-11-04 2019-07-02 Screening Room Media, Inc. Monitoring nearby mobile computing devices to prevent digital content misuse
US10417393B2 (en) 2015-11-04 2019-09-17 Screening Room Media, Inc. Detecting digital content misuse based on digital content usage clusters
US11227031B2 (en) 2015-11-04 2022-01-18 Screening Room Media, Inc. Pairing devices to prevent digital content misuse
US10452819B2 (en) 2017-03-20 2019-10-22 Screening Room Media, Inc. Digital credential system
WO2019121890A1 (en) * 2017-12-22 2019-06-27 Odass Gbr Method for reducing the computing time of a data processing unit
US11941007B2 (en) * 2017-12-22 2024-03-26 Odass Gbr Method for reducing the computing time of a data processing unit
US10805663B2 (en) * 2018-07-13 2020-10-13 Comcast Cable Communications, Llc Audio video synchronization
US11979631B2 (en) 2018-07-13 2024-05-07 Comcast Cable Communications, Llc Audio video synchronization
US11277672B1 (en) * 2020-10-07 2022-03-15 Dish Wireless L.L.C. 5G with ATSC 3.0 network extender
US11653070B2 (en) 2020-10-07 2023-05-16 Dish Wireless L.L.C. 5G with ATSC 3.0 network extender

Also Published As

Publication number Publication date
TWI420910B (en) 2013-12-21
US8855468B2 (en) 2014-10-07
US20120301108A1 (en) 2012-11-29
TW201123903A (en) 2011-07-01
WO2010120787A1 (en) 2010-10-21

Similar Documents

Publication Publication Date Title
US8855468B2 (en) Video stream index generation at a video content transmitter
US8136140B2 (en) Methods and apparatus for generating metadata utilized to filter content from a video stream using text data
US7218635B2 (en) Apparatus and method for indexing MPEG video data to perform special mode playback in a digital video recorder and indexed signal associated therewith
US8977106B2 (en) Methods and apparatus for filtering content in a video stream using closed captioning data
US6963612B2 (en) System for detecting start codes in MPEG video streams and method of operating the same
US8965177B2 (en) Methods and apparatus for displaying interstitial breaks in a progress bar of a video stream
US7230652B2 (en) System and method for providing picture-in-picture timebase management
US20100328527A1 (en) Fast Channel Switch Between Digital Television Channels
EP2061239B1 (en) Methods and apparatus for identifying video locations in a video stream using text data
KR100405975B1 (en) Method for stream jump
EP1266521B1 (en) System and method for providing multi-perspective instant replay
AU2001266732A1 (en) System and method for providing multi-perspective instant replay
US8254764B2 (en) Recording apparatus, image reproducing apparatus, and special reproduction method therefor
JP4366038B2 (en) Television broadcast processing apparatus and control method for television broadcast processing apparatus
US11076194B2 (en) Time-shifted playback
US20080145019A1 (en) Video recording and reproducing apparatus and method of reproducing video in the same
KR100708377B1 (en) Decoder for multiple video on receiving digital broadcasting
KR20140118059A (en) Method for reproducing contents
KR20050122410A (en) Method for regenerating broadcast program in digital broadcast regeneration system

Legal Events

Date Code Title Description
AS Assignment

Owner name: ECHOSTAR TECHNOLOGIES LLC, COLORADO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZETTEROWER, CHARLIE WILLIAM;KUMMER, DAVID A.;SIGNING DATES FROM 20090427 TO 20090428;REEL/FRAME:022720/0528

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION