US20170085614A1 - Method for processing multimedia streams - Google Patents

Method for processing multimedia streams Download PDF

Info

Publication number
US20170085614A1
US20170085614A1 US15/278,952 US201615278952A US2017085614A1 US 20170085614 A1 US20170085614 A1 US 20170085614A1 US 201615278952 A US201615278952 A US 201615278952A US 2017085614 A1 US2017085614 A1 US 2017085614A1
Authority
US
United States
Prior art keywords
component
packet
stream
components
logic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/278,952
Inventor
Michael K. Kounnas
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/023,258 external-priority patent/US20060133513A1/en
Application filed by Intel Corp filed Critical Intel Corp
Priority to US15/278,952 priority Critical patent/US20170085614A1/en
Publication of US20170085614A1 publication Critical patent/US20170085614A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/762Media network packet handling at the source 
    • H04L65/602
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/22Parsing or analysis of headers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/30Definitions, standards or architectural aspects of layered protocol stacks
    • H04L69/32Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level
    • H04L69/322Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions
    • H04L69/324Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions in the data link layer [OSI layer 2], e.g. HDLC
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/438Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving encoded video stream packets from an IP network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/4508Management of client data or end-user data
    • H04N21/4516Management of client data or end-user data involving client characteristics, e.g. Set-Top-Box type, software version or amount of memory available

Definitions

  • the invention is related to processing multimedia streams. More specifically, the invention is related to the intelligent selection of algorithmic processes to process multimedia streams.
  • transcoders translate a multimedia stream in one format to a stream in another format so people can communicate with each other easily and still use these varied devices and varied multimedia stream formats. Transcoding currently is utilized in a variety of different scenarios such as for one-way, two-way, or multi-way streams.
  • FIG. 1A describes an example of a transaction that may require one-way transcoding of a multimedia stream.
  • a video stream 100 is sent from a first computer 102 to a second computer 104 .
  • the video from the first computer 102 is encoded into one of many different possible formats including MPEG-2, MPEG-4, GSM (global system for mobile communications), or AMR (adaptive multi-rate) among others.
  • the second computer 104 receives, decodes, and displays the streaming video on a video monitor.
  • a person may want to record a message on their voicemail system to inform callers that they are out of the office. That person may want to play back the message to determine whether they want to save it or record again.
  • the person could be speaking through a G.711 format phone and the voicemail system may be recording the message in a WAV file. Again, this requires a transcoding step.
  • Two-way transcoding is used when two devices are sending streams to each other, such as potentially during a video call or an audio (e.g., telephone) call.
  • FIG. 1B describes an example of a two-way video call that may require two-way transcoding of multimedia streams.
  • a video stream 110 is sent from a first computer 112 to a second computer 114 , and also from the second computer 114 to the first computer 110 .
  • Multi-way transcoding is used when more than two devices are sending streams to each other, such as, for example, during a three-party video conference.
  • FIG. 1C describes an example of a three-way video conference that may require multi-way transcoding.
  • multiple computers are sending and receiving video streams between each other.
  • a first computer 122 , a second computer 124 , and a hand-held mobile computer/cell phone 126 are sending and receiving streams 120 between each other.
  • each of the computers shown in FIGS. 1A, 1B and 1C may encode and decode multimedia streams in different formats.
  • the two computers in FIG. 1C 130 and 132
  • the mobile computer/cellular phone 134 may not be capable of encoding or decoding MPEG-4 and instead encodes and decodes only a GSM-format stream.
  • a transcoding device may be utilized to intercept the streams coming from each of the three devices and translate the streams into each destination device's native format.
  • a video stream is just one of many possible examples of where a transcoder may be needed. For example, if a person using a land-line phone calls his friend on a cellular phone there is a transcoding step necessary.
  • the standard land-line phones typically operate using a G.711 format.
  • the friend who answers the call would typically have a GSM, AMR, or other cellular phone voice format technology.
  • GSM Global System for Mobile communications
  • AMR Access Management Entity
  • real-time transcoding systems usually utilize extremely powerful processors, such as servers with dedicated digital signal processors, to make sure that any form of information in the packet can be transcoded in real-time. In many occasions, this amount of processing power is not required.
  • processors such as servers with dedicated digital signal processors
  • a person who is recording an out-of-office message and then reviews their recordings is not operating in real-time. There is a delay in the range of multiple seconds between when the person speaks into the phone, and when they hear their recorded voice during review.
  • transcoding systems that perform transcoding in a non-real-time, offline manner.
  • a user on a personal computer may translate a VHS tape into an AVI file to store on a DVD.
  • no transcoding is done at all, for example, raw video, voice, or audio format streams can be transmitted across a network. This type of stream is not efficient because raw format streams are uncompressed and require a great deal of network bandwidth.
  • FIG. 1A describes an example of a transaction that may require one-way transcoding of a multimedia stream.
  • FIG. 1B describes an example of a two-way video call that may require two-way transcoding of multimedia streams.
  • FIG. 1C describes an example of a three-way video conference that may require multi-way transcoding.
  • FIG. 2 is a block diagram of one embodiment of a computer system with a multimedia stream processing management system (MSPMS).
  • MSPMS multimedia stream processing management system
  • FIG. 3 describes one embodiment of a multimedia stream processing management system (MSPMS).
  • MSPMS multimedia stream processing management system
  • FIG. 4 describes one embodiment of a method to transcode a multimedia stream packet with a single algorithmic process.
  • FIG. 5 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • FIG. 6 describes another embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • FIG. 7 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and one buffer.
  • FIG. 8 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and multiple buffers.
  • FIG. 9 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes and multiple buffers.
  • FIG. 10 describes one embodiment of an algorithmic process omitting individual components based on component priority and processing power limitations.
  • FIG. 11 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream to real-time and non-real-time algorithmic processes.
  • FIG. 12 is a flow diagram of another embodiment of a method for assigning components of a multimedia stream to real-time processes.
  • FIG. 13 is a flow diagram of an embodiment of a method for omitting components of a multimedia stream from the transcoding process.
  • FIG. 14 is a flow diagram of an embodiment of a method to skip the transcoding of a stream.
  • FIG. 15 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream that do not require real-time processing, but require processing before a determined maximum time delay, to processes that are able to process within the determined maximum time delay.
  • Embodiments of an effective method for processing multimedia streams are disclosed.
  • numerous specific details are set forth. However, it is understood that embodiments may be practiced without these specific details. In other instances, well-known elements, applications, and techniques have not been discussed in detail in order to avoid obscuring the present invention.
  • FIG. 2 is a block diagram of one embodiment of a computer system with a multimedia stream processing management system (MSPMS).
  • the computer system includes a processor 200 , a memory controller hub (MCH) 202 , and an I/O controller hub (ICH) 210 .
  • the MCH 202 and the ICH 210 comprise a chipset.
  • the processor 200 is coupled to the MCH 202 via a host bus.
  • the MCH 202 is coupled to system memory 204 .
  • system memory may be synchronous dynamic random access memory (SDRAM), double data rate SDRAM (DDR-SDRAM), Rambus DRAM (RDRAM), or one of many other formats of main system memory.
  • SDRAM synchronous dynamic random access memory
  • DDR-SDRAM double data rate SDRAM
  • RDRAM Rambus DRAM
  • the computer system includes a MSPMS 206 stored in system memory 204 and operated/executed by the processor 200 .
  • the MSPMS 206 may control a number of management decisions relating to processing a multimedia stream such as the speed the processing requires (e.g., real-time or non-real-time), the priority associated with each of the different components that comprise the stream (e.g., the audio component, the voice component, the video component, the data component, and the control component), and the assignment of the specific device that will actually handle the multimedia stream.
  • the device that actually handles the multimedia stream may be a digital signal processor, a network server, a remote desktop computer, or the computer system shown in FIG. 2 (including processor 200 ) among many other possible processing devices.
  • the MCH 202 is also coupled to a graphics module 208 .
  • the ICH 210 is coupled to a hard drive 212 and an I/O bus 214 .
  • the I/O bus 214 is coupled to a network interface card that interfaces a network.
  • the network may be a local area network, a public phone system, the Internet, a wide area network, a wireless network, or any other network that is capable of delivering data between computing devices.
  • FIG. 3 describes one embodiment of a multimedia stream processing management system (MSPMS).
  • the processing logic 300 of the MSPMS is located between a multimedia stream source 302 and a destination 318 .
  • the source 302 generates, encodes, and sends a multimedia stream across a network to a destination device 318 .
  • the stream is comprised of a number of sequential packets.
  • Each packet may comprise several components.
  • Each component of the packet consists of a type of information regarding the stream.
  • a packet for an audio/video stream can contain a voice component and a separate video component.
  • there may be other components that also make up multimedia streams such as a data component, which sends instructions within the stream, a general audio component, which includes music and background noise information (as opposed to the specific voice information), and a control component, which may include security information, authentication information, and quality of service (QOS) information among other types of control information.
  • Each component within a packet contains a predetermined amount of information for that component. The amount of information per component per packet is usually measured in time. For example, a voice component within an individual packet of a multimedia stream may consist of 100 milliseconds of the recorded voice.
  • each packet includes a header that contains address information to allow the packet to be routed to the correct destination.
  • the MSPMS processing logic 300 intercepts the stream.
  • the stream enters a stream unwrapper unit 304 .
  • the stream unwrapper unit 304 unwraps the stream into its individual components.
  • the individual components of the stream are routed to the process assigning unit 306 .
  • the video component is unwrapped and sent to the process assigning unit 306 .
  • header address information is sent directly from the stream unwrapper unit 304 to the stream wrapper unit 316 .
  • the process assigning unit 306 then assigns each of the one or more stream components to processing logic that includes one or more algorithmic processes ( 308 , 310 , 312 , 314 ).
  • An “algorithmic process” is a method or process that is capable of modifying one or more components of a multimedia stream in some way. Examples of algorithmic processes include a method to transcode a component of a multimedia stream, a method to omit certain components of a multimedia stream, a method to delay the processing of a component of a multimedia stream, a method to buffer a real-time transport protocol (RTP) stream, among others.
  • RTP real-time transport protocol
  • a user may want to leave a voicemail.
  • the user calls a phone number and a voicemail system answers the call.
  • a first algorithmic process may be buffering the actual voice data coming through an RTP stream
  • a second algorithmic process may be monitoring for a dial tone (dual-tone multi-frequency (DTMF) detection) to detect when any buttons on the phone are pressed during the user message
  • a third algorithmic process may be transcoding the RTP stream into a more storage friendly format.
  • DTMF dual-tone multi-frequency
  • the RTP stream buffer and the DTMF detection algorithmic processes must be operated in real-time to create an effective user interaction, otherwise the voicemail will not be useful, whereas the transcoding process just takes information from the buffer on an available basis.
  • the RTP and DTMF algorithmic processes may have a higher priority than the transcoding process and the processes are managed by the MSPMS accordingly.
  • prioritizing algorithmic processes may be accomplished based on the importance of a component regarding user perception.
  • the voicemail user interaction would place a higher priority on functionality to maintain a clear recording of the user's voice. Therefore, buffering an RTP stream is obviously more important than transcoding that stream to a different format because omitting any part of the RTP stream would cause a loss of information and the user's voice would not be audible in part or all of the message.
  • there are N algorithmic processes each of which may accomplish the processing of a stream component in a different way.
  • each algorithmic process prioritizes the processing of components sent to it based on each component's importance.
  • the process assigning unit 306 assigns two or more components to a single algorithmic prioritizing process, the two or more components will be processed in the order of their priority.
  • the priority may be decided based on an algorithm that takes into account factors such as network transmission rates, relative importance of different components for user comprehensibility of the stream, the bandwidth consumption of the component in relationship to the other components, or any number of other factors.
  • the video component in a multimedia stream of multi-component packets usually consumes the most bandwidth and computing power, thus, the algorithmic process may be able to just transcode video in real-time, or it may be able to transcode all of the other components in real-time (possibly audio, voice, data, and control components), therefore, four components may have priority over just one.
  • the decision may be made by the process assigning unit 306 to assign a component to a particular algorithmic process that only processes every other (or every third, fourth, etc.) packetized component, which would result in a lower frame rate.
  • two or more of the N algorithmic processes are running on a single computer system.
  • each of the two or more components also have their own assigned priority levels.
  • the voice component process may have a higher operating priority in the computer system than the video component process.
  • the computer system does not have enough resources to run both processes simultaneously, the video component process will be delayed while the voice component process completes the processing of the voice component of one or more multimedia stream packets.
  • each of the N algorithmic processes runs on its own thread in the local computer system. In this embodiment, the local computer system may assign each thread a certain priority to correspond with the priorities of the components assigned to each process-thread.
  • two or more of the N algorithmic processes run on two or more separate computer systems.
  • the process assigning unit 306 may utilize one or more of the algorithmic processes that run on the computer system with the most available resources.
  • one component may be assigned to two or more algorithmic processes.
  • the process assigning unit 306 divides up the information within a single component in the stream into portions and assigns each portion to a separate algorithmic process.
  • the process assigning unit 306 alternates sending component information between two algorithmic processes on a packet-by-packet level (i.e., a first packet's component is sent to algorithmic process one, a second packet's component is sent to algorithmic process two, a third packet's component is sent to algorithmic process three, etc).
  • the process assigning unit 306 assigns the video component to an algorithmic process to transcode from MPEG-4 to MPEG-2.
  • the video component i.e., the new MPEG-2 format video stream
  • the stream wrapper unit 316 will encapsulate the individual components back into a packetized format to be sent to the destination.
  • the stream wrapper unit 316 will take information directly from the stream unwrapper unit 304 such as the destination address of each component the stream wrapper unit 316 receives as the output of the one or more algorithmic processes.
  • the stream wrapper unit 316 receives information directly from the process assigning unit 306 such as priority information per component and packet combining information per component.
  • Priority information per component allows the stream wrapper unit 316 to prioritize the wrapping and sending of different components that arrive from one or more of the algorithmic processes.
  • Packet combining information per component allows the stream wrapper unit 316 to incorporate multiple components into a single packet or, alternatively, allow an individual component to be wrapped into a separate packet and sent for a reason such as higher priority.
  • the stream wrapper unit 316 receives the newly transcoded MPEG-2 video component and rewraps the component with the destination address information and any other pertinent information and sends the stream on to the destination 318 .
  • the process assigning unit 306 when deciding which algorithmic process to sign a given component to, takes into account information including: the network connectivity and throughput of each computer system that an algorithmic process runs on, the amount of free resources that each computer system has available, the priority of each algorithmic process on its respective computer system, among other information.
  • the process assigning unit 306 receives information regarding each computer system from the algorithmic processes running on each one (the dotted arrows in FIG. 3 ).
  • the process assigning unit 306 (and the MSPMS in general) is capable of determining how to process each component in the stream as efficiently as possible to minimize or eliminate any perceived performance degradation of the stream.
  • the MSPMS may process a component of a stream in real-time, not in real-time, or not process a component at all depending on the priority of the component and the available system resources.
  • FIGS. 4 through 9 illustrate multiple examples of MSPMS processing logic decisions regarding the processing of a multimedia stream with a video and a voice component.
  • the functionality shows that each component in the stream requires transcoding, although, in different embodiments, partial transcoding or no transcoding at all may be allowable for some or all of the video stream.
  • processing the components of a stream may include many other functions other than just transcoding such as buffering an RTP stream, DTMF detection, or any one of many other multimedia stream processing functions.
  • the transcoding examples in FIGS. 4 through 9 are just illustrative examples, which may be substituted for any number of other processes that take place involving a multimedia stream.
  • the examples are illustrative of a single packet within the stream, the embodiments include repeating the process for potentially all the packets within the stream.
  • FIG. 4 describes one embodiment of a method to transcode a multimedia stream packet with a single algorithmic process.
  • a source 500 sends a multimedia stream packet 502 across a network.
  • the packet 502 contains a header 504 that has all the routing and address information regarding the packet's destination.
  • the packet also contains a video component 506 and a voice component 508 .
  • the video component comprises one video frame (i.e., a still image) and the voice component comprises one voice frame (i.e., 100 milliseconds of voice information).
  • the header 404 is removed and a packet 410 containing only the video and voice component is sent to the algorithmic process 412 .
  • the algorithmic process 412 transcodes the information in each component and outputs a new packet containing only transcoded video and voice components.
  • the header 404 bypasses 414 the algorithmic process and may be reattached to the transcoded packet to create a new packet 416 , which is sent to the destination 418 .
  • the single algorithmic process 412 is capable of transcoding (in real-time) a multimedia stream of packets that have video and voice components.
  • the header 404 is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header 404 is not removed from the packet and the packet is sent to the algorithmic process with the header 404 intact.
  • FIG. 5 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • a source 500 sends a multimedia stream packet 502 across a network.
  • the packet 502 contains a video component, and a voice component (as described above in FIG. 4 ).
  • the header is removed and different components within the packet are also separated.
  • the video component 504 is sent to algorithmic process # 1 ( 508 ) for video transcoding and the voice component 506 is sent to algorithmic process # 2 ( 510 ) for voice transcoding.
  • algorithmic processes # 1 ( 508 ) and # 2 ( 510 ) may be located on the same computing device or on different computing devices.
  • the header which bypassed the transcoding process ( 512 and 514 ) may be reattached to separate transcoded video and voice component packets ( 516 and 518 ).
  • the video component packet 516 and voice component packet 518 arrive separately at the destination 520 .
  • the header is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact.
  • the algorithmic process itself may modify the header.
  • FIG. 6 describes another embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • a source 600 sends a multimedia stream packet 602 across a network.
  • the packet 602 contains a video component and a voice component (as described above in FIG. 4 ).
  • the header is removed and different components within the packet are also separated.
  • the video component 604 is sent to algorithmic process # 1 ( 608 ) for video transcoding and the voice component 606 is sent to algorithmic process # 2 ( 610 ) for voice transcoding.
  • algorithmic processes # 1 ( 608 ) and # 2 ( 610 ) complete their respective video and voice component transcoding they reattach the video component and the voice component into one multi-component packet 614 .
  • the header which bypassed the transcoding process may be reattached to the multi-component packet 614 and the packet is sent to the destination 616 .
  • the packet 614 is sent to the destination by the stream wrapper unit ( FIG. 3, 316 ) with just one component to minimize any bottleneck in the algorithmic process that finished first.
  • the header is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact.
  • the algorithmic process itself may modify the header.
  • FIG. 7 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and one buffer.
  • a source 700 sends a multimedia stream packet 702 across a network.
  • the packet 702 contains a video component and a voice component (as described above in FIG. 4 ).
  • the header is removed and different components within the packet are also separated.
  • the voice component 704 has priority over the video component 706 , thus the voice component 704 is sent directly to the algorithmic process 710 while the video component 706 is sent to a buffer 708 to wait for the algorithmic process 710 to finish transcoding the voice component 704 .
  • the buffer 708 releases the stored video component 706 to the algorithmic process 710 for transcoding.
  • the transcoded component is reattached to the header ( 712 or 714 ) and the individual video or voice component packet ( 716 or 718 ) is sent to the destination 720 .
  • the header is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 8 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and multiple buffers.
  • a source 800 sends a multimedia stream packet 802 across a network.
  • the packet 802 contains a video component and a voice component (as described above in FIG. 4 ).
  • the header is removed and different components within the packet are also separated.
  • the priority of the components may be changed in real-time based on certain circumstances. In this case, the priority of the video component and the priority of the voice component may increase or decrease relative to each other during transcoding operations.
  • two buffers are utilized to allow for either component to have the capability of waiting for the algorithmic process 812 to finish transcoding the other component.
  • the video component 804 is sent to buffer # 1 ( 808 ) and the voice component 806 is sent to buffer # 2 ( 810 ).
  • the algorithmic process 812 takes a single component one at a time from either buffer # 1 ( 808 ) or buffer # 2 ( 810 ). The determination of which buffer to take from depends on the priority assigned to the individual component waiting in each buffer. Once the algorithmic process 812 finishes transcoding an individual component, the transcoded component is reattached to the header ( 814 or 816 ) and the individual video or voice component packet ( 818 or 820 ) is sent to the destination 822 .
  • the header is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact.
  • the algorithmic process itself may modify the header.
  • FIG. 9 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes and multiple buffers.
  • a source 900 sends a multimedia stream packet 902 across a network.
  • the packet 902 contains a video component and a voice component (as described above in FIG. 5 ).
  • the header is removed and different components within the packet are also separated.
  • algorithmic processes # 1 ( 912 ) and # 2 ( 914 ) are located on the same computer system, but in separate processes. In this case, the priority of the video transcoding process and the priority of the voice transcoding process may increase or decrease relative not only to each other, but also to other processes running on the computer system.
  • each algorithmic process has its own buffer to allow for both components to have the capability of waiting for the algorithmic processes ( 912 and 914 ) to regain the high priority on the system.
  • the video component 904 is sent to buffer # 1 ( 908 ) and the voice component 906 is sent to buffer # 2 ( 910 ).
  • algorithmic process # 1 ( 912 ) takes individual video components from buffer # 1 ( 908 ) and algorithmic process # 2 ( 914 ) takes individual voice components from buffer # 2 ( 910 ).
  • the transcoded component is reattached to its header ( 916 or 918 ) and the individual video or voice component packet ( 920 or 922 ) is sent to the destination 924 .
  • the header is modified by the stream wrapper unit ( FIG. 3, 316 ) based on information provided from the stream unwrapper unit ( FIG. 3, 304 ) and the process assigning unit ( FIG. 3, 306 ).
  • the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact.
  • the algorithmic process itself may modify the header.
  • FIG. 10 describes one embodiment of an algorithmic process omitting individual components based on component priority and processing power limitations.
  • a stream of video components 1000 i.e., video frames
  • a stream of voice components 1002 i.e. voice frames
  • the algorithmic process 1004 in this embodiment is similar to the algorithmic process described in FIG. 8 .
  • the algorithmic process 1004 does not have the processing power to keep up with every voice frame and every video frame. Thus, if there are no other systems with algorithmic processes capable of taking some of the transcoding work, certain frames must be omitted to maintain the real-time nature of the stream.
  • a stream of four video frames 1000 and a stream of four voice frames 1002 are input into the algorithmic process 1004 .
  • the algorithmic process 1004 transcodes all voice frames and sends them out in order in the voice output stream 1006 .
  • the algorithmic process 1004 also uses its remaining processing power to transcode as many video frames as possible and sends them out in order in the video output stream 1008 .
  • the algorithmic process 1004 is omitting every other video frame to maintain the real-time aspect of the streams. Consequently, the voice stream is clear and intact, and the video stream has its frame rate cut in half.
  • FIG. 11 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream to real-time and non-real-time algorithmic processes.
  • the method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both.
  • the method begins by processing logic determining whether one or more components comprising a multimedia stream require real-time processing (block 1100 ). If processing logic determines that real-time processing is required (block 1102 ) then processing logic assigns the one or more components that require real-time processing to one or more real-time algorithmic processes (block 1104 ). Finally, processing logic assigns any remaining components to one or more non-real-time algorithmic processes (block 1106 ) and the method is finished.
  • FIG. 12 is a flow diagram of another embodiment of a method for assigning components of a multimedia stream to real-time processes.
  • the method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both.
  • processing logic begins by processing logic unwrapping a packet into its one or more components (block 1200 ).
  • processing logic determines the amount of processing required to process each of the one or more components requiring real-time processing within each packet in real-time (block 1202 ).
  • processing logic assigns each of the one or more components requiring real-time processing within the packet to a separate real-time process (block 1204 ) and the method is finished.
  • FIG. 13 is a flow diagram of an embodiment of a method for omitting components of a multimedia stream from the transcoding process.
  • the method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both.
  • processing logic begins by processing logic unwraps a packet into its one or more components (block 1300 ).
  • processing logic determines if all components of the unwrapped packet can be transcoded in real-time (block 1302 ). If they can then the method is finished. Otherwise, processing logic examines the first component (block 1304 ). Once the component is examined, processing logic determines if the component can be omitted (block 1306 ).
  • the determination whether a component can be omitted may be made based on a number of factors such as the processing priority of the component and the number of consecutive component frames that have been omitted. For example, as discussed in FIG. 10 , if the algorithmic process does not have the processing power to process everything in real-time a decision must be made as to which components to omit and when to omit them. Dropping a video component down to half its standard frame rate by omitting every other frame is less objectionable than to drop 15 frames in a row and then transcode 15 frames in a row. The end result is the same number of frames, but the quality of the viewing experience is different for those two separate options.
  • processing logic omits the component (block 1308 ).
  • processing logic determines whether all components have been examined (block 1310 ). If so, the method is finished. Otherwise, processing logic examines the next component (block 1312 ). Eventually, all components of an unwrapped packet will be examined and the method will finish.
  • determining whether to omit one or more components from a stream may include a decision regarding multimedia stream processes other than just transcoding. Thus, transcoding is used as an example and can be substituted with one or more other functions that may be performed on a multimedia stream.
  • FIG. 14 is a flow diagram of an embodiment of a method to skip the transcoding of a stream.
  • the method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both.
  • the method begins by processing logic determining the format of the stream transmitted from the source device (block 1500 ).
  • processing logic determines the format of the stream required by the destination device (block 1502 ).
  • processing logic determines whether the two formats are the same (block 1504 ). If the formats are not the same, processing logic transcodes the stream (block 1506 ) and the process is finished. Otherwise, if the formats are the same, processing logic does not transcode the stream and instead sends the stream directly from the source to the destination (block 1508 ) and the method is finished.
  • FIG. 15 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream that do not require real-time processing, but require processing before a determined maximum time delay, to processes that are able to process within the determined maximum time delay.
  • the method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both.
  • processing logic begins by processing logic determining the maximum delay allowable for each of the components in a packet that do not require real-time processing (block 1500 ).
  • a determination of the maximum allowable processing delay may be made based on a number of factors including the type of transaction the multimedia stream is being used for, the speed of the transmission medium (e.g. the required speed of processing a voicemail attached in an email is limited to the inherent delay of emails), as well as numerous other factors.
  • processing logic assigns each of the components not requiring real-time processing to a separate process, wherein each process delivers at least the amount of processing required to process each respective component in the allowable delayed time (block 1502 ) and the method is finished.
  • the packets dealt with mainly video and voice components the packets contain any one or more of video, audio, voice, data, control, or any other conceivable packet type.
  • the example embodiments referenced packets that included only two components.
  • the packets may contain any conceivable number of components (i.e., greater than or equal to one component).
  • a single algorithmic process was responsible for processing all components of a single type.
  • more than one algorithmic process may be assigned to a single type of component.
  • two or more algorithmic processes can trade off processing duties and each algorithmic process can therefore process every other packetized component of a single type (or every third, fourth, etc.).

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

A method, machine readable medium, system, and device are disclosed. In one embodiment the method comprises determining whether one or more components that comprise a multimedia stream require real-time processing, if at least one of the one or more components require real-time processing, assigning the one or more components that require real-time processing to one or more real-time processes, and assigning any remaining components to one or more non-real-time processes.

Description

    REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 12/891,681 filed on Sep. 27, 2010 and entitled, “METHOD FOR PROCESSING MULTIMEDIA STREAMS” which is a continuation of U.S. patent application Ser. No. 11/023,258 filed on Dec. 22, 2004 and entitled, “METHOD FOR PROCESSING MULTIMEDIA STREAMS”. These applications are entirely incorporated by reference.
  • FIELD OF THE INVENTION
  • The invention is related to processing multimedia streams. More specifically, the invention is related to the intelligent selection of algorithmic processes to process multimedia streams.
  • BACKGROUND OF THE INVENTION
  • Interaction with multimedia streams such as during music downloading, video recording, and phone conversations is pervasive in society. The types of electronic devices designed to transmit and receive these streams are as varied as the formats of the streams themselves. If a user of one electronic device sends information to a user of another electronic device the formats need to be compatible for the communication to effectively take place. In many cases, the format that two electronic devices use are not compatible at all, thus an intermediary device must be used to translate the information traveling between the two devices. Devices called transcoders translate a multimedia stream in one format to a stream in another format so people can communicate with each other easily and still use these varied devices and varied multimedia stream formats. Transcoding currently is utilized in a variety of different scenarios such as for one-way, two-way, or multi-way streams.
  • One-way transcoding is used in circumstances where one device is sending a multimedia stream to a second device. Video messaging and video streaming are examples of one-way transcoding. FIG. 1A describes an example of a transaction that may require one-way transcoding of a multimedia stream. In this example, a video stream 100 is sent from a first computer 102 to a second computer 104. The video from the first computer 102 is encoded into one of many different possible formats including MPEG-2, MPEG-4, GSM (global system for mobile communications), or AMR (adaptive multi-rate) among others. The second computer 104 receives, decodes, and displays the streaming video on a video monitor.
  • Another example of one-way transcoding would be voicemail. A person may want to record a message on their voicemail system to inform callers that they are out of the office. That person may want to play back the message to determine whether they want to save it or record again. The person could be speaking through a G.711 format phone and the voicemail system may be recording the message in a WAV file. Again, this requires a transcoding step.
  • Two-way transcoding is used when two devices are sending streams to each other, such as potentially during a video call or an audio (e.g., telephone) call. FIG. 1B describes an example of a two-way video call that may require two-way transcoding of multimedia streams. In this example, a video stream 110 is sent from a first computer 112 to a second computer 114, and also from the second computer 114 to the first computer 110.
  • Multi-way transcoding is used when more than two devices are sending streams to each other, such as, for example, during a three-party video conference. FIG. 1C describes an example of a three-way video conference that may require multi-way transcoding. In this example, just like the example in FIG. 1B, multiple computers are sending and receiving video streams between each other. In this example, a first computer 122, a second computer 124, and a hand-held mobile computer/cell phone 126 are sending and receiving streams 120 between each other.
  • In many cases, each of the computers shown in FIGS. 1A, 1B and 1C may encode and decode multimedia streams in different formats. For example, the two computers in FIG. 1C (130 and 132) may encode and decode an MPEG-4 format stream, but the mobile computer/cellular phone 134 may not be capable of encoding or decoding MPEG-4 and instead encodes and decodes only a GSM-format stream. In this case, a transcoding device may be utilized to intercept the streams coming from each of the three devices and translate the streams into each destination device's native format.
  • A video stream is just one of many possible examples of where a transcoder may be needed. For example, if a person using a land-line phone calls his friend on a cellular phone there is a transcoding step necessary. The standard land-line phones typically operate using a G.711 format. The friend who answers the call would typically have a GSM, AMR, or other cellular phone voice format technology. Thus, in order for the two friends to hear each other a transcoding step would be necessary while they were talking.
  • The processing power required for many of these transcoding examples varies greatly. In some examples, real-time transcoding is required, in others it isn't. “Real-time” refers to the proximate speed required for a transcoder to transcode and output a multimedia stream at the same rate the untranscoded multimedia stream arrives at the transcoder. Thus, transcoding a voice stream in real-time would require transcoding the voice information at the rate the person is actually speaking the words. Therefore, the two friends who are talking to each other over a land-line and a cellular line must have the formats transcoded in real-time because they are speaking to each other in real-time and a delay in the audio may not be acceptable. The 3-way video conference would require even more processing power because not only are there three streams that need to be transcoded in real-time, but the streams consist of packets with voice and video components, which is a lot more information per packet than just pure voice component packet.
  • Currently, real-time transcoding systems usually utilize extremely powerful processors, such as servers with dedicated digital signal processors, to make sure that any form of information in the packet can be transcoded in real-time. In many occasions, this amount of processing power is not required. A person who is recording an out-of-office message and then reviews their recordings is not operating in real-time. There is a delay in the range of multiple seconds between when the person speaks into the phone, and when they hear their recorded voice during review.
  • There are also transcoding systems that perform transcoding in a non-real-time, offline manner. A user on a personal computer may translate a VHS tape into an AVI file to store on a DVD. Finally, there are implementations where no transcoding is done at all, for example, raw video, voice, or audio format streams can be transmitted across a network. This type of stream is not efficient because raw format streams are uncompressed and require a great deal of network bandwidth.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The present invention is illustrated by way of example and is not limited by the figures of the accompanying drawings, in which like references indicate similar elements, and in which:
  • FIG. 1A describes an example of a transaction that may require one-way transcoding of a multimedia stream.
  • FIG. 1B describes an example of a two-way video call that may require two-way transcoding of multimedia streams.
  • FIG. 1C describes an example of a three-way video conference that may require multi-way transcoding.
  • FIG. 2 is a block diagram of one embodiment of a computer system with a multimedia stream processing management system (MSPMS).
  • FIG. 3 describes one embodiment of a multimedia stream processing management system (MSPMS).
  • FIG. 4 describes one embodiment of a method to transcode a multimedia stream packet with a single algorithmic process.
  • FIG. 5 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • FIG. 6 describes another embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes.
  • FIG. 7 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and one buffer.
  • FIG. 8 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and multiple buffers.
  • FIG. 9 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes and multiple buffers.
  • FIG. 10 describes one embodiment of an algorithmic process omitting individual components based on component priority and processing power limitations.
  • FIG. 11 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream to real-time and non-real-time algorithmic processes.
  • FIG. 12 is a flow diagram of another embodiment of a method for assigning components of a multimedia stream to real-time processes.
  • FIG. 13 is a flow diagram of an embodiment of a method for omitting components of a multimedia stream from the transcoding process.
  • FIG. 14 is a flow diagram of an embodiment of a method to skip the transcoding of a stream.
  • FIG. 15 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream that do not require real-time processing, but require processing before a determined maximum time delay, to processes that are able to process within the determined maximum time delay.
  • DETAILED DESCRIPTION
  • Embodiments of an effective method for processing multimedia streams are disclosed. In the following description, numerous specific details are set forth. However, it is understood that embodiments may be practiced without these specific details. In other instances, well-known elements, applications, and techniques have not been discussed in detail in order to avoid obscuring the present invention.
  • FIG. 2 is a block diagram of one embodiment of a computer system with a multimedia stream processing management system (MSPMS). The computer system includes a processor 200, a memory controller hub (MCH) 202, and an I/O controller hub (ICH) 210. The MCH 202 and the ICH 210 comprise a chipset. The processor 200 is coupled to the MCH 202 via a host bus. The MCH 202 is coupled to system memory 204. In different embodiments, system memory may be synchronous dynamic random access memory (SDRAM), double data rate SDRAM (DDR-SDRAM), Rambus DRAM (RDRAM), or one of many other formats of main system memory.
  • In one embodiment, the computer system includes a MSPMS 206 stored in system memory 204 and operated/executed by the processor 200. In different embodiments, the MSPMS 206 may control a number of management decisions relating to processing a multimedia stream such as the speed the processing requires (e.g., real-time or non-real-time), the priority associated with each of the different components that comprise the stream (e.g., the audio component, the voice component, the video component, the data component, and the control component), and the assignment of the specific device that will actually handle the multimedia stream. In different embodiments, the device that actually handles the multimedia stream may be a digital signal processor, a network server, a remote desktop computer, or the computer system shown in FIG. 2 (including processor 200) among many other possible processing devices.
  • The MCH 202 is also coupled to a graphics module 208. The ICH 210 is coupled to a hard drive 212 and an I/O bus 214. The I/O bus 214 is coupled to a network interface card that interfaces a network. In different embodiments, the network may be a local area network, a public phone system, the Internet, a wide area network, a wireless network, or any other network that is capable of delivering data between computing devices.
  • FIG. 3 describes one embodiment of a multimedia stream processing management system (MSPMS). In this embodiment, the processing logic 300 of the MSPMS is located between a multimedia stream source 302 and a destination 318. The source 302 generates, encodes, and sends a multimedia stream across a network to a destination device 318. The stream is comprised of a number of sequential packets.
  • Each packet may comprise several components. Each component of the packet consists of a type of information regarding the stream. For example, a packet for an audio/video stream can contain a voice component and a separate video component. In different examples, there may be other components that also make up multimedia streams such as a data component, which sends instructions within the stream, a general audio component, which includes music and background noise information (as opposed to the specific voice information), and a control component, which may include security information, authentication information, and quality of service (QOS) information among other types of control information. Each component within a packet contains a predetermined amount of information for that component. The amount of information per component per packet is usually measured in time. For example, a voice component within an individual packet of a multimedia stream may consist of 100 milliseconds of the recorded voice. Apart from component information, each packet includes a header that contains address information to allow the packet to be routed to the correct destination.
  • Returning to FIG. 3, if, for example, the source 302 is sending an MPEG-4 format video stream to the destination 318, which requires an MPEG-2 format, then the stream must be transcoded prior to reaching the destination 318. In this embodiment, the MSPMS processing logic 300 intercepts the stream. The stream enters a stream unwrapper unit 304. The stream unwrapper unit 304 unwraps the stream into its individual components. The individual components of the stream are routed to the process assigning unit 306. In the current example, the video component is unwrapped and sent to the process assigning unit 306. In one embodiment, header address information is sent directly from the stream unwrapper unit 304 to the stream wrapper unit 316.
  • The process assigning unit 306 then assigns each of the one or more stream components to processing logic that includes one or more algorithmic processes (308, 310, 312, 314). An “algorithmic process” is a method or process that is capable of modifying one or more components of a multimedia stream in some way. Examples of algorithmic processes include a method to transcode a component of a multimedia stream, a method to omit certain components of a multimedia stream, a method to delay the processing of a component of a multimedia stream, a method to buffer a real-time transport protocol (RTP) stream, among others.
  • Many times, to complete a certain task, more than one algorithmic process must be utilized. For example, a user may want to leave a voicemail. The user calls a phone number and a voicemail system answers the call. When the user leaves the message a first algorithmic process may be buffering the actual voice data coming through an RTP stream, a second algorithmic process may be monitoring for a dial tone (dual-tone multi-frequency (DTMF) detection) to detect when any buttons on the phone are pressed during the user message, and a third algorithmic process may be transcoding the RTP stream into a more storage friendly format. In this case, the RTP stream buffer and the DTMF detection algorithmic processes must be operated in real-time to create an effective user interaction, otherwise the voicemail will not be useful, whereas the transcoding process just takes information from the buffer on an available basis. Thus, the RTP and DTMF algorithmic processes may have a higher priority than the transcoding process and the processes are managed by the MSPMS accordingly.
  • In one embodiment, prioritizing algorithmic processes may be accomplished based on the importance of a component regarding user perception. In the above example, the voicemail user interaction would place a higher priority on functionality to maintain a clear recording of the user's voice. Therefore, buffering an RTP stream is obviously more important than transcoding that stream to a different format because omitting any part of the RTP stream would cause a loss of information and the user's voice would not be audible in part or all of the message. Accordingly, in one embodiment, there are N algorithmic processes, each of which may accomplish the processing of a stream component in a different way. In one embodiment, each algorithmic process prioritizes the processing of components sent to it based on each component's importance. Thus, in this embodiment, if the process assigning unit 306 assigns two or more components to a single algorithmic prioritizing process, the two or more components will be processed in the order of their priority. In different embodiments, the priority may be decided based on an algorithm that takes into account factors such as network transmission rates, relative importance of different components for user comprehensibility of the stream, the bandwidth consumption of the component in relationship to the other components, or any number of other factors.
  • For example, in a video conference, if the video component is not available, audio communication is still a viable alternative, but if the audio component is not available, the efficiency of the communication drops dramatically with just a video feed. In another example, the video component in a multimedia stream of multi-component packets usually consumes the most bandwidth and computing power, thus, the algorithmic process may be able to just transcode video in real-time, or it may be able to transcode all of the other components in real-time (possibly audio, voice, data, and control components), therefore, four components may have priority over just one. Additionally, if resources are not available to process all components in real-time or even in a delayed (non-real-time) manner, another option is to omit a certain percentage of the lower priority component(s). Thus, if the video component is the lowest priority, the decision may be made by the process assigning unit 306 to assign a component to a particular algorithmic process that only processes every other (or every third, fourth, etc.) packetized component, which would result in a lower frame rate.
  • In another embodiment, two or more of the N algorithmic processes are running on a single computer system. In this embodiment, each of the two or more components also have their own assigned priority levels. Thus, in one example, if a multimedia stream has a video component and a voice component, each assigned to a separate algorithmic process by the process assigning unit 306, the voice component process may have a higher operating priority in the computer system than the video component process. In this case, if the computer system does not have enough resources to run both processes simultaneously, the video component process will be delayed while the voice component process completes the processing of the voice component of one or more multimedia stream packets. In another embodiment, each of the N algorithmic processes runs on its own thread in the local computer system. In this embodiment, the local computer system may assign each thread a certain priority to correspond with the priorities of the components assigned to each process-thread.
  • In yet another embodiment, two or more of the N algorithmic processes run on two or more separate computer systems. In this embodiment, the process assigning unit 306 may utilize one or more of the algorithmic processes that run on the computer system with the most available resources. In still yet another embodiment, one component may be assigned to two or more algorithmic processes. In this embodiment, the process assigning unit 306 divides up the information within a single component in the stream into portions and assigns each portion to a separate algorithmic process. In one embodiment, the process assigning unit 306 alternates sending component information between two algorithmic processes on a packet-by-packet level (i.e., a first packet's component is sent to algorithmic process one, a second packet's component is sent to algorithmic process two, a third packet's component is sent to algorithmic process three, etc).
  • In the current example, the process assigning unit 306 assigns the video component to an algorithmic process to transcode from MPEG-4 to MPEG-2. Once the video component for each packet is processed by the one or more algorithmic processes, the video component (i.e., the new MPEG-2 format video stream) is sent to a stream wrapper unit 316. In one embodiment, the stream wrapper unit 316 will encapsulate the individual components back into a packetized format to be sent to the destination. In one embodiment, the stream wrapper unit 316 will take information directly from the stream unwrapper unit 304 such as the destination address of each component the stream wrapper unit 316 receives as the output of the one or more algorithmic processes. In another embodiment, the stream wrapper unit 316 receives information directly from the process assigning unit 306 such as priority information per component and packet combining information per component. Priority information per component allows the stream wrapper unit 316 to prioritize the wrapping and sending of different components that arrive from one or more of the algorithmic processes. Packet combining information per component allows the stream wrapper unit 316 to incorporate multiple components into a single packet or, alternatively, allow an individual component to be wrapped into a separate packet and sent for a reason such as higher priority.
  • In the current example, the stream wrapper unit 316 receives the newly transcoded MPEG-2 video component and rewraps the component with the destination address information and any other pertinent information and sends the stream on to the destination 318.
  • In one embodiment, when deciding which algorithmic process to sign a given component to, the process assigning unit 306 takes into account information including: the network connectivity and throughput of each computer system that an algorithmic process runs on, the amount of free resources that each computer system has available, the priority of each algorithmic process on its respective computer system, among other information. In this embodiment, the process assigning unit 306 receives information regarding each computer system from the algorithmic processes running on each one (the dotted arrows in FIG. 3). With computer system information relating to each algorithmic process, the process assigning unit 306 (and the MSPMS in general) is capable of determining how to process each component in the stream as efficiently as possible to minimize or eliminate any perceived performance degradation of the stream. Thus, as illustrated above, the MSPMS may process a component of a stream in real-time, not in real-time, or not process a component at all depending on the priority of the component and the available system resources.
  • FIGS. 4 through 9 illustrate multiple examples of MSPMS processing logic decisions regarding the processing of a multimedia stream with a video and a voice component. In these examples the functionality shows that each component in the stream requires transcoding, although, in different embodiments, partial transcoding or no transcoding at all may be allowable for some or all of the video stream. Additionally, processing the components of a stream may include many other functions other than just transcoding such as buffering an RTP stream, DTMF detection, or any one of many other multimedia stream processing functions. Thus, the transcoding examples in FIGS. 4 through 9 are just illustrative examples, which may be substituted for any number of other processes that take place involving a multimedia stream. Furthermore, although the examples are illustrative of a single packet within the stream, the embodiments include repeating the process for potentially all the packets within the stream.
  • FIG. 4 describes one embodiment of a method to transcode a multimedia stream packet with a single algorithmic process. In this embodiment, a source 500 sends a multimedia stream packet 502 across a network. The packet 502 contains a header 504 that has all the routing and address information regarding the packet's destination. In this embodiment, the packet also contains a video component 506 and a voice component 508. In one embodiment, the video component comprises one video frame (i.e., a still image) and the voice component comprises one voice frame (i.e., 100 milliseconds of voice information). In one embodiment, the header 404 is removed and a packet 410 containing only the video and voice component is sent to the algorithmic process 412. The algorithmic process 412 transcodes the information in each component and outputs a new packet containing only transcoded video and voice components. The header 404 bypasses 414 the algorithmic process and may be reattached to the transcoded packet to create a new packet 416, which is sent to the destination 418. In this embodiment, the single algorithmic process 412 is capable of transcoding (in real-time) a multimedia stream of packets that have video and voice components. In one embodiment, the header 404 is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header 404 is not removed from the packet and the packet is sent to the algorithmic process with the header 404 intact.
  • FIG. 5 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes. In this embodiment, a source 500 sends a multimedia stream packet 502 across a network. The packet 502 contains a video component, and a voice component (as described above in FIG. 4). In one embodiment, the header is removed and different components within the packet are also separated. The video component 504 is sent to algorithmic process #1 (508) for video transcoding and the voice component 506 is sent to algorithmic process #2 (510) for voice transcoding. In different embodiments, algorithmic processes #1 (508) and #2 (510) may be located on the same computing device or on different computing devices. When algorithmic processes #1 (508) and #2 (510) complete their respective video and voice component transcoding, the header, which bypassed the transcoding process (512 and 514) may be reattached to separate transcoded video and voice component packets (516 and 518). The video component packet 516 and voice component packet 518 arrive separately at the destination 520. In one embodiment, the header is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 6 describes another embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes. In this embodiment, a source 600 sends a multimedia stream packet 602 across a network. The packet 602 contains a video component and a voice component (as described above in FIG. 4). The header is removed and different components within the packet are also separated. The video component 604 is sent to algorithmic process #1 (608) for video transcoding and the voice component 606 is sent to algorithmic process #2 (610) for voice transcoding. When algorithmic processes #1 (608) and #2 (610) complete their respective video and voice component transcoding they reattach the video component and the voice component into one multi-component packet 614. The header, which bypassed the transcoding process may be reattached to the multi-component packet 614 and the packet is sent to the destination 616. In this embodiment, if either transcoded component packet must wait more than a pre-determined amount of time for a corresponding packet from the other algorithmic process, the packet 614 is sent to the destination by the stream wrapper unit (FIG. 3, 316) with just one component to minimize any bottleneck in the algorithmic process that finished first. In one embodiment, the header is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 7 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and one buffer. In this embodiment, a source 700 sends a multimedia stream packet 702 across a network. The packet 702 contains a video component and a voice component (as described above in FIG. 4). The header is removed and different components within the packet are also separated. In this embodiment, the voice component 704 has priority over the video component 706, thus the voice component 704 is sent directly to the algorithmic process 710 while the video component 706 is sent to a buffer 708 to wait for the algorithmic process 710 to finish transcoding the voice component 704. Once the voice component 704 transcoding process has finished, the buffer 708 releases the stored video component 706 to the algorithmic process 710 for transcoding. Once a voice component or video component has finished being transcoded, the transcoded component is reattached to the header (712 or 714) and the individual video or voice component packet (716 or 718) is sent to the destination 720. In one embodiment, the header is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 8 describes an embodiment of a method to transcode a multimedia stream packet with one algorithmic process and multiple buffers. In this embodiment, a source 800 sends a multimedia stream packet 802 across a network. The packet 802 contains a video component and a voice component (as described above in FIG. 4). The header is removed and different components within the packet are also separated. In this embodiment, there is only one algorithmic process 812. Additionally, the priority of the components may be changed in real-time based on certain circumstances. In this case, the priority of the video component and the priority of the voice component may increase or decrease relative to each other during transcoding operations. Thus, two buffers are utilized to allow for either component to have the capability of waiting for the algorithmic process 812 to finish transcoding the other component.
  • In this embodiment, the video component 804 is sent to buffer #1 (808) and the voice component 806 is sent to buffer #2 (810). In this embodiment, the algorithmic process 812 takes a single component one at a time from either buffer #1 (808) or buffer #2 (810). The determination of which buffer to take from depends on the priority assigned to the individual component waiting in each buffer. Once the algorithmic process 812 finishes transcoding an individual component, the transcoded component is reattached to the header (814 or 816) and the individual video or voice component packet (818 or 820) is sent to the destination 822.
  • In one embodiment, the header is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 9 describes an embodiment of a method to transcode a multimedia stream packet with multiple algorithmic processes and multiple buffers. In this embodiment, a source 900 sends a multimedia stream packet 902 across a network. The packet 902 contains a video component and a voice component (as described above in FIG. 5). The header is removed and different components within the packet are also separated. In this embodiment, there are two algorithmic processes (912 and 914). Additionally, the priority of the components may be changed in real-time based on certain circumstances. In one embodiment, algorithmic processes #1 (912) and #2 (914) are located on the same computer system, but in separate processes. In this case, the priority of the video transcoding process and the priority of the voice transcoding process may increase or decrease relative not only to each other, but also to other processes running on the computer system.
  • For example, the computer system that both algorithmic processes reside on receives a worm from the Internet and a process that runs a virus protection software program temporarily takes a higher priority than either the video transcoding algorithmic process or the voice transcoding algorithmic process. Thus, in this case, each algorithmic process has its own buffer to allow for both components to have the capability of waiting for the algorithmic processes (912 and 914) to regain the high priority on the system.
  • In this embodiment, the video component 904 is sent to buffer #1 (908) and the voice component 906 is sent to buffer #2 (910). In this embodiment, algorithmic process #1 (912) takes individual video components from buffer #1 (908) and algorithmic process #2 (914) takes individual voice components from buffer #2 (910). Once either algorithmic process finishes transcoding an individual component, the transcoded component is reattached to its header (916 or 918) and the individual video or voice component packet (920 or 922) is sent to the destination 924.
  • In one embodiment, the header is modified by the stream wrapper unit (FIG. 3, 316) based on information provided from the stream unwrapper unit (FIG. 3, 304) and the process assigning unit (FIG. 3, 306). In another embodiment, the header is not removed from the packet and the packet is sent to the algorithmic process with the header intact. Thus, the algorithmic process itself may modify the header.
  • FIG. 10 describes one embodiment of an algorithmic process omitting individual components based on component priority and processing power limitations. In this embodiment, a stream of video components 1000 (i.e., video frames) and a stream of voice components 1002 (i.e. voice frames) is input into a algorithmic process 1004. Thus, the algorithmic process 1004 in this embodiment is similar to the algorithmic process described in FIG. 8. In this embodiment, the algorithmic process 1004 does not have the processing power to keep up with every voice frame and every video frame. Thus, if there are no other systems with algorithmic processes capable of taking some of the transcoding work, certain frames must be omitted to maintain the real-time nature of the stream.
  • In the case that the voice component has a higher transcoding priority than the video component, certain frames of video must be omitted (i.e., dropped). Therefore, in one embodiment, a stream of four video frames 1000 and a stream of four voice frames 1002 are input into the algorithmic process 1004. The algorithmic process 1004 transcodes all voice frames and sends them out in order in the voice output stream 1006. The algorithmic process 1004 also uses its remaining processing power to transcode as many video frames as possible and sends them out in order in the video output stream 1008. In this case, the algorithmic process 1004 is omitting every other video frame to maintain the real-time aspect of the streams. Consequently, the voice stream is clear and intact, and the video stream has its frame rate cut in half.
  • FIG. 11 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream to real-time and non-real-time algorithmic processes. The method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 11, the method begins by processing logic determining whether one or more components comprising a multimedia stream require real-time processing (block 1100). If processing logic determines that real-time processing is required (block 1102) then processing logic assigns the one or more components that require real-time processing to one or more real-time algorithmic processes (block 1104). Finally, processing logic assigns any remaining components to one or more non-real-time algorithmic processes (block 1106) and the method is finished.
  • FIG. 12 is a flow diagram of another embodiment of a method for assigning components of a multimedia stream to real-time processes. The method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 12, the method begins by processing logic unwrapping a packet into its one or more components (block 1200). Next, processing logic determines the amount of processing required to process each of the one or more components requiring real-time processing within each packet in real-time (block 1202). Finally, processing logic assigns each of the one or more components requiring real-time processing within the packet to a separate real-time process (block 1204) and the method is finished.
  • FIG. 13 is a flow diagram of an embodiment of a method for omitting components of a multimedia stream from the transcoding process. The method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 13, the method begins by processing logic unwraps a packet into its one or more components (block 1300). Next, processing logic determines if all components of the unwrapped packet can be transcoded in real-time (block 1302). If they can then the method is finished. Otherwise, processing logic examines the first component (block 1304). Once the component is examined, processing logic determines if the component can be omitted (block 1306).
  • The determination whether a component can be omitted may be made based on a number of factors such as the processing priority of the component and the number of consecutive component frames that have been omitted. For example, as discussed in FIG. 10, if the algorithmic process does not have the processing power to process everything in real-time a decision must be made as to which components to omit and when to omit them. Dropping a video component down to half its standard frame rate by omitting every other frame is less objectionable than to drop 15 frames in a row and then transcode 15 frames in a row. The end result is the same number of frames, but the quality of the viewing experience is different for those two separate options.
  • Returning to FIG. 13, if the component can be omitted then processing logic omits the component (block 1308). Next, processing logic determines whether all components have been examined (block 1310). If so, the method is finished. Otherwise, processing logic examines the next component (block 1312). Eventually, all components of an unwrapped packet will be examined and the method will finish. In another embodiment, determining whether to omit one or more components from a stream may include a decision regarding multimedia stream processes other than just transcoding. Thus, transcoding is used as an example and can be substituted with one or more other functions that may be performed on a multimedia stream.
  • FIG. 14 is a flow diagram of an embodiment of a method to skip the transcoding of a stream. The method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 14, the method begins by processing logic determining the format of the stream transmitted from the source device (block 1500). Next, processing logic determines the format of the stream required by the destination device (block 1502). Next, processing logic determines whether the two formats are the same (block 1504). If the formats are not the same, processing logic transcodes the stream (block 1506) and the process is finished. Otherwise, if the formats are the same, processing logic does not transcode the stream and instead sends the stream directly from the source to the destination (block 1508) and the method is finished.
  • FIG. 15 is a flow diagram of an embodiment of a method for assigning components of a multimedia stream that do not require real-time processing, but require processing before a determined maximum time delay, to processes that are able to process within the determined maximum time delay. The method is performed by processing logic that may comprise hardware (circuitry, dedicated logic, etc.), software (such as is run on a general purpose computer system or a dedicated machine), or a combination of both. Referring to FIG. 15, the method begins by processing logic determining the maximum delay allowable for each of the components in a packet that do not require real-time processing (block 1500). In different embodiments, a determination of the maximum allowable processing delay may be made based on a number of factors including the type of transaction the multimedia stream is being used for, the speed of the transmission medium (e.g. the required speed of processing a voicemail attached in an email is limited to the inherent delay of emails), as well as numerous other factors. Next, processing logic assigns each of the components not requiring real-time processing to a separate process, wherein each process delivers at least the amount of processing required to process each respective component in the allowable delayed time (block 1502) and the method is finished.
  • In many of the preceding example embodiments, the packets dealt with mainly video and voice components. In other embodiments, the packets contain any one or more of video, audio, voice, data, control, or any other conceivable packet type. Additionally, the example embodiments referenced packets that included only two components. In other embodiments, the packets may contain any conceivable number of components (i.e., greater than or equal to one component). Finally, in many example embodiments a single algorithmic process was responsible for processing all components of a single type. In other embodiments, more than one algorithmic process may be assigned to a single type of component. Thus, in many embodiments, two or more algorithmic processes can trade off processing duties and each algorithmic process can therefore process every other packetized component of a single type (or every third, fourth, etc.).
  • Thus, embodiments for an effective method for processing multimedia streams are disclosed. These embodiments have been described with reference to specific exemplary embodiments thereof. It will, however, be evident to persons having the benefit of this disclosure that various modifications and changes may be made to these embodiments without departing from the broader spirit and scope of the embodiments described herein. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (14)

1. A device, comprising logic to:
read a first packet, the first packet one of a plurality of packets comprising a multimedia packet stream, wherein the first packet comprises a plurality of components; and
for each component of the plurality of components within the packet
determine whether the component requires real-time-processing;
assign the component to be processed by a real-time process in response to the component requiring real-time processing; and
assign the component to be processed by a non-real-time process in response to the component not requiring real-time processing.
2. The device of claim 1, wherein the logic is further operable to
determine a rate at which the plurality of packets are being received; and
process each component within each packet at least at the rate the packets are being received.
3. The device of claim 2, wherein the logic is further operable to
unwrap the first packet by separating each of the plurality of components comprising the first packet.
4. The device of claim 3, wherein the logic is further operable to
assign a priority level to a component, wherein the priority level is calculated based on the importance of the component's information.
5. The device of claim 4, wherein the logic is further operable to
determine a maximum delay allowable for a first component not requiring real-time processing; and
assign the first component to a process that delivers at least the amount of processing required to process the first component within the maximum allowable delay.
6. The device of claim 4, wherein the logic is further operable to
process a plurality of the components of the first packet sequentially, wherein the sequence of the components to be processed is in the order of the priority levels per component, a highest priority level component being processed first and a lowest priority level component being processed last.
7. The device of claim 6, wherein the logic is further operable to
store the stream of packets in real-time on a storage device immediately after packet creation to prevent information loss of one or more low priority level packet components.
8. The device of claim 4, wherein the logic is further operable to:
omit at least one component of the first packet from being processed.
9. The device of claim 4, wherein the logic is further operable to
omit a component with the lowest priority level from the first packet.
10. The device of claim 9, wherein the logic is further operable to
set a maximum number of consecutive packets in the stream with the same omitted component.
11. The device of claim 1, wherein the logic is further operable to:
determine a first format of the packets in the stream transmitted from a source device;
determine a second format of the packets in the stream required by a destination device;
cause each packet within the stream to directly transmit between the source device and the destination device without processing in response to the first and second formats being equal.
12. The device of claim 1, wherein a process comprises an executable thread on a computing device.
13. The device of claim 1, wherein the logic is further operable to
assign at least two of the plurality of components within each packet to processes on separate computing devices.
14. The device of claim 1, wherein the logic is further operable to
assign two or more of the plurality of components to the same process.
US15/278,952 2004-12-22 2016-09-28 Method for processing multimedia streams Abandoned US20170085614A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/278,952 US20170085614A1 (en) 2004-12-22 2016-09-28 Method for processing multimedia streams

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US11/023,258 US20060133513A1 (en) 2004-12-22 2004-12-22 Method for processing multimedia streams
US12/891,681 US9462333B2 (en) 2010-09-27 2010-09-27 Method for processing multimedia streams
US15/278,952 US20170085614A1 (en) 2004-12-22 2016-09-28 Method for processing multimedia streams

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/891,681 Continuation US9462333B2 (en) 2004-12-22 2010-09-27 Method for processing multimedia streams

Publications (1)

Publication Number Publication Date
US20170085614A1 true US20170085614A1 (en) 2017-03-23

Family

ID=45870606

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/891,681 Active 2033-01-31 US9462333B2 (en) 2004-12-22 2010-09-27 Method for processing multimedia streams
US15/278,952 Abandoned US20170085614A1 (en) 2004-12-22 2016-09-28 Method for processing multimedia streams

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/891,681 Active 2033-01-31 US9462333B2 (en) 2004-12-22 2010-09-27 Method for processing multimedia streams

Country Status (1)

Country Link
US (2) US9462333B2 (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9813562B1 (en) * 2010-09-01 2017-11-07 Sprint Communications Company L.P. Dual tone multi-frequency transcoding server for use by multiple session border controllers
US8995306B2 (en) * 2011-04-06 2015-03-31 Cisco Technology, Inc. Video conferencing with multipoint conferencing units and multimedia transformation units
WO2013163223A1 (en) * 2012-04-23 2013-10-31 Marvell World Trade Ltd. ADJUSTING THREAD PRIORITY TO IMPROVE THROUGHPUT BETWEEN Peer-to-peer (P2P) DEVICES
US10341673B2 (en) * 2013-05-08 2019-07-02 Integrated Device Technology, Inc. Apparatuses, methods, and content distribution system for transcoding bitstreams using first and second transcoders
US9836286B2 (en) * 2015-03-27 2017-12-05 Ca, Inc. Co-existential wrapping system for mobile applications

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020007416A1 (en) * 1998-06-23 2002-01-17 David M. Putzolu Recognizing audio and video streams over ppp links in the absence of an announcement protocol
US20020013633A1 (en) * 2000-07-28 2002-01-31 Tomoya Kodama Audio processor and audio data processing method
US20020181506A1 (en) * 2001-06-04 2002-12-05 Koninklijke Philips Electronics N.V. Scheme for supporting real-time packetization and retransmission in rate-based streaming applications
US20020196939A1 (en) * 2001-06-06 2002-12-26 Unger Robert Allan Decoding and decryption of partially encrypted information
US20040170198A1 (en) * 1999-07-15 2004-09-02 Telefonaktiebolaget Lm Ericsson (Publ) Scheduling and admission control of packet data traffic
US20060133513A1 (en) * 2004-12-22 2006-06-22 Kounnas Michael K Method for processing multimedia streams
US7385997B2 (en) * 2002-04-08 2008-06-10 International Business Machines Corporation Priority based bandwidth allocation within real-time and non-real-time traffic streams
US7647619B2 (en) * 2000-04-26 2010-01-12 Sony Corporation Scalable filtering table
US20120078402A1 (en) * 2009-06-19 2012-03-29 Dolby Laboratories Licensing Corporation Upgradeable engine framework for audio and video
US8213413B2 (en) * 2004-11-19 2012-07-03 Northrop Grumman Systems Corporation Real-time packet processing system and method

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2809962B2 (en) 1993-03-02 1998-10-15 株式会社東芝 Resource management method
JP3249334B2 (en) * 1995-04-06 2002-01-21 株式会社東芝 Digital interface device and digital interface method
US5703877A (en) * 1995-11-22 1997-12-30 General Instrument Corporation Of Delaware Acquisition and error recovery of audio data carried in a packetized data stream
JPH10240548A (en) 1997-03-03 1998-09-11 Toshiba Corp Task scheduling device and method
US6052384A (en) 1997-03-21 2000-04-18 Scientific-Atlanta, Inc. Using a receiver model to multiplex variable-rate bit streams having timing constraints
US6351471B1 (en) 1998-01-14 2002-02-26 Skystream Networks Inc. Brandwidth optimization of video program bearing transport streams
JP2955561B1 (en) * 1998-05-29 1999-10-04 株式会社ディジタル・ビジョン・ラボラトリーズ Stream communication system and stream transfer control method
US6483543B1 (en) 1998-07-27 2002-11-19 Cisco Technology, Inc. System and method for transcoding multiple channels of compressed video streams using a self-contained data unit
US7088725B1 (en) 1999-06-30 2006-08-08 Sony Corporation Method and apparatus for transcoding, and medium
US6438135B1 (en) * 1999-10-21 2002-08-20 Advanced Micro Devices, Inc. Dynamic weighted round robin queuing
ES2269375T3 (en) 2000-04-11 2007-04-01 Koninklijke Philips Electronics N.V. DATA FLOW ADAPTATION SERVER.
US6747991B1 (en) 2000-04-26 2004-06-08 Carnegie Mellon University Filter and method for adaptively modifying the bit rate of synchronized video and audio streams to meet packet-switched network bandwidth constraints
US6847656B1 (en) 2000-09-25 2005-01-25 General Instrument Corporation Statistical remultiplexing with bandwidth allocation among different transcoding channels
US6793625B2 (en) 2000-11-13 2004-09-21 Draeger Medical Systems, Inc. Method and apparatus for concurrently displaying respective images representing real-time data and non real-time data
JP2002185943A (en) 2000-12-12 2002-06-28 Nec Corp Broadcasting viewing method, broadcasting transmission server, portable terminal and multi-spot speaking and broadcasting control viewing equipment
US7376731B2 (en) 2002-01-29 2008-05-20 Acme Packet, Inc. System and method for providing statistics gathering within a packet network
CN1647001A (en) 2002-04-05 2005-07-27 松下电器产业株式会社 Improved utilization of data broadcasting technology with handheld control apparatus
KR100462321B1 (en) 2002-12-16 2004-12-17 한국전자통신연구원 system for down-link packet scheduling of mobile connuvication and method thereof, its program stored recording medium
CN1933465B (en) 2005-09-13 2010-09-08 华为技术有限公司 Method and device for eliminating differential transmission delay difference

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020007416A1 (en) * 1998-06-23 2002-01-17 David M. Putzolu Recognizing audio and video streams over ppp links in the absence of an announcement protocol
US20040170198A1 (en) * 1999-07-15 2004-09-02 Telefonaktiebolaget Lm Ericsson (Publ) Scheduling and admission control of packet data traffic
US7647619B2 (en) * 2000-04-26 2010-01-12 Sony Corporation Scalable filtering table
US20020013633A1 (en) * 2000-07-28 2002-01-31 Tomoya Kodama Audio processor and audio data processing method
US20020181506A1 (en) * 2001-06-04 2002-12-05 Koninklijke Philips Electronics N.V. Scheme for supporting real-time packetization and retransmission in rate-based streaming applications
US20020196939A1 (en) * 2001-06-06 2002-12-26 Unger Robert Allan Decoding and decryption of partially encrypted information
US7385997B2 (en) * 2002-04-08 2008-06-10 International Business Machines Corporation Priority based bandwidth allocation within real-time and non-real-time traffic streams
US8213413B2 (en) * 2004-11-19 2012-07-03 Northrop Grumman Systems Corporation Real-time packet processing system and method
US20060133513A1 (en) * 2004-12-22 2006-06-22 Kounnas Michael K Method for processing multimedia streams
US20120078402A1 (en) * 2009-06-19 2012-03-29 Dolby Laboratories Licensing Corporation Upgradeable engine framework for audio and video

Also Published As

Publication number Publication date
US20120076155A1 (en) 2012-03-29
US9462333B2 (en) 2016-10-04

Similar Documents

Publication Publication Date Title
EP1851966A2 (en) Method for processing multimedia streams
US20170085614A1 (en) Method for processing multimedia streams
US7817557B2 (en) Method and system for buffering audio/video data
US10045089B2 (en) Selection of encoder and decoder for a video communications session
US8713167B1 (en) Distributive data capture
KR101008764B1 (en) Method and system for improving interactive media response systems using visual cues
KR101552365B1 (en) Multimedia communications method
US7324802B2 (en) Method and system for managing communication in emergency communication system
US20120030682A1 (en) Dynamic Priority Assessment of Multimedia for Allocation of Recording and Delivery Resources
US9021062B2 (en) Sharing audio and video device on a client endpoint device between local use and hosted virtual desktop use
US10679636B2 (en) Methods and apparatus for supporting encoding, decoding and/or transcoding of content streams in a communication system
US8787547B2 (en) Selective audio combination for a conference
US8984156B2 (en) Multi-party mesh conferencing with stream processing
EP4026345A1 (en) Combining video streams in composite video stream with metadata
US7865634B2 (en) Managing a buffer for media processing
US20100218210A1 (en) Emergency broadcast system
CN102246491A (en) System and method for storing multi-source multimedia presentations
EP4315819A1 (en) Method and system for integrating video content in a video conference session
CN113365108A (en) Audio and video transcoding system and method based on color ring back tone
US7113514B2 (en) Apparatus and method for implementing a packet based teleconference bridge
US12069121B1 (en) Adaptive video quality for large-scale video conferencing
US20070207778A1 (en) Messaging System And A Telephone Incorporating Such A System
US20100014509A1 (en) Digital telecommunications system, program product for, and method of managing such a system

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION