WO2010049879A1 - Method for buffering streaming data and a terminal device - Google Patents

Method for buffering streaming data and a terminal device Download PDF

Info

Publication number
WO2010049879A1
WO2010049879A1 PCT/IB2009/054731 IB2009054731W WO2010049879A1 WO 2010049879 A1 WO2010049879 A1 WO 2010049879A1 IB 2009054731 W IB2009054731 W IB 2009054731W WO 2010049879 A1 WO2010049879 A1 WO 2010049879A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
rendering
data rate
server
terminal
Prior art date
Application number
PCT/IB2009/054731
Other languages
French (fr)
Inventor
Nicolas Delahaye
Original Assignee
Nxp B.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nxp B.V. filed Critical Nxp B.V.
Priority to CN2009801427406A priority Critical patent/CN102197656A/en
Priority to US13/126,394 priority patent/US8612552B2/en
Priority to EP09747924.0A priority patent/EP2351371B1/en
Publication of WO2010049879A1 publication Critical patent/WO2010049879A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23406Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving management of server-side video buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/41407Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a portable device, e.g. video client on a mobile phone, PDA, laptop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44004Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving video buffer management, e.g. video decoder buffer or video display buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6131Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via a mobile phone network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/173Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
    • H04N7/17309Transmission or handling of upstream communications
    • H04N7/17318Direct or substantially direct transmission and handling of requests

Definitions

  • the invention relates to a method for buffering streaming data in audio/video applications and a terminal device.
  • Buffering is a common technique deployed in streaming use cases, but it may result in a poor user experience. Buffering is required in the following cases: (1) streaming session start-up (including channel zapping, from one to another one), and (2) jump/seek action initiated by the end user.
  • An example of a data processing system including a buffer for real-time date is known from US 6,247,072, which discloses apparatus and methods for matching data rates which is useful for a receiver receiving real-time data over a medium.
  • Implementations feature a process establishing a buffer in a receiver; receiving source data from a source having a nominal source data rate, the received source data arriving at an incoming data rate that differs from time-to-time from the nominal source data rate; filling the buffer with source data as it is received at the incoming data rate and emptying the buffer to provide data for consumption in real time at a consumption data rate; setting a rate-matching factor M, the factor M affecting the rate at which the buffer is emptied; and tracking the level of data in the buffer and resetting the value of M to increase the rate at which the buffer is emptied when the buffer fills above a target range, and resetting the value of M to decrease the rate at which the buffer is emptied when the buffer empties below a target range.
  • Analog media is sampled to be digitalized. This process uses different sample rates depending on audio or video.
  • audio is commonly sampled from 8kHz to 48kHz depending on the use case ( ⁇ 16kHz for voice content, > 16kHz for music content).
  • video may be sampled at 24Hz by cinema camera, 25Hz by PAL : European TV standard, about 30fps by NTSC : US TV standard.
  • both the server and the terminal send/consume the media at the same data rate; the server data rate and the rendering data rate are equal and match a real time clock.
  • Cable or terrestrial digital television have constant data throughput, for example the server data rate and the reception data rate are equal. Thus, sent data will be received and rendered by the terminal after a constant delay (transmission time).
  • Fig.l shows a schematic representation of a data processing system according to the prior art.
  • a server S is in charge of sending media to a terminal T via a network N.
  • the server S has real-time behaviour, e.g. it sends media data associated to the server clock.
  • the network N is in charge of carrying data from the server S to the terminal T.
  • a common way to model a network N is to use a buffer B.
  • This buffer B contains data RFS received from the server, and not yet sent to the terminal STT.
  • the network jitter corresponds to the network buffer duration BL.
  • the terminal T - also referred to as "client" or “receiver” - is in charge of receiving and rendering the media from the network N.
  • a terminal T has a media renderer and a buffer AL to manage reception rate variation from the network and consumption rate from the media renderer.
  • the main end-user insights to have a good end-user experience are: a large number of channels (not addressed by this invention), media rendering must not be interrupted in case of network congestion, quick start up time ( ⁇ 2s) (from channel selection to first image display) must be provided, fast zapping time ( ⁇ 2s) (from switch command to a new channel, and first image display of the new channel) must be provided and a quick jump/seek time ( ⁇ 2s) (from jump/seek command and first image display) must be provided.
  • the data throughput is not constant, for example the server data rate C s and the reception data rate C rec are not the same.
  • the server data rate C s can be higher than the reception data rate C rec . This triggers an increase of the network buffer B and a decrease of the terminal buffer AL decrease (as the consumption rate is constant, equal to the server output rate).
  • the terminal buffer AL has to support this variation to avoid media rendering interruption, e.g. terminal buffer time to be null.
  • the terminal has to buffer a significant amount of data. For instance, it pauses rendering (rendering data rate C ren is null), and waits for the terminal buffer time to increase. Once the terminal buffer time reaches a threshold, the media rendering starts (rendering data rate equals server data rate). Thus, during this buffering time the playback is paused, and the playback starts again once buffering has finished.
  • the buffering latency is equal to the end-user latency, e.g. the time elapsing between receiving the first media data and the first image display. This results in a poor end-user experience.
  • a common buffering time is about 8s.
  • a common buffering time is about 6s.
  • This object is solved by a method of buffering streaming data according to claim 1 and a terminal device according to claim 6.
  • a method of streaming data from a server at a server data rate via a network to at least one terminal at a terminal reception data rate is provided.
  • a streaming section from the server is requested by the terminal.
  • Streaming data is forwarded from the server to the network at a server data rate and from the network to the terminal at a reception data rate.
  • Data received from the network is buffered in the terminal buffer for at least a first period.
  • the rendering of the buffered data is initiated after the first period at a first rendering rate, which is lower than the server data rate or the reception data rate.
  • the first rendering data rate is adapted according to the filling of the terminal buffer with received streaming data until the rendering data rate corresponds to the server data rate.
  • the initial part of the rendering can be faster or slower than the server data rate.
  • the rendering data rate will only return to its normal value after the first period, i.e. after a certain amount of data rate is buffered in the terminal buffer. This will enable an improved rendering of the data as the rendering will be continuous. This is not possible according to the prior art as there the rendering will be stopped if not sufficient data is present in the terminal buffer.
  • the adapting of the first rendering data rate is performed by resampling the streaming data buffered in the terminal buffer.
  • an audio time stretching is performed to resample the streaming data.
  • the first rendering data rate is decreased with a decreasing filling of the terminal buffer.
  • the rendering data rate is increased with an increased filling of the terminal buffer.
  • the invention also relates to a terminal device for receiving streaming data from a server via a network.
  • the terminal device comprises at least one terminal buffer for buffering data received from the network at a reception data rate for at least a first period.
  • the terminal device furthermore comprises a rendering unit for rendering streaming data buffered in the terminal buffer.
  • the terminal device also comprises at least one control unit for requesting a streaming session from the server, for initiating a rendering of the buffered data after the first period at a first rendering rate by the rendering unit and for adapting the first rendering data rate according to the filling of the terminal buffer with received streaming data until the rendering data rate corresponds to the server data rate.
  • the first rendering data rate is lower than the server data rate when the rendering of the streaming data is initiated.
  • the invention also relates to a data processing system having a server, at least one terminal device and a network to which the server and the terminal device are coupled to.
  • the server sends stream data via the network to the at least one terminal device.
  • the terminal device is implemented as described above.
  • the invention proposes to reduce buffering latency (for example to about 2s) regardless of the network type, and to reduce the rendering data rate (playback speed), and then smoothly restoring normal rendering data rate (e.g. to the same rate as the server data rate).
  • This change can be performed in such a way that there is no visible/hearable distortion, in other words such that an end-user noticeable artifact can be avoided.
  • This is achieved by changing the video sample duration (e.g. display each frame during 50ms rather than 40ms) and using an audio time stretching algorithm to change sample duration without impacting the pitch.
  • Pitch represents the perceived fundamental frequency of a sound. It is one of the three major auditory attributes of sounds along with loudness and timbre. While the actual fundamental frequency can be precisely determined through physical measurement, it may differ from the perceived pitch because of overtones, also known as partials, harmonic or otherwise, in the sound. Depending on the stretching algorithm quality, a +/-10% consume rate tuning can be achieved without noticeable artifact, i.e. without affecting the audio pitch.
  • the method according to the invention can comprise performing a normal buffering (preferably no longer than 2s); starting the rendering, with a lower rendering rate; fine-tuning the rendering rate depending on the terminal buffer to achieve server rate and targeted buffer duration; rendering rate equals server rate, and terminal has the targeted buffer protection.
  • a normal buffering preferably no longer than 2s
  • starting the rendering with a lower rendering rate
  • fine-tuning the rendering rate depending on the terminal buffer to achieve server rate and targeted buffer duration
  • rendering rate equals server rate, and terminal has the targeted buffer protection.
  • Fig. 1 shows a schematic representation of a data processing system according to the prior art
  • Fig. 2 shows a schematic representation of a data processing system according to a first embodiment
  • Fig. 3 shows a schematic representation of a data processing system according to a second embodiment
  • Fig. 4 shows a schematic representation of a resampling in a method of buffering streaming data according to a third embodiment
  • Fig. 5 shows a diagram of the buffer filling versus the reception data rate according to the invention
  • Fig. 6 shows a block diagram of a terminal device according to a fourth embodiment.
  • Fig. 2 shows a schematic representation of a data processing system according to a first embodiment.
  • the data processing system comprises a server S, at least one terminal T and a network N.
  • Streaming data can be transmitted from the server S at a server data rate Cs via the network N and from the network N at a reception data rate C rec to the terminal T.
  • the terminal comprises a buffer AL for buffering the streaming data.
  • the terminal T can render the buffered data at a rending data rate C ren .
  • the network N can comprise a buffer B with a network buffering time BL.
  • the buffer B contains data received from the server RFS but not yet sent to the terminal STT.
  • a method of buffering the streaming data is depicted, where normal buffering during a shorter time slot is implemented.
  • the first step comprises performing buffering within a time slot that comprises substantially 1/3 of the time slot for normal terminal buffering.
  • the first step represents the initial stage; the player or terminal has sent a command to the server S to start a streaming session.
  • the server S starts sending media data to the network N, and the network N brings the media data to the terminal T.
  • the network N can act as a buffer B or not (i. e. a network buffering time BL can be present).
  • the terminal T does not consume the data, so the terminal buffer AL has a terminal buffer time ALT which is growing to finally reach a threshold ALTl.
  • a few media data ⁇ 0.5s
  • prefetch overall media rendering chain e.g.
  • the terminal buffer time threshold ALTl is generally at least equal to 1.5s. Once the ALTl threshold has been achieved, the terminal T moves to the second step.
  • Fig. 3 shows a schematic representation of a data processing system according to a second embodiment.
  • the data processing system comprises a server S, a terminal T and a network N for transporting data from the server S to the terminal T.
  • the network can comprise a buffer B with a network buffer time BL.
  • the buffer B contains data received from the server RFS but not yet sent to the terminal STT.
  • a situation is depicted, where the rendering of buffered data with a lower rendering rate C ren is started.
  • the second step comprises starting the rendering with a lower rendering rate.
  • the terminal T starts the rendering with a low rendering rate.
  • the server S still sends its data to the network at the same data rate Cs, and the network N brings them to the terminal T which receives them at a reception data rate C rec .
  • the main difference with step 1 is that the terminal consumes data at the rendering rate C ren .
  • this rate C ren is lower than the server data rate Cs (e.g. ⁇ 10% in a preferred implementation, other reductions are also possible), so for example:
  • Fig. 4 shows a schematic representation of a resampling in the method of buffering streaming data according to a third embodiment.
  • the original rendering data rate ORDR is resampled to a lower rendering data rate LRDR.
  • the lowering of the rendering data rate requires a resampling of digital content to a lower rate.
  • Media resampling is a well known technique as will be explained with reference to Fig. 4.
  • Video resampling (lower) can be performed by increasing the laps time between two video frames (e.g. moving from 40 ms to 44.4ms with 10% lower rate). It does not require any complex interpolation as deviation is very limited. Moreover this processing does not create visible artifacts.
  • a +/- 10% rendering rate tuning can be achieved without impacting the audio pitch.
  • a 10% lower speed rate will be 13.5 frames per seconds (73 milli-seconds between frames), and without any noticeable audio artifact.
  • this step is a transitory step. Terminal does not stay in this step, and moves to the third step.
  • a third step in the method according to the invention is described. It relates to the fine tuning of the rendering rate depending on the terminal buffer in order to achieve a nominal rendering rate and a targeted buffer duration.
  • the rendering rate speed-up according to the invention must be as smooth as possible to avoid artifacts in case of a sudden rendering rate change.
  • step 3 will move smoothly from 10% (step 2) to 0% (step 4) by 1% substeps.
  • the buffer filling over the time is shown in Fig. 5.
  • a graph is depicted showing the relation between the rendering data ratio RDRR with respect to the normalized buffer filling BF. So according to the terminal buffer filling, the appropriate reception data rate can be tuned in order to match the targeted terminal filling.
  • Step 3 takes some time to operate, but it does not have incidence on the end-user experience, as media rendering has started at step 2.
  • the first criterion comprises the number of steps (and step value) and the rules to move from lower rate (eg 10%) to 0%. This is an implementation choice and its mainly depends on the audio time stretcher quality. It is noted that several rules can be used (linear or logarithm) with a high number of steps or a very limited number of steps.
  • the fourth step relates to a situation where the rendering rate equals the server rate and the terminal has a targeted buffer protection.
  • the terminal has achieved the targeted buffer protection ALT2.
  • the step 3 approach can also be used during overall rendering experience. Since the reception data rate is not constant, the terminal buffer is not stable, so using step 3 allows minimizing the media streaming interruption during rendering experience.
  • Fig. 6 shows a block diagram of a terminal device according to a fourth embodiment.
  • the terminal device comprises a terminal buffer AL for a buffering of incoming streaming data, a control unit CU and a rendering unit RU for rendering streaming data buffered in the terminal buffer AL.
  • the control unit CU is adapted to request a streaming session from the server S. Furthermore, when the streaming session has started and streaming data is buffered in the terminal buffer for a first period, the control unit CU is adapted to initiate a rendering of the buffered data at a first rendering rate after the first period. The first rendering data rate is lower than the server data rate. Moreover, the control unit CU is adapted to adapt the first rendering data rate according to the filling of the terminal buffer with the received streaming data.
  • the terminal device is adapted to perform the method of buffering streaming data as described according to the first, second or third embodiment.
  • the invention targets any terminal performing media streaming on a network with a variable reception data rate. It mainly addresses mobile terminals, but it can also address PC's with an over-the-air network (Wifi, WiMax, 2.5G , 3G, LTE etc.).

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Engineering & Computer Science (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A method of streaming data from a server (S) at a server data rate (Cs) via a network to at least one terminal at a terminal reception data rate (Crec) is provided. A streaming section from the server (S) is requested by the terminal (T). Streaming data is forwarded from the server (S) to the network (N) at a server data rate (Cs) and from the network (N) to the terminal (T) at a reception data rate (Crec). Data received from the network (N) is buffered in the terminal buffer (AL) for at least a first period. The rendering of the buffered data is initiated after the first period at a first rendering rate (Cren), which is lower than the server data rate (Cs) or the reception data rate (Crec). The first rendering data rate (Cren) is adapted according to the filling of the terminal buffer (AL) with received streaming data until the rendering data rate (Cren) corresponds to the server data rate (Cs).

Description

METHOD FOR BUFFERING STREAMING DATA AND A TERMINAL DEVICE
FIELD OF THE INVENTION
The invention relates to a method for buffering streaming data in audio/video applications and a terminal device.
BACKGROUND OF THE INVENTION
Buffering is a common technique deployed in streaming use cases, but it may result in a poor user experience. Buffering is required in the following cases: (1) streaming session start-up (including channel zapping, from one to another one), and (2) jump/seek action initiated by the end user. An example of a data processing system including a buffer for real-time date is known from US 6,247,072, which discloses apparatus and methods for matching data rates which is useful for a receiver receiving real-time data over a medium. Implementations feature a process establishing a buffer in a receiver; receiving source data from a source having a nominal source data rate, the received source data arriving at an incoming data rate that differs from time-to-time from the nominal source data rate; filling the buffer with source data as it is received at the incoming data rate and emptying the buffer to provide data for consumption in real time at a consumption data rate; setting a rate-matching factor M, the factor M affecting the rate at which the buffer is emptied; and tracking the level of data in the buffer and resetting the value of M to increase the rate at which the buffer is emptied when the buffer fills above a target range, and resetting the value of M to decrease the rate at which the buffer is emptied when the buffer empties below a target range.
Analog media is sampled to be digitalized. This process uses different sample rates depending on audio or video. On one side, audio is commonly sampled from 8kHz to 48kHz depending on the use case (< 16kHz for voice content, > 16kHz for music content). On the other side, video may be sampled at 24Hz by cinema camera, 25Hz by PAL : European TV standard, about 30fps by NTSC : US TV standard. During a streaming use case, both the server and the terminal send/consume the media at the same data rate; the server data rate and the rendering data rate are equal and match a real time clock. Cable or terrestrial digital television have constant data throughput, for example the server data rate and the reception data rate are equal. Thus, sent data will be received and rendered by the terminal after a constant delay (transmission time).
Fig.l shows a schematic representation of a data processing system according to the prior art. A server S is in charge of sending media to a terminal T via a network N. The server S has real-time behaviour, e.g. it sends media data associated to the server clock. The network N is in charge of carrying data from the server S to the terminal T. A common way to model a network N is to use a buffer B. This buffer B contains data RFS received from the server, and not yet sent to the terminal STT. In this case the network jitter corresponds to the network buffer duration BL. The terminal T - also referred to as "client" or "receiver" - is in charge of receiving and rendering the media from the network N. A terminal T has a media renderer and a buffer AL to manage reception rate variation from the network and consumption rate from the media renderer.
In the context of cable and terrestrial digital television the server sent rate Cs and the terminal reception rate Crec are equal. Thus, both the network N and the terminal buffer AL are constant. Moreover, the sum of these two buffers is about 2s. Thus, video services have been designed to maximize end-user experience, and thereby the revenues of the broadcaster. The main end-user insights to have a good end-user experience are: a large number of channels (not addressed by this invention), media rendering must not be interrupted in case of network congestion, quick start up time (~2s) (from channel selection to first image display) must be provided, fast zapping time (~2s) (from switch command to a new channel, and first image display of the new channel) must be provided and a quick jump/seek time (~2s) (from jump/seek command and first image display) must be provided.
In a mobile network the data throughput is not constant, for example the server data rate Cs and the reception data rate Crec are not the same. For instance, when a mobile terminal has a limited network coverage, the server data rate Cs can be higher than the reception data rate Crec. This triggers an increase of the network buffer B and a decrease of the terminal buffer AL decrease (as the consumption rate is constant, equal to the server output rate).
Thus, the terminal buffer AL has to support this variation to avoid media rendering interruption, e.g. terminal buffer time to be null. In order to protect against media rendering interruption, the terminal has to buffer a significant amount of data. For instance, it pauses rendering (rendering data rate Cren is null), and waits for the terminal buffer time to increase. Once the terminal buffer time reaches a threshold, the media rendering starts (rendering data rate equals server data rate). Thus, during this buffering time the playback is paused, and the playback starts again once buffering has finished.
The longer the buffering time, the better protection against bandwidth variation is available. Nevertheless, as the server sends data at real-time, the buffering latency is equal to the end-user latency, e.g. the time elapsing between receiving the first media data and the first image display. This results in a poor end-user experience. In a 2.5G network, a common buffering time is about 8s. In a 3G network, a common buffering time is about 6s.
This long buffering time is noticeable in two main use cases: (1) at start-up of the streaming session on a media channel (also when the end-user zaps to another channel), and (2) at jump/seek time. Unfortunately the current situation prevents a massive video streaming adoption by the end-user in mobile networks because the main end-user insights have not been achieved. This poor streaming experience prevents operators to sell and deploy additional value services such as video streaming services, even if they are requested by the end-user. Thus, operators are benchmarking terminal solutions to select the one which does have a minimal terminal buffering time.
SUMMARY OF THE INVENTION
It is an object of the invention to provide a method of buffering streaming data and a terminal device which enables the end-user experience by shortening significantly the buffering time.
This object is solved by a method of buffering streaming data according to claim 1 and a terminal device according to claim 6.
Therefore, a method of streaming data from a server at a server data rate via a network to at least one terminal at a terminal reception data rate is provided. A streaming section from the server is requested by the terminal. Streaming data is forwarded from the server to the network at a server data rate and from the network to the terminal at a reception data rate. Data received from the network is buffered in the terminal buffer for at least a first period. The rendering of the buffered data is initiated after the first period at a first rendering rate, which is lower than the server data rate or the reception data rate. The first rendering data rate is adapted according to the filling of the terminal buffer with received streaming data until the rendering data rate corresponds to the server data rate.
Accordingly, it is possible to start with the rendering of the received streaming data earlier than in the prior art. The initial part of the rendering can be faster or slower than the server data rate. As the rendering data rate will only return to its normal value after the first period, i.e. after a certain amount of data rate is buffered in the terminal buffer. This will enable an improved rendering of the data as the rendering will be continuous. This is not possible according to the prior art as there the rendering will be stopped if not sufficient data is present in the terminal buffer. According to an aspect of the invention, the adapting of the first rendering data rate is performed by resampling the streaming data buffered in the terminal buffer.
According to a further aspect of the invention, an audio time stretching is performed to resample the streaming data.
According to a further aspect of the invention, the first rendering data rate is decreased with a decreasing filling of the terminal buffer.
According to a further aspect of the invention, the rendering data rate is increased with an increased filling of the terminal buffer.
The invention also relates to a terminal device for receiving streaming data from a server via a network. The terminal device comprises at least one terminal buffer for buffering data received from the network at a reception data rate for at least a first period. The terminal device furthermore comprises a rendering unit for rendering streaming data buffered in the terminal buffer. The terminal device also comprises at least one control unit for requesting a streaming session from the server, for initiating a rendering of the buffered data after the first period at a first rendering rate by the rendering unit and for adapting the first rendering data rate according to the filling of the terminal buffer with received streaming data until the rendering data rate corresponds to the server data rate. The first rendering data rate is lower than the server data rate when the rendering of the streaming data is initiated.
This can be e.g. solved by tuning the rendering clock and stretching the audio playback to avoid noticeable effects of such tuning. The invention also relates to a data processing system having a server, at least one terminal device and a network to which the server and the terminal device are coupled to. The server sends stream data via the network to the at least one terminal device. The terminal device is implemented as described above.
The invention proposes to reduce buffering latency (for example to about 2s) regardless of the network type, and to reduce the rendering data rate (playback speed), and then smoothly restoring normal rendering data rate (e.g. to the same rate as the server data rate). This change can be performed in such a way that there is no visible/hearable distortion, in other words such that an end-user noticeable artifact can be avoided. This is achieved by changing the video sample duration (e.g. display each frame during 50ms rather than 40ms) and using an audio time stretching algorithm to change sample duration without impacting the pitch.
Pitch represents the perceived fundamental frequency of a sound. It is one of the three major auditory attributes of sounds along with loudness and timbre. While the actual fundamental frequency can be precisely determined through physical measurement, it may differ from the perceived pitch because of overtones, also known as partials, harmonic or otherwise, in the sound. Depending on the stretching algorithm quality, a +/-10% consume rate tuning can be achieved without noticeable artifact, i.e. without affecting the audio pitch.
In particular, the method according to the invention can comprise performing a normal buffering (preferably no longer than 2s); starting the rendering, with a lower rendering rate; fine-tuning the rendering rate depending on the terminal buffer to achieve server rate and targeted buffer duration; rendering rate equals server rate, and terminal has the targeted buffer protection.
Further aspects of the invention are defined in the dependent claims. The steps will be explained with reference to the accompanying drawings. An important advantage of the invention is that an operator can provide a better end-user experience without performing any infrastructure change, neither of the server nor of the network.
Embodiments and advantages of the invention will now be described in more detail with reference to the figures.
Fig. 1 shows a schematic representation of a data processing system according to the prior art,
Fig. 2 shows a schematic representation of a data processing system according to a first embodiment,
Fig. 3 shows a schematic representation of a data processing system according to a second embodiment, Fig. 4 shows a schematic representation of a resampling in a method of buffering streaming data according to a third embodiment, Fig. 5 shows a diagram of the buffer filling versus the reception data rate according to the invention, and Fig. 6 shows a block diagram of a terminal device according to a fourth embodiment. DESCRIPTION OF PREFERRED EMBODIMENTS
Fig. 2 shows a schematic representation of a data processing system according to a first embodiment. The data processing system comprises a server S, at least one terminal T and a network N. Streaming data can be transmitted from the server S at a server data rate Cs via the network N and from the network N at a reception data rate Crec to the terminal T. The terminal comprises a buffer AL for buffering the streaming data. The terminal T can render the buffered data at a rending data rate Cren. The network N can comprise a buffer B with a network buffering time BL. The buffer B contains data received from the server RFS but not yet sent to the terminal STT. In Fig. 2, a method of buffering the streaming data is depicted, where normal buffering during a shorter time slot is implemented.
The first step comprises performing buffering within a time slot that comprises substantially 1/3 of the time slot for normal terminal buffering. The first step represents the initial stage; the player or terminal has sent a command to the server S to start a streaming session. As described in the above scheme, the server S starts sending media data to the network N, and the network N brings the media data to the terminal T. Depending on the network congestion the network N can act as a buffer B or not (i. e. a network buffering time BL can be present). During this phase the terminal T does not consume the data, so the terminal buffer AL has a terminal buffer time ALT which is growing to finally reach a threshold ALTl. On most mobile platforms a few media data (~ 0.5s) is required to prefetch overall media rendering chain (e.g. avoid low level platform interruption). Moreover, extra second(s) is/are generally required to absorb video frame size variations in case of a video coding algorithm based on a predictive scheme such as MPEG-4 or H.263/H.264. So the terminal buffer time threshold ALTl is generally at least equal to 1.5s. Once the ALTl threshold has been achieved, the terminal T moves to the second step.
Fig. 3 shows a schematic representation of a data processing system according to a second embodiment. Here, the data processing system comprises a server S, a terminal T and a network N for transporting data from the server S to the terminal T. The network can comprise a buffer B with a network buffer time BL. The buffer B contains data received from the server RFS but not yet sent to the terminal STT. In Fig. 3, a situation is depicted, where the rendering of buffered data with a lower rendering rate Cren is started. The second step comprises starting the rendering with a lower rendering rate. During this phase, the terminal T starts the rendering with a low rendering rate. The server S still sends its data to the network at the same data rate Cs, and the network N brings them to the terminal T which receives them at a reception data rate Crec. The main difference with step 1 is that the terminal consumes data at the rendering rate Cren. However, this rate Cren is lower than the server data rate Cs (e.g. ~ 10% in a preferred implementation, other reductions are also possible), so for example:
Figure imgf000008_0001
Having a rendering data rate Cren lower than the server data rate Cs has major advantages. It minimizes rendering interruption, because there are more data received than consumed. It is noted that that interruption may still occur if Crec < Crenwhich happens if network conditions are very bad. In this case, interruption occurrence will be reduced by means of the method according to the invention. It continues filling the terminal buffer (see the third step).
Fig. 4 shows a schematic representation of a resampling in the method of buffering streaming data according to a third embodiment. The original rendering data rate ORDR is resampled to a lower rendering data rate LRDR. The lowering of the rendering data rate requires a resampling of digital content to a lower rate. Media resampling is a well known technique as will be explained with reference to Fig. 4. Video resampling (lower) can be performed by increasing the laps time between two video frames (e.g. moving from 40 ms to 44.4ms with 10% lower rate). It does not require any complex interpolation as deviation is very limited. Moreover this processing does not create visible artifacts.
However, it should be noted that the same technique can not be used for audio, because it will lead unpleasant effects (like speeding up/down the audio tape playback). Basic resampling changes the sampling frequency and the audio pitch which creates a noticeable artifact. To avoid this artifact, the proposed solution uses an audio time stretcher algorithm. This algorithm allows resampling an audio stream without changing its pitch or its sampling frequency.
Depending on the audio time stretching quality, a +/- 10% rendering rate tuning can be achieved without impacting the audio pitch. In this case rather than displaying a 15 frames per second stream (66 milli-seconds between frames) a 10% lower speed rate will be 13.5 frames per seconds (73 milli-seconds between frames), and without any noticeable audio artifact. It is noted that this step is a transitory step. Terminal does not stay in this step, and moves to the third step. In the following, a third step in the method according to the invention is described. It relates to the fine tuning of the rendering rate depending on the terminal buffer in order to achieve a nominal rendering rate and a targeted buffer duration.
The rendering rate speed-up according to the invention must be as smooth as possible to avoid artifacts in case of a sudden rendering rate change. According to a preferred embodiment, if a 1% linear step is used, for example with a 10% lower rate at step 2, then step 3 will move smoothly from 10% (step 2) to 0% (step 4) by 1% substeps. The buffer filling over the time is shown in Fig. 5. In Fig. 5, a graph is depicted showing the relation between the rendering data ratio RDRR with respect to the normalized buffer filling BF. So according to the terminal buffer filling, the appropriate reception data rate can be tuned in order to match the targeted terminal filling. Step 3 takes some time to operate, but it does not have incidence on the end-user experience, as media rendering has started at step 2. This lead time mainly depends on two criteria. The first criterion comprises the number of steps (and step value) and the rules to move from lower rate (eg 10%) to 0%. This is an implementation choice and its mainly depends on the audio time stretcher quality. It is noted that several rules can be used (linear or logarithm) with a high number of steps or a very limited number of steps. The second criterion comprises the reception data rate crec- This value is dynamic; it depends on the network load. Its variation will impact the buffer filling, so the terminal may have to increase or decrease the rendering data rate according to the buffer filling. This stage ends (moves to step 4) once the targeted terminal buffer time ALT2 has been reached with a 0% rendering data rate ratio: eg Cren = Cs
In the following, the fourth step in the method according to the invention is described. The fourth step relates to a situation where the rendering rate equals the server rate and the terminal has a targeted buffer protection. In this step, the terminal has achieved the targeted buffer protection ALT2. The rendering data rate equals the server data rate (Cren = Cs), so the terminal is in an optimal streaming condition. It is noted that the step 3 approach can also be used during overall rendering experience. Since the reception data rate is not constant, the terminal buffer is not stable, so using step 3 allows minimizing the media streaming interruption during rendering experience.
Fig. 6 shows a block diagram of a terminal device according to a fourth embodiment. The terminal device comprises a terminal buffer AL for a buffering of incoming streaming data, a control unit CU and a rendering unit RU for rendering streaming data buffered in the terminal buffer AL. The control unit CU is adapted to request a streaming session from the server S. Furthermore, when the streaming session has started and streaming data is buffered in the terminal buffer for a first period, the control unit CU is adapted to initiate a rendering of the buffered data at a first rendering rate after the first period. The first rendering data rate is lower than the server data rate. Moreover, the control unit CU is adapted to adapt the first rendering data rate according to the filling of the terminal buffer with the received streaming data.
It should be noted, that the terminal device according to the fourth embodiment is adapted to perform the method of buffering streaming data as described according to the first, second or third embodiment. The invention targets any terminal performing media streaming on a network with a variable reception data rate. It mainly addresses mobile terminals, but it can also address PC's with an over-the-air network (Wifi, WiMax, 2.5G , 3G, LTE etc.).
It is remarked that the scope of protection of the invention is not restricted to the embodiments described herein. Neither is the scope of protection of the invention restricted by the reference symbols in the claims. The word 'comprising' does not exclude other parts than those mentioned in a claim. The word 'a(n)' preceding an element does not exclude a plurality of those elements. Means forming part of the invention may both be implemented in the form of dedicated hardware or in the form of a programmed general-purpose processor. The invention resides in each new feature or combination of features.

Claims

CLAIM:
1. Method of buffering streaming data from a server (S) at a server data rate (Cs) via a network (N) to at least one terminal device (T) at a terminal reception data rate (Crec), comprising the steps of: requesting a streaming season from the server (S) by the terminal device (T); forwarding streaming data from the server (S) to the network (N) at a server data rate
(Ls) and from the network (N) to the terminal device (T) at a reception data rate (Crec); buffering data received from the network (N) in a terminal buffer (AL) of the terminal (T) for at least a first period; initiate rendering the buffered data after the first period at a first rendering rate (Cren) which is lower than the server data rate (Ls) or the reception data rate (Crec); adapting the first rendering data rate (Cren) according to a filling of the terminal buffer (AL) with received streaming data until the rendering data rate (Cren) corresponds to the server data rate (Cs).
2. Method according to claim 2, wherein the adapting of the first rendering data rate (Cren) is performed by resampling the streaming data buffered in the terminal buffer (AL).
3. Method according to claim 1 or 2, wherein an audio time stretching is performed to resample the streaming data.
4. Method according to claim 1, 2 or 3, wherein the first rendering data rate (Cren) is increased with an increasing filling of the terminal buffer (AL).
5. Method according to any one of the claims 1 to 4, wherein the rendering data rate (Cren) is decreased with decreased filling of the terminal buffer (AL).
6. Terminal device (T) for receiving streaming data from a server (S) via a network (N), comprising: at least one terminal buffer (AL) for buffering data received from the network (N) at a reception data rate (Crec) for at least a first period; a rendering unit (RU) for rendering streaming data buffered in the terminal buffer (AL); and at least one control unit (CU) for requesting a streaming season from the server (S); for initiating a rendering the buffered data after the data has been buffered for a first period at a first rendering rate (Cren) by the rendering unit (RU), wherein the first rendering data rate (Cren) is lower than the server data rate (Cs) or the reception data rate (Crec); and for adapting the first rendering data rate (Cren) according to filling of the terminal buffer with received streaming data until the rendering data rate (Cren) corresponds to the server data rate (Cs).
7. Data processing system, comprising: a server (S) for forwarding streaming data, a network (N) for receiving data from the server (S), and at least one terminal device according to claim 6.
PCT/IB2009/054731 2008-10-28 2009-10-26 Method for buffering streaming data and a terminal device WO2010049879A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN2009801427406A CN102197656A (en) 2008-10-28 2009-10-26 Method for buffering streaming data and a terminal device
US13/126,394 US8612552B2 (en) 2008-10-28 2009-10-26 Method for buffering streaming data and a terminal device
EP09747924.0A EP2351371B1 (en) 2008-10-28 2009-10-26 Method for buffering streaming data and a terminal device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR08291010.0 2008-10-28
EP08291010 2008-10-28

Publications (1)

Publication Number Publication Date
WO2010049879A1 true WO2010049879A1 (en) 2010-05-06

Family

ID=42028107

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2009/054731 WO2010049879A1 (en) 2008-10-28 2009-10-26 Method for buffering streaming data and a terminal device

Country Status (1)

Country Link
WO (1) WO2010049879A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9880803B2 (en) 2016-04-06 2018-01-30 International Business Machines Corporation Audio buffering continuity

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6247072B1 (en) * 1998-01-27 2001-06-12 Cisco Technology, Inc. Real-time data rate matching across a medium
US20060104397A1 (en) * 2004-11-13 2006-05-18 Microsoft Corporation System and method for clock drift correction for broadcast audio/video streaming
US20070204056A1 (en) * 2006-02-28 2007-08-30 Sharp Laboratories Of America, Inc. Systems and methods for reducing the effects of variations on the playback of streaming media
US20070236599A1 (en) * 2006-03-31 2007-10-11 Sharp Laboratories Of America, Inc. Accelerated media coding for robust low-delay video streaming over time-varying and bandwidth limited channels

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6247072B1 (en) * 1998-01-27 2001-06-12 Cisco Technology, Inc. Real-time data rate matching across a medium
US20060104397A1 (en) * 2004-11-13 2006-05-18 Microsoft Corporation System and method for clock drift correction for broadcast audio/video streaming
US20070204056A1 (en) * 2006-02-28 2007-08-30 Sharp Laboratories Of America, Inc. Systems and methods for reducing the effects of variations on the playback of streaming media
US20070236599A1 (en) * 2006-03-31 2007-10-11 Sharp Laboratories Of America, Inc. Accelerated media coding for robust low-delay video streaming over time-varying and bandwidth limited channels

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9880803B2 (en) 2016-04-06 2018-01-30 International Business Machines Corporation Audio buffering continuity

Similar Documents

Publication Publication Date Title
US20180091572A1 (en) Streaming Manifest Quality Control
US8040864B2 (en) Map indicating quality of service for delivery of video data to wireless device
US8819750B2 (en) Personal media broadcasting system with output buffer
US7643422B1 (en) Dynamic trans-framing and trans-rating for interactive playback control
US9332050B2 (en) Media streaming with adaptation
US9148679B2 (en) Modification of delivery of video stream to wireless device based upon position/motion of wireless device
US8209733B2 (en) Edge device that enables efficient delivery of video to handheld device
US11765400B2 (en) Systems and methods for selecting an initial streaming bitrate
US9560106B2 (en) Systems and methods for controlling the encoding of a segmented media stream using segment transmit times
US11431777B2 (en) Adaptive bitrate streaming techniques
US20220070519A1 (en) Systems and methods for achieving optimal network bitrate
US9473549B2 (en) Method of playing internet video and related electronic device
CN103905820A (en) Client side video quality self-adaption method and system based on SVC
US10728630B2 (en) Adaptive bitrate streaming techniques
JP5140952B2 (en) Content distribution system, content distribution server, content reproduction terminal, program, and content distribution method
EP2351371B1 (en) Method for buffering streaming data and a terminal device
US20090300687A1 (en) Edge device establishing and adjusting wireless link parameters in accordance with qos-desired video data rate
WO2010049879A1 (en) Method for buffering streaming data and a terminal device
US20240223832A1 (en) Video stream bitrate adjustment method and apparatus, computer device, and storage medium
US20090300686A1 (en) Edge device reception verification/non-reception verification links to differing devices
EP4195626A1 (en) Streaming media content as media stream to a client system
CN118055277A (en) Video processing method, device, equipment and computer storage medium
CN118055260A (en) Video processing method, device, equipment and computer storage medium

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980142740.6

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09747924

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 13126394

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2009747924

Country of ref document: EP