US20230171301A1 - Monitoring Network Conditions - Google Patents

Monitoring Network Conditions Download PDF

Info

Publication number
US20230171301A1
US20230171301A1 US18/096,764 US202318096764A US2023171301A1 US 20230171301 A1 US20230171301 A1 US 20230171301A1 US 202318096764 A US202318096764 A US 202318096764A US 2023171301 A1 US2023171301 A1 US 2023171301A1
Authority
US
United States
Prior art keywords
network
jitter buffer
size
delay
measure
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/096,764
Inventor
Venu Annamraju
Kiran Kumar Ravuri
Mallikarjuna Kamarthi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Imagination Technologies Ltd
Original Assignee
Imagination Technologies Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Imagination Technologies Ltd filed Critical Imagination Technologies Ltd
Priority to US18/096,764 priority Critical patent/US20230171301A1/en
Publication of US20230171301A1 publication Critical patent/US20230171301A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/80Responding to QoS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/11Identifying congestion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0852Delays
    • H04L43/0858One way delays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0852Delays
    • H04L43/087Jitter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • H04L43/106Active monitoring, e.g. heartbeat, ping or trace-route using time related information in packets, e.g. by adding timestamps
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/24Traffic characterised by specific attributes, e.g. priority or QoS
    • H04L47/2416Real-time traffic
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/25Flow control; Congestion control with rate being modified by the source upon detecting a change of network conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/28Flow control; Congestion control in relation to timing considerations
    • H04L47/283Flow control; Congestion control in relation to timing considerations in response to processing delays, e.g. caused by jitter or round trip time [RTT]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/29Flow control; Congestion control using a combination of thresholds
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/70Admission control; Resource allocation
    • H04L47/80Actions related to the user profile or the type of traffic
    • H04L47/801Real time traffic
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/61Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
    • H04L65/612Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/752Media network packet handling adapting media to network capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5646Cell characteristics, e.g. loss, delay, jitter, sequence integrity
    • H04L2012/5649Cell delay or jitter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/5003Managing SLA; Interaction between SLA and QoS
    • H04L41/5019Ensuring fulfilment of SLA
    • H04L41/5025Ensuring fulfilment of SLA by proactively reacting to service quality change, e.g. by reconfiguration after service quality degradation or upgrade
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/508Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement
    • H04L41/509Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement wherein the managed service relates to media content delivery, e.g. audio, video or TV

Definitions

  • This invention relates to a method and device for identifying network conditions.
  • Real-time streaming of multimedia content over the internet has become an increasingly common application in recent years.
  • a wide range of multimedia applications such as on-demand TV, live TV viewing, audio streaming, video conferencing, net meetings, video telephony, voice over internet protocol (VoIP) and many others rely on end-to-end streaming solutions.
  • streaming media applications Unlike a “downloaded” media file, which may be retrieved first in “non-real” time and played back later, streaming media applications require a media source to encode and to transmit a media signal over a network to a media receiver, which must decode and playback the media signal in real time.
  • problems can arise when streaming media across a network, such as the Internet.
  • a network such as the Internet.
  • the network link between a pair of network end-points may become congested, which can lead to a deterioration in the quality of service of that link.
  • the effects of network congestion includes queueing delay, packet loss and blocking of new connections. This, in turn, can lead to a degradation in the quality and timeliness of the media being played out at the receiving end-point.
  • Network elements usually employ buffers to accommodate bursty incoming traffic. These buffers begin to be filled or bloated whenever incoming bandwidth exceeds outgoing bandwidth. This can cause packets to become queued in those buffers for a significant period of time, especially in network elements that have large buffers. In a first-in first-out queuing system, overly large buffers of network elements result in longer queues and higher latency which causes the network to be congested.
  • the device transmitting the media stream may quickly adapt its transmission parameters in response to changes in the network so that real-time playback can be maintained at the receiving device. For example, when the network becomes congested, the transmitting device may reduce the transmission bandwidth of the media stream by reducing the quality of the media so that the media stream does not congest the network further and so that it takes less time for packets to reach the receiving device over the congested network.
  • the transmitting device may reduce the transmission bandwidth of the media stream by reducing the quality of the media so that the media stream does not congest the network further and so that it takes less time for packets to reach the receiving device over the congested network.
  • a method of identifying a network condition between a pair of network devices comprising: monitoring a measure of delay in receiving media packets over the network; monitoring a size of the jitter buffer; and identifying a network condition in dependence on a change in the measure of delay and a variation in the size of the jitter buffer.
  • the identifying step may comprise identifying congestion in the network if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer decreases below a threshold size.
  • the identifying step may comprise identifying a change in a network route between the pair of network devices in dependence on a variation in the size of the jitter buffer about a threshold size.
  • the identifying step may comprise identifying an increase in network delay if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer temporarily decreases below a threshold size.
  • the step of monitoring a measure of delay may comprise: determining a first time period between receiving a first-received packet for an initial media frame and receiving a first-received packet for a subsequent media frame, wherein each received packet comprises a timestamp; determining a second time period between the timestamp of the packet for the initial media frame and the timestamp of the packet for the subsequent media frame; and forming the measure of delay in dependence on the difference between the first and second time periods.
  • the step of monitoring a measure of delay may comprise: determining a first time period between receiving an initial media frame and receiving a subsequent media frame, wherein each received frame comprises a timestamp; determining a second time period between the timestamp of the initial media frame and the timestamp of the subsequent media frame; and forming the measure of delay in dependence on the difference between the first and second time periods.
  • the method may further comprise: adjusting the measure of delay in dependence on the size of the jitter buffer; at a first one of the network devices, sending an indication of said adjusted measure to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • the method may further comprise: at a first one of the network devices, sending an indication of said identified network condition to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • the method may further comprise: at a first one of the network devices, sending an indication of the measure of network delay and the size of the jitter buffer to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • the packets may be RTP packets.
  • the measure of delay may be determined in dependence on RTP timestamps.
  • a data processing device for receiving a stream of media packets via a network, the device comprising: a transceiver configured to receive media packets from another device via the network; a jitter buffer configured to store the received packets; and a controller configured to: monitor a measure of delay in receiving the media packets over the network; monitor a size of the jitter buffer; and identify a network condition in dependence on a change in the measure of delay and a variation in the size of the jitter buffer.
  • the controller may be further configured to identify congestion in the network if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer decreases below a threshold size.
  • the controller may be further configured to identify a change in a network route between the data processing device and said another device in dependence on a variation in the size of the jitter buffer about a threshold size.
  • the controller may be further configured to identify an increase in network delay if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer temporarily decreases below a threshold size.
  • the controller may be further configured to: determine a first time period between receiving a first-received packet for an initial media frame and receiving a first-received packet for a subsequent media frame, wherein each received packet comprises a timestamp; determine a second time period between the timestamp of the packet for the initial media frame and the timestamp of the packet for the subsequent media frame; and form the measure of delay in dependence on the difference between the first and second time periods.
  • the controller may be further configured to: determine a first time period between receiving an initial media frame and receiving a subsequent media frame, wherein each received frame comprises a timestamp; determine a second time period between the timestamp of the initial media frame and the timestamp of the subsequent media frame; and form the measure of delay in dependence on the difference between the first and second time periods.
  • the packets may be RTP packets.
  • the timestamps may be RTP timestamps.
  • the controller may be further configured to adjust the measure of delay in dependence on the size of the jitter buffer; and the transceiver is further configured to send an indication of said adjusted measure to said another device.
  • the transceiver may be configured to send an indication of said identified network condition to said another device.
  • the transceiver may be configured to send an indication of the measure of network delay and the size of the jitter buffer to said another device.
  • a system comprising: a data processing device as described above; said another device; and the network, wherein said another device is configured to receive the indication and adjust a bandwidth for transmission of media to the data processing device in dependence on the indication.
  • a machine readable non-transitory storage medium having encoded thereon machine readable code for implementing the method described above.
  • FIG. 1 shows an example of a transmitting device and a receiving device
  • FIGS. 2 a - 2 d illustrate packet reception timings under various network conditions
  • FIG. 3 depicts a process for identifying network conditions
  • FIG. 4 depicts a process for estimating network delay
  • FIGS. 5 a - 5 d illustrate variations in network delay and jitter buffer size for various network conditions.
  • FIG. 6 depicts a process for identifying network conditions by monitoring the jitter buffer size.
  • FIG. 1 depicts a transmitting device 10 , which may be any suitable device that is capable of generating packet based data such as a computer, smartphone, videophone, etc.
  • the transmitting device 10 comprises a transceiver 11 for connection to a communications network 12 such as the internet or other packet based networks.
  • the transmitting device 10 can transmit and/or receive packets to and/or from the communications network 12 via the transceiver 11 .
  • the transmitting device 10 comprises an encoder 13 for encoding media data (e.g. video and/or audio data) that is to be transmitted over the communications network 12 to a receiving device 20 .
  • the media data may be from a media source such as a camera and/or microphone, storage device, etc (not shown).
  • the encoder 13 may encode the media data into media frames. Each frame may comprise a timestamp or sequence number which indicates the playout order of the frame and/or a time that the frame is to be played out. The timestamp may indicate the time the media from the media source was sampled (i.e. a sampling instant) according to a reference clock or wallclock that is synchronised at the transmitting and receiving devices 10 and 20 .
  • the frames may be encoded so that they are played out at a specified frame rate.
  • the encoder 13 may encode the media data according to a coding standard such as ITU-T Recommendation H.264 or ISO/IEC International Standard 14496-10 (both also known as Advanced Video Coding (AVC)), MPEG-DASH, HTTP Live Streaming or any other suitable codec.
  • a coding standard such as ITU-T Recommendation H.264 or ISO/IEC International Standard 14496-10 (both also known as Advanced Video Coding (AVC)), MPEG-DASH, HTTP Live Streaming or any other suitable codec.
  • the transmitting device 10 comprises a packetiser 14 which receives the encoded media frames from the encoder 13 and packetises the frames into a sequence of packets for transmission over the network 12 via the transceiver 11 .
  • Each frame may be packetised into one or more packets.
  • Each packet may indicate the timestamp for its frame, e.g, in the packet header. Packets that belong to the same media frame may comprise the same timestamp. As mentioned above, the timestamp may be derived from a reference clock or wallclock that is common between the transmitting device 10 and receiving device 20 .
  • the packetiser 14 may packetise the media in accordance with a Real-time Transport Protocol (RTP) standard. Other standardised packet formats may be used.
  • RTP Real-time Transport Protocol
  • the receiving device 20 comprises a transceiver 21 for receiving packets from the network 12 .
  • the packets are provided to a buffer 22 , which may be a jitter buffer that is capable of ordering the packets according to a playout sequence of the media data in the packets. This sequence may be indicated by a sequence number or timestamp contained in each packet.
  • a decoder 23 decodes the packets in the order provided to it by the buffer 22 to form a media stream. The decoder 23 decodes packets according to the codec used by the encoder 13 .
  • a media consumer 24 receives the decoded media stream for playback.
  • the media consumer 24 may be an audio and/or video player, the data packets carrying media data may be RTP packets and the media stream may be video and/or audio stream.
  • condition of the network 12 may change (e.g. become congested), which may lead to a delay in receiving packets at the receiving device 20 .
  • the transmission device 10 may comprise a quality controller 15 , which is capable of adjusting the transmission properties (such as bandwidth, media quality, packet size, etc) of the media stream.
  • the receiving device 20 comprises a quality controller 25 for identifying changes in the network 12 .
  • the quality controller 25 is capable of determining when packets are received by the receiving device 20 .
  • the time that a packet is received may be derived from the wallclock or an internal clock (not shown), which may not necessarily be synchronised with the wallclock.
  • the quality controller 25 is also capable of determining the time indicated by timestamp comprised in each packet.
  • the quality controller 25 is capable of identifying changes in the condition of the network by comparing the times that packets are received with the times indicated by the timestamps of those packets, as discussed in further detail below.
  • the quality controller 25 may send an indication of the identified change to the transmitting device 10 so that it can appropriately adjust its transmission parameters.
  • the quality controller 25 may also adjust some of its reception parameters (e.g. target jitter buffer size) in response to some network conditions.
  • the quality controller 25 uses a congestion measure to identify network conditions between the transmitting and receiving devices such as congestion, route changes, etc.
  • the congestion measure is determined using the inter-arrival times between each newly received frame and a reference frame.
  • each frame may be made up from one or more packets having the same timestamp.
  • the inter-arrival times between frames is determined by comparing the first-received packet for each frame.
  • Each frame has a different timestamp and so the quality controller 25 is able to determine that a packet for a new frame has arrived if the timestamp of that packet has not previously been received.
  • the first-received packet for each frame may not necessarily be the first packet that has been packetised or transmitted by the transmitting device 10 because of packet loss, varying network paths, etc in the network 12 .
  • the first-received packet is the first packet received by the receiving device 20 for a particular frame. Thus, a complete frame is not required to be received (i.e. all of the packets required to build the complete frame) in order to compare the arrival times between frames.
  • Equation 1 is an example of how the comparison could be made:
  • CurrentFrameRx is the time when the first packet is received for the current frame
  • RefFrameRX is the time when the first packet is received for the reference frame
  • CurrentFrameTS is the timestamp for the current frame
  • RefFrameTS is the timestamp in the reference frame.
  • the timestamp for a frame may be comprised in each packet for that frame. Alternatively, the timestamp may be determined from building a complete frame by receiving all of the packets necessary for building that frame.
  • Equation 2 is an example of how the congestion measure can be determined in a system which uses RTP packets to send media data:
  • CurrentFrameRx is the time in milliseconds when the first RTP packet is received for the current frame
  • RefFrameRX is the time milliseconds when the first RTP packet is received for the reference frame
  • CurrentFrameTS is the timestamp comprised in the RTP header of the first RTP packet for the current frame
  • RefFrameTS is the timestamp comprised in the RTP header of the first RTP packet for the reference frame.
  • the reference frame is considered to be a frame that has arrived on-time according to the best available knowledge. Initially, the first frame of the media stream is chosen to be the reference frame. Subsequently, the reference frame is updated whenever a frame arrives on-time or within some threshold.
  • FIGS. 2 a - 2 d illustrate various scenarios which show how the congestion measure can be used to determine the condition of a network.
  • FIG. 2 a illustrates a scenario where the network is operating under normal conditions.
  • Each arrow represents a media frame sent using the RTP protocol.
  • the RTP time stamp (RTPTS) for that frame is indicated above the arrows.
  • the time when then the first packet for each of the frames is received is indicated by the time below the arrows (in milliseconds). This time is shown as a time relative to the first frame, but could also be an absolute time. In this case the first frame is selected to be the reference frame.
  • the congestion measure for each frame is calculated as follows:
  • the congestion measure is a constant value of zero from the 1 st frame to the 21 st frame. This indicates that the frames are reaching the receiving device 20 on-time and so the network is operating satisfactorily for the purposes of the media stream.
  • FIG. 2 b illustrates a scenario where there is congestion in the network.
  • the congestion measure for each frame is calculated as follows:
  • This scenario differs from the FIG. 2 a scenario as the first-received packet for the 2 nd to 21 st frames are being received in an increasingly delayed manner.
  • the congestion measure is proportional to the mismatch between the sending bitrate and the bandwidth of the network. At other times, the congestion measure is close to zero or below some threshold indicating no congestion.
  • the manner in which the congestion measure increases depends on the sending bandwidth, choke bandwidth (or throttling) and network usage. It is possible to infer where the cause of the congestion may lie between devices 10 and 20 .
  • a first set of congestion measure values may be determined at device 20 for packets sent from device 10 and a second set of congestion measure values may be determined at device 10 for packets sent from device 20 .
  • Devices 10 and 20 may share their determined congestion measure values with each other. It may be determined from those shared values that the path from device 10 to device 20 is congested whilst the path from device 20 to device 10 is operating satisfactorily.
  • the congestion may be caused by the sending bandwidth at device 10 rather than due to a reduction in the available bandwidth of network 12 because the congestion is only seen in one direction rather than both directions through the network 12 .
  • it is determined that both paths are congested then it is possible to infer that there is a problem with the network 12 (such as a reduction in the available bandwidth).
  • FIG. 2 c illustrates a scenario where there is a single increase in network delay, which occurs between the second and third frame (as indicated in the figure).
  • the congestion measure for each frame is calculated as follows:
  • the congestion measure increases for the third frame and remains at that increased value for the subsequent frames.
  • the delay is constant (rather than increasing, as is the case for FIG. 2 b )
  • This increase in delay indicates, for example, that there has been a change in the network path (as depicted by the route change arrow in FIG. 2 c ) which leads to a longer amount of time for packets to be transported from the transmitting device 10 to the receiving device 20 .
  • the constant value (50, in this example) is subtracted from the calculation of the congestion measure until the reference frame has been updated.
  • FIG. 2 d illustrates a scenario where there is a single decrease in network delay, which occurs between the third and fourth frames, as indicated in the figure.
  • the congestion measure for each frame is calculated as follows:
  • the congestion measure decreases for the fourth frame.
  • the decrease indicates, for example, that there has been a change in the network path (as depicted by the route change arrow in FIG. 2 d ) which leads to a shorter amount of time for packets to be transported from the transmitting device to the receiving device.
  • the fourth frame is considered to be received on-time and so the reference frame is updated from the first frame to the fourth frame for calculating the congestion measure for subsequent frames.
  • the reference frame is updated occasionally or periodically since clock skew between the sender and receiver may lead to a gradual change in the congestion measure value. It is advantageous to update the reference frame at least once after initialisation, since it may not be known whether the first frame is “on-time” or experiencing congestion.
  • the quality controller 25 may monitor the congestion measure to determine the condition of the network, as described by the scenarios above. The quality controller 25 may then transmit information indicative of the condition of the network to the transmitting device 10 . Alternatively or additionally, the quality controller 25 may send the congestion measure and/or the CurrentFrameRx and RefFrameRx values and/or the CurrentFrameTS and RefFrameTS values to the transmitting device 10 so that quality controller 15 can similarly determine the condition of the network 12 . The transmitting device 10 may then adjust its media transmission in dependence on the received information. For example, if it is determined that the network is congested, the quality controller 15 may, in response, cause the transmitting device 10 to reduce its transmission bandwidth.
  • the transmission bandwidth will be reduced to be lower than the congested network bandwidth so that further congestion can be avoided. This will ensure that transmitted media packets arrive at the receiving device in a timely manner for playback.
  • the quality controller 15 or 25 may determine the condition of the network based on two or more values of the congestion measure exceeding a threshold. For example, a threshold value for congestion may be provided and if, e.g., three consecutively determined congestion measure values exceed this threshold, then it may be determined that the network is congested. This prevents any spurious or spiked values from falsely classifying the network as congested. Additionally or alternatively, an average of two or more congestion measure values may be used for determining the network condition rather than a single value. For example, the last three congestion measure values may be averaged to determine a current congestion measure value. Preferably, the number of congestion measure values used in either case is not excessive such that it causes a delay in responding to the detection of a network condition in a timely manner.
  • FIG. 3 is a flow diagram illustrating how network conditions between the transmitting and receiving devices 10 and 20 may be identified.
  • a packet is received and the time that the packet is received is stored.
  • the time for receiving the packet may be determined from an internal clock of the receiving device.
  • the media frame associated with the packet is determined.
  • the media frame associated with a packet may be identified in the header of the packet.
  • step 303 it is determined if the packet is the first-received packet for the identified frame. If the packet is the first-received packet for that frame (i.e. no other packets associated with that frame have been received previously), then the process moves on to step 304 . If the packet is not the first-received packet for that frame (i.e. the packet is associated with a frame for which another packet has previously been received), then the process loops back to step 301 .
  • step 304 it is determined if a reference frame has been previously set. If not, then the frame associated with received packet is set to be the reference frame at step 305 , where the time for receiving that packet (at 301 ) is stored along with the timestamp associated with the identified frame. The process then loops back to step 301 . If a reference frame had been set previously, then the process moves on to step 306 .
  • the time that the packet was received (stored at step 301 ) and the time that the first-received packet for the reference frame was received is compared to determine a first time period.
  • the timestamp of the frame associated with the received packet and the timestamp of the reference frame is compared to determine a second time period.
  • a difference between the first time period (determined at step 306 ) and the second time period (determined at step 307 ) is determined to provide a congestion measure value.
  • the congestion measure value is used to identify network conditions.
  • the congestion measure value determined for one frame or the values for a number of consecutive frames may be used to identify network conditions. For example, if a single congestion measure value is negative then this may indicate that the delay in the network has decreased (e.g. as described above with reference to FIG. 2 d ).
  • the reference frame may then be set to be the frame associated with the negative congestion measure.
  • the congestion measure increases over a consecutive number of frames, then this may indicate that the network is congested (as described above with reference to FIG. 2 b ).
  • the identified condition may then be used to adapt transmission properties of the media stream (as mentioned above).
  • the process of FIG. 3 may be performed for every packet received as part of the same media stream (as indicated by the process looping back to step 301 ).
  • the buffer 22 may be a jitter buffer.
  • a jitter buffer absorbs delay variation in the network delivery of media packets.
  • the jitter buffer works by holding media data for a period of time after it is received and before it is played out. Packets that arrive early are held in the jitter buffer longer. If packets are delayed excessively by the network, they may be discarded if they miss their playout time.
  • the jitter buffer can be considered to be a time window with the early edge aligned with the delay corresponding to the earliest arriving packet and the late edge representing the maximum permissible delay before a late arriving packet would be discarded.
  • the delay applied to packets that arrive on time or at their expected arrival time is known as the nominal delay, and this is equivalent to the time difference/buffer size difference between the insertion point of the on-time packets and the point at which the packets are read out to be played.
  • the jitter buffer 22 may be an adaptive jitter buffer.
  • An adaptive jitter buffer can adapt to changes in the network's delay and has variable size or variable delay.
  • the quality controller 25 can monitor the variation in the size of the adaptive jitter buffer 22 to determine the condition of the network 12 .
  • the quality controller 25 contemporaneously monitors the variation in the size of the adaptive jitter buffer 22 and a measure of delay in receiving packets over the network 12 .
  • the measure of delay may be determined from the time taken for packets to travel from the transmitting device 10 to the receiving device 20 and from the receiving device 20 to the transmitting device 10 , as illustrated in FIG. 4 .
  • Packets 240 and 241 may be packets for communicating media or other data (such as network statistics) between the transmitting and receiving devices 10 and 20 .
  • the packets 240 and 241 may each have one or more of the following information contained within it: Local Packet Generation Time (LPGT) and/or Remote Packet Generation Time (RPGT) and/or Packet Idle Time (PIT).
  • LPGT Local Packet Generation Time
  • RPGT Remote Packet Generation Time
  • PIT Packet Idle Time
  • every packet transmitted for the call may incorporate the LPGT information, which for packet 240 is T 0 .
  • the packet generation time at device 10 (i.e. T 0 ) of the last received packet 240 is included in packet 241 .
  • Device 10 can then estimate the network delay for a round trip from the information in packet 241 as:
  • Network delay packet 241 received time ( T 3) ⁇ RPGT ( T 0) ⁇ PIT ( T 2 ⁇ T 1) (3)
  • the network delay (T 1 ⁇ T 0 )+(T 3 ⁇ T 2 ), which is the time spent in transmission over the network for a round trip.
  • the measure of delay in receiving packets over the network may be:
  • the congestion measure mentioned above is used as the measure of delay in receiving packets over the network.
  • the congestion measure is a measure of delay in a single direction (i.e. from transmitting device to receiving device) rather than a round-trip measure of delay and thus may provide a more instantaneous measure of network delay than the method described in relation to FIG. 4 .
  • the congestion measure is insensitive to variations in packet loss, out-of-order packets, retransmission of packets and packet sizes, as mentioned above.
  • FIGS. 5 a - 5 d illustrate various scenarios which show how the monitoring the jitter buffer size and a measure of network delay can be used to determine the condition of a network.
  • the congestion measure is used as the measure of delay.
  • any other measure of delay could be used (such as the network delay calculated using equation 3).
  • FIG. 5 a illustrates a scenario where the network 12 becomes congested.
  • a first graph 401 which shows the variation in the congestion measure over a period of time and a second graph 402 which shows the variation in the adaptive jitter buffer size over the same period of time.
  • the average jitter buffer size is within a small tolerance of a target jitter buffer size (JB EST ) 403 .
  • the tolerance is ⁇ , as shown in the diagram.
  • the tolerance may be ⁇ 10 ms of the target jitter buffer size.
  • the network starts to become congested. This causes the congestion measure to increase (which indicates that the network delay is increasing) and the size of the adaptive jitter buffer to decrease.
  • the size of the jitter buffer decreases to below the tolerance level, ⁇ , and towards zero. This indicates that frames are being received at a rate that is slower than the rate at which they should be played out (and thus at a rate that is slower than they are transmitted). This in turn indicates that the network is congested.
  • a further indicator that the network is congested is that the gradient of the decrease in the jitter buffer size is proportional to the gradient of the increase in the congestion measure (as depicted at 405 ).
  • FIG. 5 b illustrates a scenario where there is a one way increase in network delay.
  • there is an increase in the network delay This may be caused by, for example, a change in the network route between the transmitting and receiving devices 10 and 20 .
  • the congestion measure 401 increases and the jitter buffer size 402 begins to decrease.
  • the jitter buffer size 402 decreases to below the tolerance level, ⁇ .
  • the measure of delay does not increase (i.e. remains substantially constant after time 406 )
  • the jitter buffer size begins to increase again to be within the tolerance level.
  • the increase in network delay causes the congestion measure 401 to increase.
  • the congestion measure remains constant for packets received after the increase in delay (as described with reference to FIG. 2 c ).
  • This constant bias added to the congestion measure 401 may provide false readings to other components that use the congestion measure 401 .
  • the congestion measure may be used by the transmitting device to adapt its transmission bandwidth.
  • the constant bias above zero may cause the transmitting device to, for example, incorrectly adapt its bandwidth.
  • the bias is determined by averaging two or more of the latest congestion measure values (e.g. the last three congestion measure values).
  • the stars 407 in FIG. 5 b depict the calculated bias based on averaging the last three congestion measure values.
  • the bias value is only updated when the condition of the network is stable.
  • the jitter buffer size being within the tolerance level may indicate that the network is stable.
  • the bias value is not updated from just prior to the jitter buffer size dropping to below the tolerance level (at time 408 ) to just after the jitter buffer size returning to within the tolerance level (at time 409 ).
  • the bias may then be removed from the congestion measure value to provide an adjusted congestion measure value 410 , which may then be used by other components (e.g. the quality controller 15 at the transmitting device 10 ), as mentioned above.
  • FIG. 5 c illustrates a scenario where there is a one way decrease in network delay.
  • there is a decrease in the network delay This may be caused by, for example, a change in the network route between the transmitting and receiving devices 10 and 20 .
  • the congestion measure 401 decreases and the jitter buffer size 402 begins to increase.
  • the jitter buffer size 402 increases to above the tolerance level, + ⁇ .
  • the network delay is constant, the jitter buffer size 402 begins to decrease again to be within the tolerance level. This is because the decrease in network delay causes a temporary increase in the amount of packets arriving at the jitter buffer.
  • FIG. 5 c illustrates the behaviour of the congestion measure when there is a decrease in network delay.
  • the reference frame is updated to be the latest frame, as mentioned above.
  • the congestion measure may then become positive until another decrease in network delay is detected.
  • FIG. 5 d illustrates a scenario where the network is bursty. Due to the burstiness, there are no packets received at time 412 . This causes the jitter buffer size 402 to decrease and eventually drop to below the tolerance level, ⁇ . Then, at time 413 , a burst of packets arrive. This causes the congestion measure 401 to spike and the jitter buffer size 402 to suddenly increase above the tolerance level + ⁇ . The burst then ends and the packets begin to arrive at the rate prior to time 412 , which causes the jitter buffer size 402 to return to be within the tolerance level. The rapid decrease and increase below and above the tolerance levels of the jitter buffer indicates that the network is bursty and so the spike in the congestion measure does not indicate that the network is congested.
  • the quality controller 25 of the receiving device 20 contemporaneously monitors the size of the jitter buffer 22 and the network delay (which may be indicated by the congestion measure) to determine the condition of the network, as described by the scenarios above.
  • the quality controller 25 may then transmit information indicative of the condition of the network to the transmitting device 10 .
  • the quality controller 25 may send the information indicating the jitter buffer size and the network delay (e.g. the congestion measure and/or the CurrentFrameRx and RefFrameRx values and/or the CurrentFrameTS and RefFrameTS values) to the transmitting device 10 so that the quality controller 15 of the transmitting device 10 itself can determine the condition of the network 12 .
  • the transmitting device 10 may then adjust its media transmission in dependence on the received information. For example, if it is determined that the network is congested, the transmitting device, in response, may reduce its transmission bandwidth. This may be achieved by, for example, reducing the quality of the media that is to be transmitted, increasing packet size (which reduces the packetisation overhead), reducing error correction redundancy, etc. Preferably, the transmission bandwidth will be reduced to be lower than the congested network bandwidth so that further congestion can be avoided. This will ensure that transmitted media packets arrive at the receiving device in a timely manner for playback.
  • the quality controller 25 at the receiving device 20 may also adjust some reception parameters in response to determining the condition of the network. For example, if it is determined that the network is bursty, the receiving device may increase it's target jitter buffer size to help ensure that the media is played out without any gaps.
  • FIG. 6 is a flow diagram illustrating how network conditions between the transmitting and receiving devices 10 and 20 may be identified from monitoring the size of the jitter buffer.
  • media packets are received at the receiving device 20 .
  • the media packets are temporarily stored at the jitter buffer 22 .
  • a measure of delay in receiving the packets over the network 12 is monitored.
  • the measure of delay may be the congestion measure (as described in relation to FIGS. 2 a - 3 ) or the network delay (as described in relation to FIG. 4 ) or any other suitable measure of delay (such as the other examples mentioned above).
  • the size of the jitter buffer 22 is monitored.
  • the measure of delay and the size of the jitter buffer 22 may be monitored concurrently.
  • the variation in the size of the jitter buffer 22 and the change in the measure of delay with time is analysed so as to identify network conditions, such as those described in relation to FIGS. 5 a - 5 d.
  • the process of FIG. 6 may be performed periodically such that the condition of the network can be monitored for the duration of the media stream.
  • the transmission of the media stream may be adapted to optimise the stream for the new network condition.
  • the transmitting and receiving devices configured in accordance with the examples described herein could be embodied in hardware, software or any suitable combination of hardware and software.
  • the transmitting devices may have the same capabilities as the receiving devices and vice versa.
  • the devices as described herein could comprise, for example, software for execution at one or more processors (such as at a CPU and/or GPU), and/or one or more dedicated processors (such as ASICs), and/or one or more programmable processors (such as FPGAs) suitably programmed so as to provide functionalities of the devices, and/or heterogeneous processors comprising one or more dedicated, programmable and general purpose processing functionalities.
  • the devices described herein can comprise one or more processors and one or more memories having program code stored thereon, the processors and the memories being such as to, in combination, provide the claimed devices and/or perform the claimed methods.
  • Data processing units described herein need not be provided as discrete units and represent functionalities that could (a) be combined in any manner, and (b) themselves comprise one or more data processing entities.
  • Data processing units could be provided by any suitable hardware or software functionalities, or combinations of hardware and software functionalities.
  • software includes executable code for processors (e.g. CPUs and/or GPUs), firmware, bytecode, programming language code such as C or OpenCL, and modules for reconfigurable logic devices such as FPGAs.
  • Machine-readable code includes software and code for defining hardware, such as register transfer level (RTL) code as might be generated in Verilog or VHDL.
  • RTL register transfer level
  • any one or more of the methods described herein could be performed by one or more physical processing units executing program code that causes the unit(s) to perform the methods.
  • the or each physical processing unit could be any suitable processor, such as a CPU or GPU (or a core thereof), or fixed function or programmable hardware.
  • the program code could be stored in non-transitory form at a machine readable medium such as an integrated circuit memory, or optical or magnetic storage.
  • a machine readable medium might comprise several memories, such as on-chip memories, computer working memories, and non-volatile storage devices.

Abstract

A method of identifying a network condition between a pair of network devices, wherein one of the devices comprises a jitter buffer for storing packets received via a network, the method comprising: monitoring a measure of delay in receiving media packets over the network; monitoring a size of the jitter buffer; and identifying a network condition in dependence on a change in the measure of delay and a variation in the size of the jitter buffer.

Description

    BACKGROUND OF THE INVENTION
  • This invention relates to a method and device for identifying network conditions.
  • Real-time streaming of multimedia content over the internet has become an increasingly common application in recent years. A wide range of multimedia applications, such as on-demand TV, live TV viewing, audio streaming, video conferencing, net meetings, video telephony, voice over internet protocol (VoIP) and many others rely on end-to-end streaming solutions. Unlike a “downloaded” media file, which may be retrieved first in “non-real” time and played back later, streaming media applications require a media source to encode and to transmit a media signal over a network to a media receiver, which must decode and playback the media signal in real time.
  • Problems can arise when streaming media across a network, such as the Internet. For example, the network link between a pair of network end-points may become congested, which can lead to a deterioration in the quality of service of that link. The effects of network congestion includes queueing delay, packet loss and blocking of new connections. This, in turn, can lead to a degradation in the quality and timeliness of the media being played out at the receiving end-point.
  • One of the causes of network congestion is buffer-bloat. Network elements usually employ buffers to accommodate bursty incoming traffic. These buffers begin to be filled or bloated whenever incoming bandwidth exceeds outgoing bandwidth. This can cause packets to become queued in those buffers for a significant period of time, especially in network elements that have large buffers. In a first-in first-out queuing system, overly large buffers of network elements result in longer queues and higher latency which causes the network to be congested.
  • When streaming real-time media, for example, it may be desirable for the device transmitting the media stream to quickly adapt its transmission parameters in response to changes in the network so that real-time playback can be maintained at the receiving device. For example, when the network becomes congested, the transmitting device may reduce the transmission bandwidth of the media stream by reducing the quality of the media so that the media stream does not congest the network further and so that it takes less time for packets to reach the receiving device over the congested network. In order to quickly and appropriately adapt the transmission properties of the media stream in response to changeable network conditions, there is a need to be able to detect changes in the condition of the network and to determine the cause of those changes.
  • SUMMARY OF THE INVENTION
  • According to a first aspect there is provided a method of identifying a network condition between a pair of network devices, wherein one of the devices comprises a jitter buffer for storing packets received via a network, the method comprising: monitoring a measure of delay in receiving media packets over the network; monitoring a size of the jitter buffer; and identifying a network condition in dependence on a change in the measure of delay and a variation in the size of the jitter buffer.
  • The identifying step may comprise identifying congestion in the network if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer decreases below a threshold size.
  • The identifying step may comprise identifying a change in a network route between the pair of network devices in dependence on a variation in the size of the jitter buffer about a threshold size.
  • The identifying step may comprise identifying an increase in network delay if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer temporarily decreases below a threshold size.
  • The step of monitoring a measure of delay may comprise: determining a first time period between receiving a first-received packet for an initial media frame and receiving a first-received packet for a subsequent media frame, wherein each received packet comprises a timestamp; determining a second time period between the timestamp of the packet for the initial media frame and the timestamp of the packet for the subsequent media frame; and forming the measure of delay in dependence on the difference between the first and second time periods.
  • The step of monitoring a measure of delay may comprise: determining a first time period between receiving an initial media frame and receiving a subsequent media frame, wherein each received frame comprises a timestamp; determining a second time period between the timestamp of the initial media frame and the timestamp of the subsequent media frame; and forming the measure of delay in dependence on the difference between the first and second time periods.
  • The method may further comprise: adjusting the measure of delay in dependence on the size of the jitter buffer; at a first one of the network devices, sending an indication of said adjusted measure to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • The method may further comprise: at a first one of the network devices, sending an indication of said identified network condition to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • The method may further comprise: at a first one of the network devices, sending an indication of the measure of network delay and the size of the jitter buffer to the other network device; and at said other network device, receiving said indication and adjusting a bandwidth for transmission of media to said first network device in dependence on said indication.
  • The packets may be RTP packets. The measure of delay may be determined in dependence on RTP timestamps.
  • According to a second aspect there is provided a data processing device for receiving a stream of media packets via a network, the device comprising: a transceiver configured to receive media packets from another device via the network; a jitter buffer configured to store the received packets; and a controller configured to: monitor a measure of delay in receiving the media packets over the network; monitor a size of the jitter buffer; and identify a network condition in dependence on a change in the measure of delay and a variation in the size of the jitter buffer.
  • The controller may be further configured to identify congestion in the network if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer decreases below a threshold size.
  • The controller may be further configured to identify a change in a network route between the data processing device and said another device in dependence on a variation in the size of the jitter buffer about a threshold size.
  • The controller may be further configured to identify an increase in network delay if the change in the measure of delay indicates an increase in network delay and the size of the jitter buffer temporarily decreases below a threshold size.
  • The controller may be further configured to: determine a first time period between receiving a first-received packet for an initial media frame and receiving a first-received packet for a subsequent media frame, wherein each received packet comprises a timestamp; determine a second time period between the timestamp of the packet for the initial media frame and the timestamp of the packet for the subsequent media frame; and form the measure of delay in dependence on the difference between the first and second time periods.
  • The controller may be further configured to: determine a first time period between receiving an initial media frame and receiving a subsequent media frame, wherein each received frame comprises a timestamp; determine a second time period between the timestamp of the initial media frame and the timestamp of the subsequent media frame; and form the measure of delay in dependence on the difference between the first and second time periods.
  • The packets may be RTP packets. The timestamps may be RTP timestamps.
  • The controller may be further configured to adjust the measure of delay in dependence on the size of the jitter buffer; and the transceiver is further configured to send an indication of said adjusted measure to said another device.
  • The transceiver may be configured to send an indication of said identified network condition to said another device.
  • The transceiver may be configured to send an indication of the measure of network delay and the size of the jitter buffer to said another device.
  • According to a third aspect there is provided a system comprising: a data processing device as described above; said another device; and the network, wherein said another device is configured to receive the indication and adjust a bandwidth for transmission of media to the data processing device in dependence on the indication.
  • According to a fourth aspect there is provided machine readable code for implementing the method described above.
  • According to a fifth aspect there is provided a machine readable non-transitory storage medium having encoded thereon machine readable code for implementing the method described above.
  • DESCRIPTION OF THE DRAWINGS
  • The present invention will now be described by way of example with reference to the accompanying drawings. In the drawings:
  • FIG. 1 shows an example of a transmitting device and a receiving device;
  • FIGS. 2 a-2 d illustrate packet reception timings under various network conditions;
  • FIG. 3 depicts a process for identifying network conditions;
  • FIG. 4 depicts a process for estimating network delay;
  • FIGS. 5 a-5 d illustrate variations in network delay and jitter buffer size for various network conditions; and
  • FIG. 6 depicts a process for identifying network conditions by monitoring the jitter buffer size.
  • DETAILED DESCRIPTION OF THE DRAWINGS
  • The following description is presented to enable any person skilled in the art to make and use the invention, and is provided in the context of a particular application. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art.
  • The general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present invention. Thus, the present invention is not intended to be limited to the embodiments shown, but is to be accorded the widest scope consistent with the principles and features disclosed herein.
  • FIG. 1 depicts a transmitting device 10, which may be any suitable device that is capable of generating packet based data such as a computer, smartphone, videophone, etc. The transmitting device 10 comprises a transceiver 11 for connection to a communications network 12 such as the internet or other packet based networks. The transmitting device 10 can transmit and/or receive packets to and/or from the communications network 12 via the transceiver 11.
  • The transmitting device 10 comprises an encoder 13 for encoding media data (e.g. video and/or audio data) that is to be transmitted over the communications network 12 to a receiving device 20. The media data may be from a media source such as a camera and/or microphone, storage device, etc (not shown). The encoder 13 may encode the media data into media frames. Each frame may comprise a timestamp or sequence number which indicates the playout order of the frame and/or a time that the frame is to be played out. The timestamp may indicate the time the media from the media source was sampled (i.e. a sampling instant) according to a reference clock or wallclock that is synchronised at the transmitting and receiving devices 10 and 20. The frames may be encoded so that they are played out at a specified frame rate. The encoder 13 may encode the media data according to a coding standard such as ITU-T Recommendation H.264 or ISO/IEC International Standard 14496-10 (both also known as Advanced Video Coding (AVC)), MPEG-DASH, HTTP Live Streaming or any other suitable codec.
  • The transmitting device 10 comprises a packetiser 14 which receives the encoded media frames from the encoder 13 and packetises the frames into a sequence of packets for transmission over the network 12 via the transceiver 11. Each frame may be packetised into one or more packets. Each packet may indicate the timestamp for its frame, e.g, in the packet header. Packets that belong to the same media frame may comprise the same timestamp. As mentioned above, the timestamp may be derived from a reference clock or wallclock that is common between the transmitting device 10 and receiving device 20. The packetiser 14 may packetise the media in accordance with a Real-time Transport Protocol (RTP) standard. Other standardised packet formats may be used. The packetiser 14 provides the packets to the transceiver 11 for transmission over the network 12 to receiving device 20.
  • The receiving device 20 comprises a transceiver 21 for receiving packets from the network 12. The packets are provided to a buffer 22, which may be a jitter buffer that is capable of ordering the packets according to a playout sequence of the media data in the packets. This sequence may be indicated by a sequence number or timestamp contained in each packet. A decoder 23 decodes the packets in the order provided to it by the buffer 22 to form a media stream. The decoder 23 decodes packets according to the codec used by the encoder 13. A media consumer 24 receives the decoded media stream for playback. In the example shown in FIG. 1 , the media consumer 24 may be an audio and/or video player, the data packets carrying media data may be RTP packets and the media stream may be video and/or audio stream.
  • As mentioned above, the condition of the network 12 may change (e.g. become congested), which may lead to a delay in receiving packets at the receiving device 20.
  • This delay may lead to complete frames being received too late for them to be played out on time by the media consumer 24. Thus, it is desirable to determine if the condition of the network 12 has changed and how it has changed so that the transmitting device 10 can appropriately adapt its transmission in order to compensate for the change. The transmission device 10 may comprise a quality controller 15, which is capable of adjusting the transmission properties (such as bandwidth, media quality, packet size, etc) of the media stream.
  • The receiving device 20 comprises a quality controller 25 for identifying changes in the network 12. The quality controller 25 is capable of determining when packets are received by the receiving device 20. The time that a packet is received may be derived from the wallclock or an internal clock (not shown), which may not necessarily be synchronised with the wallclock. The quality controller 25 is also capable of determining the time indicated by timestamp comprised in each packet. The quality controller 25 is capable of identifying changes in the condition of the network by comparing the times that packets are received with the times indicated by the timestamps of those packets, as discussed in further detail below. The quality controller 25 may send an indication of the identified change to the transmitting device 10 so that it can appropriately adjust its transmission parameters. The quality controller 25 may also adjust some of its reception parameters (e.g. target jitter buffer size) in response to some network conditions.
  • The quality controller 25 uses a congestion measure to identify network conditions between the transmitting and receiving devices such as congestion, route changes, etc. The congestion measure is determined using the inter-arrival times between each newly received frame and a reference frame. As mentioned above, each frame may be made up from one or more packets having the same timestamp. The inter-arrival times between frames is determined by comparing the first-received packet for each frame. Each frame has a different timestamp and so the quality controller 25 is able to determine that a packet for a new frame has arrived if the timestamp of that packet has not previously been received. The first-received packet for each frame may not necessarily be the first packet that has been packetised or transmitted by the transmitting device 10 because of packet loss, varying network paths, etc in the network 12. The first-received packet is the first packet received by the receiving device 20 for a particular frame. Thus, a complete frame is not required to be received (i.e. all of the packets required to build the complete frame) in order to compare the arrival times between frames.
  • The inter-arrival times of the first received packet between a newly received frame and a reference frame is compared against the difference in time between the timestamps of those frames. This comparison is used to form the congestion measure. Equation 1 is an example of how the comparison could be made:

  • Congestion measure=(CurrentFrameRx−RefFrameRx)−(CurrentFrameTS−RefFrameTS)   (1)
  • Where CurrentFrameRx is the time when the first packet is received for the current frame, RefFrameRX is the time when the first packet is received for the reference frame, CurrentFrameTS is the timestamp for the current frame and RefFrameTS is the timestamp in the reference frame. As mentioned above, the timestamp for a frame may be comprised in each packet for that frame. Alternatively, the timestamp may be determined from building a complete frame by receiving all of the packets necessary for building that frame.
  • Equation 2 is an example of how the congestion measure can be determined in a system which uses RTP packets to send media data:

  • Congestion measure=(CurrentFrameRx−RefFrameRx)−[(CurrentFrameTS−RefFrameTS)/90]  (2)
  • Where CurrentFrameRx is the time in milliseconds when the first RTP packet is received for the current frame, RefFrameRX is the time milliseconds when the first RTP packet is received for the reference frame, CurrentFrameTS is the timestamp comprised in the RTP header of the first RTP packet for the current frame and RefFrameTS is the timestamp comprised in the RTP header of the first RTP packet for the reference frame. Dividing by 90 converts the RTP time units into milliseconds.
  • If the congestion measure is zero or within some threshold that is close to zero, then it is considered that packets received from the network are “on-time” and so the network is operating satisfactorily for the purposes of the media stream. The reference frame is considered to be a frame that has arrived on-time according to the best available knowledge. Initially, the first frame of the media stream is chosen to be the reference frame. Subsequently, the reference frame is updated whenever a frame arrives on-time or within some threshold.
  • As mentioned above, the time when the first-received packet for a frame is used to determine when a frame has arrived rather than when all of the packets to complete the frame has arrived. This provides a congestion measure that is insensitive to variations in packet loss, out-of-order packets, retransmission of packets and packet sizes.
  • FIGS. 2 a-2 d illustrate various scenarios which show how the congestion measure can be used to determine the condition of a network.
  • FIG. 2 a illustrates a scenario where the network is operating under normal conditions. Each arrow represents a media frame sent using the RTP protocol. The RTP time stamp (RTPTS) for that frame is indicated above the arrows. The time when then the first packet for each of the frames is received is indicated by the time below the arrows (in milliseconds). This time is shown as a time relative to the first frame, but could also be an absolute time. In this case the first frame is selected to be the reference frame. Using equation 2 for RTP packets, the congestion measure for each frame is calculated as follows:
  • 1st frame:

  • Congestion measure=(0−0)−((10000−10000)/90)=0
  • 2nd frame:

  • Congestion measure=(100−0)−((19000−10000)/90)=(100−100)=0
  • 3rd frame:

  • Congestion measure=(200−0)−((28000−10000)/90)=(200−200)=0
  • 4th frame:

  • Congestion measure=(300−0)−((37000−10000)/90)=(300−300)=0
  • 11th frame:

  • Congestion measure=(10000−0)−((910000−10000)/90)=(10000−10000)=0
  • 21st frame:

  • Congestion measure=(20000−0)−((1810000−10000)/90)=(20000−20000)=0
  • In this scenario, the congestion measure is a constant value of zero from the 1st frame to the 21st frame. This indicates that the frames are reaching the receiving device 20 on-time and so the network is operating satisfactorily for the purposes of the media stream.
  • FIG. 2 b illustrates a scenario where there is congestion in the network. Using equation 2, the congestion measure for each frame is calculated as follows:
  • 1st frame:

  • Congestion measure=(0−0)−((10000−10000)/90)=0
  • 2nd frame:

  • Congestion measure=(200−0)−((19000−10000)/90)=(200−100)=100
  • 3rd frame:

  • Congestion measure=(400−0)−((28000−10000)/90)=(400−200)=200
  • 4th frame:

  • Congestion measure=(1000−0)−((37000−10000)/90)=(1000−300)=700
  • 11th frame:

  • Congestion measure=(20000−0)−((910000−10000)/90)=(20000−10000)=10000
  • 21st frame:

  • Congestion measure=(40000−0)−((1810000−10000)/90)=(40000−20000)=20000
  • This scenario differs from the FIG. 2 a scenario as the first-received packet for the 2nd to 21st frames are being received in an increasingly delayed manner. During congestion, the congestion measure, over time, is proportional to the mismatch between the sending bitrate and the bandwidth of the network. At other times, the congestion measure is close to zero or below some threshold indicating no congestion.
  • The manner in which the congestion measure increases depends on the sending bandwidth, choke bandwidth (or throttling) and network usage. It is possible to infer where the cause of the congestion may lie between devices 10 and 20. For example, a first set of congestion measure values may be determined at device 20 for packets sent from device 10 and a second set of congestion measure values may be determined at device 10 for packets sent from device 20. Devices 10 and 20 may share their determined congestion measure values with each other. It may be determined from those shared values that the path from device 10 to device 20 is congested whilst the path from device 20 to device 10 is operating satisfactorily. From this it is possible to infer that the congestion may be caused by the sending bandwidth at device 10 rather than due to a reduction in the available bandwidth of network 12 because the congestion is only seen in one direction rather than both directions through the network 12. Similarly, if it is determined that both paths are congested then it is possible to infer that there is a problem with the network 12 (such as a reduction in the available bandwidth).
  • FIG. 2 c illustrates a scenario where there is a single increase in network delay, which occurs between the second and third frame (as indicated in the figure). Using equation 2, the congestion measure for each frame is calculated as follows:
  • 1st frame:

  • Congestion measure=(0−0)−((10000−10000)/90)=0
  • 2nd frame:

  • Congestion measure=(100−0)−((19000−10000)/90)=(100−100)=0
  • 3rd frame:

  • Congestion measure=(250−0)−((28000−10000)/90)=(250−200)=50
  • 4th frame:

  • Congestion measure=(350−0)−((37000−10000)/90)=(350−300)=50
  • 11th frame:

  • Congestion measure=(10050−0)−((910000−10000)/90)=(10050−10000)=50
  • 21st frame:

  • Congestion measure=(20050−0)−((1810000−10000)/90)=(20050−20000)=50
  • In this scenario the congestion measure increases for the third frame and remains at that increased value for the subsequent frames. As the delay is constant (rather than increasing, as is the case for FIG. 2 b ), this indicates that the delay is not due to congestion. This increase in delay indicates, for example, that there has been a change in the network path (as depicted by the route change arrow in FIG. 2 c ) which leads to a longer amount of time for packets to be transported from the transmitting device 10 to the receiving device 20. Preferably, when this scenario is determined by the quality controller 25, the constant value (50, in this example), is subtracted from the calculation of the congestion measure until the reference frame has been updated.
  • FIG. 2 d illustrates a scenario where there is a single decrease in network delay, which occurs between the third and fourth frames, as indicated in the figure. Using equation 2, the congestion measure for each frame is calculated as follows:
  • 1st frame:

  • Congestion measure=(0−0)−((10000−10000)/90)=0
  • 2nd frame:

  • Congestion measure=(100−0)−((19000−10000)/90)=(100−100)=0
  • 3rd frame:

  • Congestion measure=(200−0)−((28000−10000)/90)=(200−200)=0
  • 4th frame:

  • Congestion measure=(250−0)−((37000−10000)/90)=(250−300)=−50
  • Update the reference frame, RefFrame=FourthFrame
  • 11th frame:

  • Congestion measure=(9950−250)−((910000−37000)/90)=(9700−9700)=0
  • 21st frame:

  • Congestion measure=(19950−250)−((1810000−37000)/90)=(19700−19700)=0
  • In this scenario the congestion measure decreases for the fourth frame. The decrease indicates, for example, that there has been a change in the network path (as depicted by the route change arrow in FIG. 2 d ) which leads to a shorter amount of time for packets to be transported from the transmitting device to the receiving device. In this case, the fourth frame is considered to be received on-time and so the reference frame is updated from the first frame to the fourth frame for calculating the congestion measure for subsequent frames.
  • Preferably (in addition to the FIG. 2 d scenario), the reference frame is updated occasionally or periodically since clock skew between the sender and receiver may lead to a gradual change in the congestion measure value. It is advantageous to update the reference frame at least once after initialisation, since it may not be known whether the first frame is “on-time” or experiencing congestion.
  • The quality controller 25 may monitor the congestion measure to determine the condition of the network, as described by the scenarios above. The quality controller 25 may then transmit information indicative of the condition of the network to the transmitting device 10. Alternatively or additionally, the quality controller 25 may send the congestion measure and/or the CurrentFrameRx and RefFrameRx values and/or the CurrentFrameTS and RefFrameTS values to the transmitting device 10 so that quality controller 15 can similarly determine the condition of the network 12. The transmitting device 10 may then adjust its media transmission in dependence on the received information. For example, if it is determined that the network is congested, the quality controller 15 may, in response, cause the transmitting device 10 to reduce its transmission bandwidth. This may be achieved by, for example, reducing the quality of the media that is to be transmitted, increasing packet size (which reduces the packetisation overhead), reducing error correction redundancy, etc. Preferably, the transmission bandwidth will be reduced to be lower than the congested network bandwidth so that further congestion can be avoided. This will ensure that transmitted media packets arrive at the receiving device in a timely manner for playback.
  • The quality controller 15 or 25 may determine the condition of the network based on two or more values of the congestion measure exceeding a threshold. For example, a threshold value for congestion may be provided and if, e.g., three consecutively determined congestion measure values exceed this threshold, then it may be determined that the network is congested. This prevents any spurious or spiked values from falsely classifying the network as congested. Additionally or alternatively, an average of two or more congestion measure values may be used for determining the network condition rather than a single value. For example, the last three congestion measure values may be averaged to determine a current congestion measure value. Preferably, the number of congestion measure values used in either case is not excessive such that it causes a delay in responding to the detection of a network condition in a timely manner.
  • FIG. 3 is a flow diagram illustrating how network conditions between the transmitting and receiving devices 10 and 20 may be identified.
  • At step 301, a packet is received and the time that the packet is received is stored. As mentioned above, the time for receiving the packet may be determined from an internal clock of the receiving device.
  • At step 302, the media frame associated with the packet is determined. As mentioned above, the media frame associated with a packet may be identified in the header of the packet.
  • At step 303, it is determined if the packet is the first-received packet for the identified frame. If the packet is the first-received packet for that frame (i.e. no other packets associated with that frame have been received previously), then the process moves on to step 304. If the packet is not the first-received packet for that frame (i.e. the packet is associated with a frame for which another packet has previously been received), then the process loops back to step 301.
  • At step 304, it is determined if a reference frame has been previously set. If not, then the frame associated with received packet is set to be the reference frame at step 305, where the time for receiving that packet (at 301) is stored along with the timestamp associated with the identified frame. The process then loops back to step 301. If a reference frame had been set previously, then the process moves on to step 306.
  • At step 306, the time that the packet was received (stored at step 301) and the time that the first-received packet for the reference frame was received is compared to determine a first time period. At step 307, the timestamp of the frame associated with the received packet and the timestamp of the reference frame is compared to determine a second time period.
  • At step 308, a difference between the first time period (determined at step 306) and the second time period (determined at step 307) is determined to provide a congestion measure value.
  • At step 309, the congestion measure value is used to identify network conditions. As mentioned above, the congestion measure value determined for one frame or the values for a number of consecutive frames may be used to identify network conditions. For example, if a single congestion measure value is negative then this may indicate that the delay in the network has decreased (e.g. as described above with reference to FIG. 2 d ). As mentioned above, the reference frame may then be set to be the frame associated with the negative congestion measure. In another example, if the congestion measure increases over a consecutive number of frames, then this may indicate that the network is congested (as described above with reference to FIG. 2 b ). The identified condition may then be used to adapt transmission properties of the media stream (as mentioned above).
  • The process of FIG. 3 may be performed for every packet received as part of the same media stream (as indicated by the process looping back to step 301).
  • As mentioned above, the buffer 22 may be a jitter buffer. A jitter buffer absorbs delay variation in the network delivery of media packets. The jitter buffer works by holding media data for a period of time after it is received and before it is played out. Packets that arrive early are held in the jitter buffer longer. If packets are delayed excessively by the network, they may be discarded if they miss their playout time.
  • The jitter buffer can be considered to be a time window with the early edge aligned with the delay corresponding to the earliest arriving packet and the late edge representing the maximum permissible delay before a late arriving packet would be discarded. The delay applied to packets that arrive on time or at their expected arrival time is known as the nominal delay, and this is equivalent to the time difference/buffer size difference between the insertion point of the on-time packets and the point at which the packets are read out to be played. The jitter buffer 22 may be an adaptive jitter buffer. An adaptive jitter buffer can adapt to changes in the network's delay and has variable size or variable delay. The quality controller 25 can monitor the variation in the size of the adaptive jitter buffer 22 to determine the condition of the network 12.
  • The quality controller 25 contemporaneously monitors the variation in the size of the adaptive jitter buffer 22 and a measure of delay in receiving packets over the network 12.
  • In one example, the measure of delay may be determined from the time taken for packets to travel from the transmitting device 10 to the receiving device 20 and from the receiving device 20 to the transmitting device 10, as illustrated in FIG. 4 .
  • Packets 240 and 241 may be packets for communicating media or other data (such as network statistics) between the transmitting and receiving devices 10 and 20. The packets 240 and 241 may each have one or more of the following information contained within it: Local Packet Generation Time (LPGT) and/or Remote Packet Generation Time (RPGT) and/or Packet Idle Time (PIT).
  • When devices 10 and 20 are communicating, for example during a VoIP call, every packet transmitted for the call may incorporate the LPGT information, which for packet 240 is T0. When packet 240 is received at device 20, the time instant at reception (T1) is measured and when device 20 sends the next packet 241, the time elapsed between the time at reception of the last packet 240 and the time of transmission (T2) of the current packet 241 is incorporated into the packet 241 as the PIT (i.e. PIT=T2−T1 in the example shown in FIG. 4 ). The packet generation time at device 10 (i.e. T0) of the last received packet 240 is included in packet 241. Device 10 can then estimate the network delay for a round trip from the information in packet 241 as:

  • Network delay=packet 241 received time (T3)−RPGT (T0)−PIT (T2−T1)   (3)
  • Therefore, the network delay=(T1−T0)+(T3−T2), which is the time spent in transmission over the network for a round trip.
  • In another examples, the measure of delay in receiving packets over the network may be:
  • measuring inter-arrival times between complete frames.
  • measuring the time period between timestamps of frames or packets received over a fixed period (e.g. one second) and comparing the time period with the fixed period.
  • measuring the spread between pairs of packets of varying sizes and estimate the available bandwidth as a function of the time spread between packets in a pair, and their size. Depending upon the available bandwidth, the packet pairs spread in time as they traverse the network elements.
  • In a preferred embodiment, the congestion measure mentioned above is used as the measure of delay in receiving packets over the network. The congestion measure is a measure of delay in a single direction (i.e. from transmitting device to receiving device) rather than a round-trip measure of delay and thus may provide a more instantaneous measure of network delay than the method described in relation to FIG. 4 . Furthermore, advantageously, the congestion measure is insensitive to variations in packet loss, out-of-order packets, retransmission of packets and packet sizes, as mentioned above.
  • FIGS. 5 a-5 d illustrate various scenarios which show how the monitoring the jitter buffer size and a measure of network delay can be used to determine the condition of a network. In these examples, the congestion measure is used as the measure of delay. However, any other measure of delay could be used (such as the network delay calculated using equation 3).
  • FIG. 5 a illustrates a scenario where the network 12 becomes congested. A first graph 401 which shows the variation in the congestion measure over a period of time and a second graph 402 which shows the variation in the adaptive jitter buffer size over the same period of time. Preferably, the average jitter buffer size is within a small tolerance of a target jitter buffer size (JBEST) 403. The tolerance is ±δ, as shown in the diagram. For example, the tolerance may be ±10 ms of the target jitter buffer size. At time 404, the network starts to become congested. This causes the congestion measure to increase (which indicates that the network delay is increasing) and the size of the adaptive jitter buffer to decrease. The size of the jitter buffer decreases to below the tolerance level, −δ, and towards zero. This indicates that frames are being received at a rate that is slower than the rate at which they should be played out (and thus at a rate that is slower than they are transmitted). This in turn indicates that the network is congested. A further indicator that the network is congested is that the gradient of the decrease in the jitter buffer size is proportional to the gradient of the increase in the congestion measure (as depicted at 405).
  • FIG. 5 b illustrates a scenario where there is a one way increase in network delay. At time 406, there is an increase in the network delay. This may be caused by, for example, a change in the network route between the transmitting and receiving devices 10 and 20. At time 406, the congestion measure 401 increases and the jitter buffer size 402 begins to decrease. Eventually, the jitter buffer size 402 decreases to below the tolerance level, −δ. However, unlike the congestion condition, as the measure of delay does not increase (i.e. remains substantially constant after time 406), the jitter buffer size begins to increase again to be within the tolerance level. This is because the network delay causes a temporary decrease in the amount of packets in the jitter buffer, but as the network is not congested the packets arrive at the same rate as they did previously and so the buffer begins to fill again. Thus, this indicates that there is a one way increase in network delay rather than congestion.
  • The increase in network delay causes the congestion measure 401 to increase. The congestion measure remains constant for packets received after the increase in delay (as described with reference to FIG. 2 c ). This constant bias added to the congestion measure 401 may provide false readings to other components that use the congestion measure 401. For example, as mentioned above, the congestion measure may be used by the transmitting device to adapt its transmission bandwidth. The constant bias above zero may cause the transmitting device to, for example, incorrectly adapt its bandwidth. Thus, it may be advantageous to quickly remove the bias from the congestion measure. Preferably, the bias is determined by averaging two or more of the latest congestion measure values (e.g. the last three congestion measure values). The stars 407 in FIG. 5 b depict the calculated bias based on averaging the last three congestion measure values. Preferably, the bias value is only updated when the condition of the network is stable. The jitter buffer size being within the tolerance level may indicate that the network is stable. Thus, preferably, the bias value is not updated from just prior to the jitter buffer size dropping to below the tolerance level (at time 408) to just after the jitter buffer size returning to within the tolerance level (at time 409). The bias may then be removed from the congestion measure value to provide an adjusted congestion measure value 410, which may then be used by other components (e.g. the quality controller 15 at the transmitting device 10), as mentioned above.
  • FIG. 5 c illustrates a scenario where there is a one way decrease in network delay. At time 411, there is a decrease in the network delay. This may be caused by, for example, a change in the network route between the transmitting and receiving devices 10 and 20. At time 411, the congestion measure 401 decreases and the jitter buffer size 402 begins to increase. Eventually, the jitter buffer size 402 increases to above the tolerance level, +δ. As the network delay is constant, the jitter buffer size 402 begins to decrease again to be within the tolerance level. This is because the decrease in network delay causes a temporary increase in the amount of packets arriving at the jitter buffer. As the network is not congested the packets then start to arrive at the same rate as they previously did and so the buffer begins to empty again and remain within the tolerance level. Thus, this indicates that there is a one way decrease in network delay. FIG. 5 c illustrates the behaviour of the congestion measure when there is a decrease in network delay. Preferably, when a decrease in network delay is detected, the reference frame is updated to be the latest frame, as mentioned above. Thus subsequent to any update, the congestion measure may then become positive until another decrease in network delay is detected.
  • FIG. 5 d illustrates a scenario where the network is bursty. Due to the burstiness, there are no packets received at time 412. This causes the jitter buffer size 402 to decrease and eventually drop to below the tolerance level, −δ. Then, at time 413, a burst of packets arrive. This causes the congestion measure 401 to spike and the jitter buffer size 402 to suddenly increase above the tolerance level +δ. The burst then ends and the packets begin to arrive at the rate prior to time 412, which causes the jitter buffer size 402 to return to be within the tolerance level. The rapid decrease and increase below and above the tolerance levels of the jitter buffer indicates that the network is bursty and so the spike in the congestion measure does not indicate that the network is congested.
  • As mentioned above, the quality controller 25 of the receiving device 20 contemporaneously monitors the size of the jitter buffer 22 and the network delay (which may be indicated by the congestion measure) to determine the condition of the network, as described by the scenarios above. The quality controller 25 may then transmit information indicative of the condition of the network to the transmitting device 10. Alternatively or additionally, the quality controller 25 may send the information indicating the jitter buffer size and the network delay (e.g. the congestion measure and/or the CurrentFrameRx and RefFrameRx values and/or the CurrentFrameTS and RefFrameTS values) to the transmitting device 10 so that the quality controller 15 of the transmitting device 10 itself can determine the condition of the network 12. The transmitting device 10 may then adjust its media transmission in dependence on the received information. For example, if it is determined that the network is congested, the transmitting device, in response, may reduce its transmission bandwidth. This may be achieved by, for example, reducing the quality of the media that is to be transmitted, increasing packet size (which reduces the packetisation overhead), reducing error correction redundancy, etc. Preferably, the transmission bandwidth will be reduced to be lower than the congested network bandwidth so that further congestion can be avoided. This will ensure that transmitted media packets arrive at the receiving device in a timely manner for playback. The quality controller 25 at the receiving device 20 may also adjust some reception parameters in response to determining the condition of the network. For example, if it is determined that the network is bursty, the receiving device may increase it's target jitter buffer size to help ensure that the media is played out without any gaps.
  • FIG. 6 is a flow diagram illustrating how network conditions between the transmitting and receiving devices 10 and 20 may be identified from monitoring the size of the jitter buffer.
  • At step 601, media packets are received at the receiving device 20. At step 602, the media packets are temporarily stored at the jitter buffer 22.
  • At step 603, a measure of delay in receiving the packets over the network 12 is monitored. As mentioned above the measure of delay may be the congestion measure (as described in relation to FIGS. 2 a -3) or the network delay (as described in relation to FIG. 4 ) or any other suitable measure of delay (such as the other examples mentioned above).
  • At step 604, the size of the jitter buffer 22 is monitored. The measure of delay and the size of the jitter buffer 22 may be monitored concurrently. At step 605, the variation in the size of the jitter buffer 22 and the change in the measure of delay with time is analysed so as to identify network conditions, such as those described in relation to FIGS. 5 a -5 d. The process of FIG. 6 may be performed periodically such that the condition of the network can be monitored for the duration of the media stream. When there is a change in the condition of the network, the transmission of the media stream may be adapted to optimise the stream for the new network condition.
  • The transmitting and receiving devices configured in accordance with the examples described herein could be embodied in hardware, software or any suitable combination of hardware and software. The transmitting devices may have the same capabilities as the receiving devices and vice versa. The devices as described herein could comprise, for example, software for execution at one or more processors (such as at a CPU and/or GPU), and/or one or more dedicated processors (such as ASICs), and/or one or more programmable processors (such as FPGAs) suitably programmed so as to provide functionalities of the devices, and/or heterogeneous processors comprising one or more dedicated, programmable and general purpose processing functionalities. The devices described herein can comprise one or more processors and one or more memories having program code stored thereon, the processors and the memories being such as to, in combination, provide the claimed devices and/or perform the claimed methods.
  • Data processing units described herein (e.g. encoder, quality controller and packetizer) need not be provided as discrete units and represent functionalities that could (a) be combined in any manner, and (b) themselves comprise one or more data processing entities. Data processing units could be provided by any suitable hardware or software functionalities, or combinations of hardware and software functionalities.
  • The term software as used herein includes executable code for processors (e.g. CPUs and/or GPUs), firmware, bytecode, programming language code such as C or OpenCL, and modules for reconfigurable logic devices such as FPGAs. Machine-readable code includes software and code for defining hardware, such as register transfer level (RTL) code as might be generated in Verilog or VHDL.
  • Any one or more of the methods described herein could be performed by one or more physical processing units executing program code that causes the unit(s) to perform the methods. The or each physical processing unit could be any suitable processor, such as a CPU or GPU (or a core thereof), or fixed function or programmable hardware. The program code could be stored in non-transitory form at a machine readable medium such as an integrated circuit memory, or optical or magnetic storage. A machine readable medium might comprise several memories, such as on-chip memories, computer working memories, and non-volatile storage devices.
  • The applicant hereby discloses in isolation each individual feature described herein and any combination of two or more such features, to the extent that such features or combinations are capable of being carried out based on the present specification as a whole in the light of the common general knowledge of a person skilled in the art, irrespective of whether such features or combinations of features solve any problems disclosed herein, and without limitation to the scope of the claims. The applicant indicates that aspects of the present invention may consist of any such individual feature or combination of features. In view of the foregoing description it will be evident to a person skilled in the art that various modifications may be made within the scope of the invention.

Claims (20)

What is claimed is:
1. A method of identifying a network condition between a receiving device and a transmitting device, wherein the receiving device comprises a jitter buffer for storing packets received from the transmitting device via a network, the method comprising:
monitoring, over a period of time, a variation in a measure of delay in receiving media packets over the network;
monitoring, over the period of time, a variation in a size of the jitter buffer; and
identifying a network condition by monitoring, over the period of time, for a scenario indicative of that network condition in which:
the measure of delay is increasing, remaining substantially constant and/or decreasing; and
contemporaneously, the size of the jitter buffer is increasing, remaining substantially constant and/or decreasing.
2. The method as claimed in claim 1, wherein the identifying step comprises identifying a decrease in network delay in response to determining that, over the period of time:
(i) the measure of delay decreases and subsequently remains substantially constant; and
(ii) the size of the jitter buffer increases above a threshold size and subsequently decreases below the threshold size whilst the measure of delay remains substantially constant.
3. The method as claimed in claim 2, wherein the threshold size is determined in dependence on a target jitter buffer size of the jitter buffer comprised by the receiving device.
4. The method as claimed in claim 2, wherein the identifying step further comprises identifying that the decrease in network delay is caused by a change in the network route between the transmitting device and the receiving device.
5. The method as claimed in claim 1, wherein the identifying step comprises identifying burstiness in response to determining that, over the period of time:
(i) the size of the jitter buffer decreases below a lower threshold size;
(ii) subsequently to (i), the measure of delay and the size of the jitter buffer both increase, wherein the size of the jitter buffer increases above an upper threshold size; and
(iii) subsequently to (ii), the size of the jitter buffer decreases below the upper threshold size whilst the measure of delay remains substantially constant.
6. The method as claimed in claim 5, wherein the identifying step comprises identifying burstiness in response to determining that, over the period of time:
(ii), subsequently to (i), the measure of delay and the size of the jitter buffer contemporaneously increase, wherein the size of the jitter buffer increases above an upper threshold size.
7. The method as claimed in claim 5, wherein the upper threshold size and the lower threshold size are determined in dependence on a target jitter buffer size of the jitter buffer comprised by the receiving device.
8. The method as claimed in claim 1, wherein the identifying step comprises identifying congestion in the network in response to determining that, over the period of time:
(i) the measure of delay increases; and
(ii) the size of the jitter buffer decreases.
9. The method as claimed in claim 8, wherein the identifying step comprises identifying congestion in the network in response to determining that, over the period of time:
(ii) the size of the jitter buffer decreases below a threshold size.
10. The method as claimed in claim 9, wherein the threshold size is determined in dependence on a target jitter buffer size of the jitter buffer comprised by the receiving device.
11. The method as claimed in claim 8, wherein the identifying step comprises identifying congestion in the network in response to further determining that, over the period of time, the gradient of the decrease in the size of jitter buffer is proportional to the gradient of the increase in the measure of delay.
12. The method as claimed in claim 1, wherein the identifying step comprises identifying an increase in network delay in response to determining that, over the period of time:
(i) the measure of delay increases and subsequently remains substantially constant; and
(ii) the size of the jitter buffer decreases below a threshold size and subsequently increases above the threshold size whilst the measure of delay remains substantially constant.
13. The method as claimed in claim 12, wherein the threshold size is determined in dependence on a target jitter buffer size of the jitter buffer comprised by the receiving device.
14. The method as claimed in claim 12, wherein the identifying step further comprises identifying that the increase in network delay is caused by a change in the network route between the transmitting device and the receiving device.
15. The method as claimed in claim 12, the method further comprising, prior to using the measure of delay for other purposes, adjusting the measure of delay in dependence on a calculated bias.
16. The method as claimed in claim 1, the method further comprising adjusting a transmission parameter of the transmitting device in dependence on the identified network condition.
17. The method as claimed in claim 16, the method comprising adjusting a bandwidth for transmission of media by the transmitting device in dependence on the identified network condition.
18. The method as claimed in claim 1, the method further comprising adjusting a reception parameter of the receiving device in dependence on the identified network condition.
19. The method as claimed in claim 18, the method comprising adjusting a target jitter buffer size of the jitter buffer comprised by the receiving device in dependence on the identified network condition.
20. A receiving device configured to receive a stream of media packets via a network, the receiving device comprising:
a jitter buffer configured to store media packets received from a transmitting device via the network; and
a controller configured to:
monitor, over a period of time, a variation in a measure of delay in receiving the media packets over the network;
monitor, over the period of time, a variation in a size of the jitter buffer; and
identify a network condition by monitoring, over the period of time, for a scenario indicative of that network condition in which:
the measure of delay is increasing, remaining substantially constant and/or decreasing; and
contemporaneously, the size of the jitter buffer is increasing, remaining substantially constant and/or decreasing.
US18/096,764 2015-07-31 2023-01-13 Monitoring Network Conditions Pending US20230171301A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/096,764 US20230171301A1 (en) 2015-07-31 2023-01-13 Monitoring Network Conditions

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
GB1513522.1A GB2535819B (en) 2015-07-31 2015-07-31 Monitoring network conditions
GB1513522.1 2015-07-31
PCT/GB2016/051953 WO2017021682A1 (en) 2015-07-31 2016-06-29 Monitoring network conditions
US201815749369A 2018-01-31 2018-01-31
US18/096,764 US20230171301A1 (en) 2015-07-31 2023-01-13 Monitoring Network Conditions

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US15/749,369 Continuation US11595461B2 (en) 2015-07-31 2016-06-29 Monitoring network conditions
PCT/GB2016/051953 Continuation WO2017021682A1 (en) 2015-07-31 2016-06-29 Monitoring network conditions

Publications (1)

Publication Number Publication Date
US20230171301A1 true US20230171301A1 (en) 2023-06-01

Family

ID=54062962

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/749,369 Active 2037-12-04 US11595461B2 (en) 2015-07-31 2016-06-29 Monitoring network conditions
US18/096,764 Pending US20230171301A1 (en) 2015-07-31 2023-01-13 Monitoring Network Conditions

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/749,369 Active 2037-12-04 US11595461B2 (en) 2015-07-31 2016-06-29 Monitoring network conditions

Country Status (5)

Country Link
US (2) US11595461B2 (en)
EP (1) EP3329641B1 (en)
CN (1) CN107852348B (en)
GB (1) GB2535819B (en)
WO (1) WO2017021682A1 (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2535819B (en) * 2015-07-31 2017-05-17 Imagination Tech Ltd Monitoring network conditions
EP4033770A1 (en) * 2015-11-17 2022-07-27 Livestreaming Sweden AB Video distribution synchronization
CN108011771A (en) * 2016-11-01 2018-05-08 阿里巴巴集团控股有限公司 Detection method, device and the equipment of network link available bandwidth
DE112017007782B4 (en) * 2017-08-28 2022-04-28 Mitsubishi Electric Corporation Wireless communication device, wireless communication method and wireless communication program
US10742532B2 (en) * 2017-12-18 2020-08-11 Futurewei Technologies, Inc. Non-intrusive mechanism to measure network function packet processing delay
TWI696961B (en) * 2018-12-12 2020-06-21 財團法人工業技術研究院 Deep neural networks (dnn) hardware accelerator and operation method thereof
CN109819310B (en) * 2019-01-25 2021-07-27 青岛海信传媒网络技术有限公司 Streaming media data playing method and device
US11277326B2 (en) 2020-03-31 2022-03-15 Netflix, Inc. Techniques for detecting changes to circuit delays in telecommunications networks
CN111901678B (en) * 2020-07-31 2022-08-23 成都云格致力科技有限公司 Anti-jitter smoothing method and system for TCP real-time video stream
CN113220231B (en) * 2021-05-11 2022-07-22 无锡众星微系统技术有限公司 Adaptive flow control method and device supporting STP application
CN113747253B (en) * 2021-08-17 2023-04-28 中移(杭州)信息技术有限公司 Network bandwidth determining method, video RTP receiving end and storage medium
CN114285776B (en) * 2022-01-14 2023-07-14 广州广哈通信股份有限公司 Network line state monitoring method and device and video conference terminal

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10587518B2 (en) * 2015-07-31 2020-03-10 Imagination Technologies Limited Identifying network conditions
US11595461B2 (en) * 2015-07-31 2023-02-28 Imagination Technologies Limited Monitoring network conditions

Family Cites Families (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1997022201A2 (en) 1995-12-12 1997-06-19 The Board Of Trustees Of The University Of Illinois Method and system for transmitting real-time video
US6360271B1 (en) * 1999-02-02 2002-03-19 3Com Corporation System for dynamic jitter buffer management based on synchronized clocks
US6442139B1 (en) 1998-01-29 2002-08-27 At&T Adaptive rate control based on estimation of message queuing delay
KR100584394B1 (en) 1998-12-30 2006-08-30 삼성전자주식회사 Vocoder parameter control method by real time network monitoring
US20020101885A1 (en) 1999-03-15 2002-08-01 Vladimir Pogrebinsky Jitter buffer and methods for control of same
DE60032458T2 (en) * 2000-04-14 2007-04-12 Alcatel Self-adapting dither buffer
US6862298B1 (en) * 2000-07-28 2005-03-01 Crystalvoice Communications, Inc. Adaptive jitter buffer for internet telephony
JP2002314595A (en) 2001-04-10 2002-10-25 Nec Corp Communication repeater and jitter suppression method
JP3900413B2 (en) 2002-02-14 2007-04-04 Kddi株式会社 Video information transmission method and program
JP3769752B2 (en) 2002-12-24 2006-04-26 ソニー株式会社 Information processing apparatus and information processing method, data communication system, and program
WO2004092981A2 (en) 2003-04-07 2004-10-28 Belarc, Inc. Software update and patch audit subsystem for use in a computer information database system
WO2004092891A2 (en) 2003-04-14 2004-10-28 Telchemy, Incorporated System for identifying and locating network problems
FR2857540A1 (en) 2003-07-11 2005-01-14 France Telecom Voice signal processing delay evaluating method for packet switching network e.g. Internet protocol network, involves determining temporary gap between voice signals and calculating delay from gap and predetermined decoding time
JP4349114B2 (en) 2003-12-10 2009-10-21 ソニー株式会社 Transmission device and method, reception device and method, recording medium, and program
CN102209078B (en) 2004-02-13 2013-01-02 诺基亚公司 Timing experience quality metric
KR100641159B1 (en) 2004-07-23 2006-11-02 엘지전자 주식회사 Adaptive method for multimedia transmission rate estimation based on rtcppacket
US8238376B2 (en) * 2005-04-13 2012-08-07 Sony Corporation Synchronized audio/video decoding for network devices
KR100789902B1 (en) * 2005-12-09 2008-01-02 한국전자통신연구원 Apparatus and Method for Transport of a VoIP Packet with Multiple Speech Frames
US7742413B1 (en) * 2006-02-01 2010-06-22 Sprint Communications Company, L.P. Utilizing a null jitter buffer to monitor session traffic
US8069260B2 (en) * 2007-01-12 2011-11-29 Microsoft Corporation Dynamic buffer settings for media playback
TW201019649A (en) 2008-11-06 2010-05-16 Inst Information Industry Network system, adjusting method of data transmission rate and computer program procut thereof
KR101071124B1 (en) * 2009-06-12 2011-10-07 양선주 Method for estimating quality of server of voice and video service in packet networks
EP2302845B1 (en) * 2009-09-23 2012-06-20 Google, Inc. Method and device for determining a jitter buffer level
JP5257313B2 (en) 2009-09-25 2013-08-07 富士通株式会社 Delay time measuring apparatus, delay time measuring program and method
JP5538652B2 (en) * 2010-12-20 2014-07-02 三菱電機株式会社 Network status monitoring method
US8868731B1 (en) 2011-06-06 2014-10-21 Cisco Technology, Inc. Technique for false positives prevention in high availability network
GB2495929B (en) * 2011-10-25 2014-09-03 Skype Jitter buffer
US8627468B2 (en) 2011-11-03 2014-01-07 Verizon Patent And Licensing Inc. Optimizing performance information collection
US9313338B2 (en) * 2012-06-24 2016-04-12 Audiocodes Ltd. System, device, and method of voice-over-IP communication
US9094307B1 (en) 2012-09-18 2015-07-28 Cisco Technology, Inc. Measuring latency within a networking device
PT3321935T (en) * 2013-06-21 2019-09-12 Fraunhofer Ges Forschung Time scaler, audio decoder, method and a computer program using a quality control
US9973402B2 (en) 2013-06-26 2018-05-15 Nec Corporation Transmission device, receiving device, and relay device
US9882818B2 (en) 2013-09-30 2018-01-30 Apple Inc. Adjusting a jitter buffer based on inter arrival jitter
GB2518909B (en) 2013-12-16 2015-10-28 Imagination Tech Ltd Encoder adaptation
US9521069B2 (en) * 2015-05-08 2016-12-13 Ooma, Inc. Managing alternative networks for high quality of service communications

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10587518B2 (en) * 2015-07-31 2020-03-10 Imagination Technologies Limited Identifying network conditions
US11595461B2 (en) * 2015-07-31 2023-02-28 Imagination Technologies Limited Monitoring network conditions

Also Published As

Publication number Publication date
US20180227349A1 (en) 2018-08-09
US11595461B2 (en) 2023-02-28
GB201513522D0 (en) 2015-09-16
GB2535819B (en) 2017-05-17
CN107852348A (en) 2018-03-27
WO2017021682A1 (en) 2017-02-09
GB2535819A (en) 2016-08-31
CN107852348B (en) 2023-09-12
EP3329641A1 (en) 2018-06-06
EP3329641B1 (en) 2021-11-03

Similar Documents

Publication Publication Date Title
US20230171301A1 (en) Monitoring Network Conditions
US11489781B2 (en) Bandwidth management
US10355995B2 (en) Identifying a network condition using estimated processor load
CN106301684B (en) Media data transmission method and device
US10805196B2 (en) Packet loss and bandwidth coordination
US10659380B2 (en) Media buffering
JPWO2006054442A1 (en) Transmitting apparatus, receiving apparatus, and communication system
US11916798B2 (en) Estimating network bandwidth using probe packets
US10587518B2 (en) Identifying network conditions
EP1931068A1 (en) Method of adaptively dejittering packetized signals buffered at the receiver of a communication network node

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED