WO2022002395A1 - Latency control for a communication network - Google Patents
Latency control for a communication network Download PDFInfo
- Publication number
- WO2022002395A1 WO2022002395A1 PCT/EP2020/068580 EP2020068580W WO2022002395A1 WO 2022002395 A1 WO2022002395 A1 WO 2022002395A1 EP 2020068580 W EP2020068580 W EP 2020068580W WO 2022002395 A1 WO2022002395 A1 WO 2022002395A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- communication network
- latency
- user device
- service
- throughput
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W72/00—Local resource management
- H04W72/50—Allocation or scheduling criteria for wireless resources
- H04W72/54—Allocation or scheduling criteria for wireless resources based on quality criteria
- H04W72/543—Allocation or scheduling criteria for wireless resources based on quality criteria based on requested quality, e.g. QoS
Definitions
- the present disclosure relates generally to the field of wireless communication. More particularly, it relates to latency control in wireless communication scenarios.
- the physical product may comprise one or more parts, such as controlling circuitry in the form of one or more controllers, one or more processors, or the like.
- a first aspect is a method for latency control in a communication network.
- the method comprises identifying that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded, predicting an upcoming reduction of communication network throughput for the user device, and configuring a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
- the prediction and configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
- predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future communication network throughput value for the user device is lower than a throughput threshold value.
- predicting the upcoming reduction of communication network throughput for the user device comprises determining that a probability for upcoming reduction of communication network throughput for the user device is larger than a reduction probability threshold value.
- predicting the upcoming reduction of communication network throughput for the user device comprises predicting the upcoming reduction of communication network throughput in an upcoming time window.
- a start time of the upcoming time window and/or a length of the upcoming time window is based on one or more of: an expected duration of the communication network throughput reduction, a maximum acceptable latency of the service, and a buffering setting of the service.
- configuring the scheduler to temporarily increase resource allocation for the user device comprises enforcing scheduling prioritization of the user device over one or more other user devices.
- predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future throughput is lower than a throughput of the latency sensitive service.
- configuring the scheduler to temporarily increase resource allocation for the user device is further in response to a traffic load that is less than a traffic load threshold.
- predicting the upcoming reduction of communication network throughput for the user device is based on one or more cell-specific prediction criteria and/or on one or more network-general prediction criteria.
- predicting the upcoming reduction of communication network throughput for the user device is based on one or more of: a capability category of the user device, a subscription type of the user device, an original equipment manufacturer (OEM) of the user device, a current time indication, a current and/or previous location of the user device, and a current and/or previous signal quality of the user device a current and/or previous communication network throughput of the user device, a current and/or previous timing advance (TA) of the user device, and a load of a serving cell and/or a target cell.
- OEM original equipment manufacturer
- predicting the upcoming reduction of communication network throughput for the user device is based on communication network throughput statistics.
- identifying that a service is currently associated with a user device, wherein the deviation between the latency requirement of the service and the internal latency performance of the communication network is bounded comprises one or more of: detecting that a service class identifier is indicative of the service, detecting that a bearer dedicated for low latency requirements is assigned for the service, and determining that a traffic pattern of the service matches a latency sensitive traffic pattern.
- the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network comprises one or more of: a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network not exceeding a bounding threshold, a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being in a same order of magnitude, a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being equal, and a required end-to-end round-trip-time of the service falling within a time range specified relative an internal round-trip-time of the communication network.
- the service has a maximum allowable latency which is lower than that of mobile broadband (MBB) services and/or higher than that of ultra-reliable low latency communication (URLLC) services.
- MBB mobile broadband
- URLLC ultra-reliable low latency communication
- the latency control comprises one or more of: decrease of latency variance associated with the communication network for the user device, decrease of a maximum latency associated with the communication network for the user device, decrease of a number of latency events associated with the communication network for the user device, that exceed a latency threshold value, and decrease of an average latency associated with the communication network for the user device.
- a second aspect is a computer program product comprising a non-transitory computer readable medium, having thereon a computer program comprising program instructions.
- the computer program is loadable into a data processing unit and configured to cause execution of the method according to the first aspect when the computer program is run by the data processing unit.
- a third aspect is an apparatus for latency control in a communication network.
- the apparatus comprises controlling circuitry configured to cause identification that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded, prediction of an upcoming reduction of communication network throughput for the user device, and configuration of a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
- a fourth aspect is a network node comprising the apparatus of the third aspect.
- any of the above aspects may additionally have features identical with or corresponding to any of the various features as explained above for any of the other aspects.
- An advantage of some embodiments is that alternative approaches for latency control are provided. Generally, the alternative approaches for latency control may be used instead of, or together with, other approaches for latency control, as suitable.
- An advantage of some embodiments is that the probability of occurrences with relatively large latency (e.g., latency spikes) may be reduced.
- An advantage of some embodiments is that the average latency may be reduced.
- An advantage of some embodiments is that the latency variance may be reduced.
- An advantage of some embodiments is that accommodation of latency variations of a communication network is enabled, under the constraint of latency requirements on communication between end points of a service.
- An advantage of some embodiments is that efficient resource allocation is enabled.
- An advantage of some embodiments is that robust resource allocation is enabled.
- Figure 1 is a schematic block diagram illustrates an example communication scenario according to some embodiments
- Figure 2 is a schematic block diagram illustrates an example communication scenario according to some embodiments
- Figure 3 is a flowchart illustrating example method steps according to some embodiments.
- Figure 4 is a schematic drawing illustrating example principles according to some embodiments.
- Figure 5 is a plot diagram illustrating example principles according to some embodiments
- Figure 6 is a schematic block diagram illustrating an example apparatus according to some embodiments.
- Figure 7 is a schematic drawing illustrating an example computer readable medium according to some embodiments.
- Figure 1 schematically illustrates a communication scenario 10 for demonstrating a type of situation where latency control may be challenging.
- the communication scenario 10 comprises two communication end points 30, 40 and a communication network 20.
- the communication network 20 is for carrying information (e.g., data and/or control information) from end point 30 to end point 40 as illustrated by 33, 23, and 43 and/or from end point 40 to end point 30 as illustrated by 44, 24, and 34.
- the end points 30, 40 may be any suitable communication end points.
- One example of a communication end point pair is an application client-server pair.
- latency of communication between the end points 30, 40 may be defined as one or more of: a time for transfer of information from end point 30 to end point 40 (possibly defined as a time between information entering a transmission buffer 31 associated with the end point 30 and the same information being dispatched from a reception buffer 41 associated with the end point 40), a time for transfer of information from end point 40 to end point 30 (possibly defined as a time between information entering a transmission buffer 42 associated with the end point 40 and the same information being dispatched from a reception buffer 32 associated with the end point 30), a time from issuing of first information at end point 30 (possibly defined as a time when the first information enters a transmission buffer 31 associated with the end point 30) to reception of second information at end point 30 (possibly defined as a time when the second information is dispatched from a reception buffer 32 associated with the end point 30) wherein the second information is issued by end point 40 in response to reception of the first information, and a time from issuing of first information at end point 40 (possibly
- latency of communication between end points may be characterized by one or more of: an average duration of end point to end point transfer, a minimum duration of end point to end point transfer, a maximum duration of end point to end point transfer, a variance of the duration of end point to end point transfer, and a probability that duration of end point to end point transfer exceeds a duration threshold.
- end point to end point transfer may refer to a one way transfer or to a round-trip- time (RTT).
- RTT round-trip- time
- the latency requirements on the communication between the end points may be defined according to any of the above, or other suitable, definitions and characterization.
- a specific service may require that the time from issuing of first information at end point 30 to reception of second information at end point 30 (wherein the second information is issued by end point 40 in response to reception of the first information) is below a maximum duration value and/or has a variance below a maximum variance value.
- the communication network 20 may be any suitable communication network.
- a communication network is any wireless communication network operating in accordance with a standard advocated by the third generation partnership project (3GPP); e.g., the universal mobile telecommunication system (UMTS), UMTS long term evolution (LTE), or a fifth generation (5G) system.
- the communication network may, for example, comprise a radio access network (RAN) and/or a core network (CN).
- RAN radio access network
- CN core network
- the communication network 20 typically has an internal latency performance, schematically illustrated by 25.
- the internal latency performance 25 of the communication network 20 determines (e.g., limits) how quickly information delivered to the communication network at 36 can be transferred through the communication network over 23 and provided at 46 and/or how quickly information delivered to the communication network at 47 can be transferred through the communication network over 24 and provided at 37.
- the internal latency performance 25 of the communication network 20 may be characterized in terms of the duration (delay) of the transfer over 23 and/or 24.
- the internal latency performance 25 of the communication network 20 may be characterized by one or more of: an average duration of transfer through the communication network, a minimum duration of transfer through the communication network, a maximum duration of transfer through the communication network, a variance of the duration of transfer through the communication network, and a probability that duration of transfer through the communication network exceeds a duration threshold.
- transfer through the communication network may refer to a one way transfer or to a round-trip-time (RTT).
- RTT round-trip-time
- the internal latency performance 25 of the communication network 20 may be caused by one or more of various (standardized or non-standardized) settings and limitations of the communication network.
- Some example settings and limitations of a communication network that inherently introduce latency include - but are not limited to - standardized time domain dimensions of communication resources (e.g., time duration of one or more units for communication), scheduling principles, protocols (e.g., retransmission protocols such as hybrid automatic repeat request - HARQ), and response requirements (e.g., for acknowledgement - ACK).
- the end- to-end communication scenario is unproblematic from a latency perspective.
- Such situations may, for example, occur when an average duration of transfer through the communication network is much lower than a required average duration of end point to end point transfer, when a maximum duration of transfer through the communication network is much lower than a required maximum (or average) duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is much lower than a required maximum variance of duration of end point to end point transfer.
- Such situations may, for example, occur when an average duration of transfer through the communication network is much higher than a required average duration of end point to end point transfer, when a minimum duration of transfer through the communication network is higher than a required minimum (or average) duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is higher than a required maximum variance of duration of end point to end point transfer.
- end-to-end communication scenario comprising a service which is latency sensitive (i.e., with latency requirements on the communication between the end points) in relation to the internal latency performance of the communication network (i.e., latency sensitive service).
- Such situations may, for example, occur when an average duration of transfer through the communication network is similar to a required average duration of end point to end point transfer, when a maximum duration of transfer through the communication network is similar to a required maximum duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is similar to a required maximum variance of duration of end point to end point transfer.
- a latency sensitive service associated with a user of a communication network may be defined as a service with latency requirements on the communication between the end points which are similar to the internal latency performance of the communication network.
- a latency sensitive service associated with a user of a communication network may be defined as a service with one or more latency requirement parameter value (e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.) for the communication between the end points being in the same order of magnitude as the value of a corresponding parameter of the internal latency performance of the communication network.
- one or more latency requirement parameter value e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.
- a latency sensitive service associated with a user of a communication network may be defined as a service with one or more latency requirement parameter value (e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.) for the communication between the end points deviating from the value of a corresponding, or otherwise relevant, parameter of the internal latency performance of the communication network by less than a threshold value.
- a latency requirement parameter value e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.
- a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of maximum duration of transfer for the communication between the end points which is lower than a maximum duration of transfer through the communication network.
- a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of average duration of transfer for the communication between the end points deviates from an average duration of transfer through the communication network by less than a threshold value.
- a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of variance of duration of transfer for the communication between the end points which is lower than a value based on a variance of duration of transfer through the communication network (e.g., lower than the variance of duration of transfer through the communication network, or lower than the variance of duration of transfer through the communication network plus or minus a bias value).
- the problems associated with latency sensitive services may be solved in the same way as situations where the end-to-end communication scenario is infeasible from a latency perspective, i.e., by application of a different communication network, or a specifically designed communication type within the communication network, to more easily accommodate the latency requirements on the communication between the end points.
- application of a communication network (or a specifically designed communication type within a communication network) which accommodates strict latency requirements on the communication between the end points is typically inefficient in terms of throughput and/or capacity. For example, increasing the amount of allocated communication resources is one approach that is helpful to accommodate strict latency requirements on the communication between the end points, but has a negative impact on overall throughput of the communication network.
- MBB mobile broadband
- MBB traffic is not particularly latency sensitive but can be very throughput demanding.
- URLLC ultra-reliable low latency communication
- URLLC may be particularly suitable for industrial applications.
- features are developed to support these new URLLC services and use cases. This exemplifies situations when the latency requirements on the communication between the end points are impossible to fully accommodate by the internal latency performance of the communication network, the end-to-end communication scenario is infeasible from a latency perspective, and a specifically designed communication type within the communication network is applied to accommodate the latency requirements on the communication between the end points.
- Embodiments presented herein are particularly applicable in situations which are neither of the above (MBB, voice, and URLLC), i.e., situations when the latency requirements on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network (referred to herein as latency sensitive services).
- a relatively high throughput is also required (which is typically not the case for services requiring a specifically designed communication type, e.g., URLLC).
- Some typical example services where embodiments may be particularly applicable - e.g., in the context of a 3GPP-based communication network - are gaming applications (gaming with or without rendering, and including multi-user gaming), augmented reality (AR), virtual reality (VR), and tele-operated vehicle control (e.g., driving).
- gaming applications gaming with or without rendering, and including multi-user gaming
- AR augmented reality
- VR virtual reality
- tele-operated vehicle control e.g., driving
- the latency through the radio network (RAN), the core network (CN), and all the way to the communication end points (e.g., application client and application server) needs to be considered in view of latency requirements on the communication between the end points.
- One approach to reduce the impact of CN latency and/or of latency between the communication network and the application server, is to apply an edge cloud deployment of the application.
- some example latency requirements include a maximum round-trip-time (RTT) for communication between end points (end-to-end, E2E, RTT) in any of the ranges 10-100 ms, 30-100 ms, 30-50 ms, and 80-100 ms, and/or some example throughput requirements include a throughput in the range 5-10 Mbps or there over; up to 400 Mbps (e.g., for VR streaming applications).
- RTT round-trip-time
- reliability of the communication e.g., measured as the probability of delivering traffic within a specified time duration, i.e., fulfilling the latency requirement.
- the reliability is tightly coupled with the latency requirements (without any latency requirement, the traffic can always be delivered, e.g., by using sufficiently many retransmissions).
- reliability is a relevant metric when a communication network is tuned for latency sensitive traffic.
- some communication networks are typically dimensioned and configured to provide services (e.g., for MBB traffic) with high throughput and relatively relaxed latency requirements.
- latency is typically considered in such communication networks (e.g., in relation to transmission control protocol, TCP, throughput and ramp-up times)
- predictable latency i.e., low latency variance
- TCP transmission control protocol
- TCP transmission control protocol
- ramp-up times predictable latency (i.e., low latency variance) is typically not required.
- the timing requirements in some human-machine interaction e.g., web-browsing and video streaming
- the timing requirements in some human-machine interaction e.g., web-browsing and video streaming
- buffers For latency sensitive services, however, extensive use of buffers is not possible due to the nature of the applications (e.g., quick reaction times required for gaming, fast control response required for vehicle tele-operation, etc.).
- a latency spike will have negative impact on the application experience/performance for latency sensitive services.
- some typical communication networks apply a handover mechanism where service by one cell is released before service setup towards a target cell is completed.
- This mechanism causes a brief communication interruption during the handover procedure.
- the interruption may, for example, be in the range 30-60 ms, or considerably longer (e.g., up to 100 ms, or up to several hundreds of ms, such as 200 ms, 500 ms, or 900 ms).
- these handover interruptions typically do not negatively affect the quality of the services, since the latency caused by the interrupt can be hidden with buffer management.
- the length of these handover interruptions may be in the same order of magnitude as the latency requirements of the service, and a handover can negatively affect the quality of the service.
- some approaches for latency control may comprise avoiding unnecessary handovers, at least for latency sensitive services, while performing handovers that are necessary to maintain connection between the communication network and a device operating at the communication end point.
- some approaches for latency control may comprise controlling the setting of one or more network configuration parameter, at least for latency sensitive services.
- network configuration parameters that potentially impact the latency of a communication network. Examples include: the scheduling request periodicity (for UE initiated transmissions), the coding and modulation selected for HARQ retransmissions, the maximum number of HARQ retransmissions, timer settings in medium access control (MAC), timer settings in radio link control (RLC).
- MAC medium access control
- RLC radio link control
- the setting of these configuration parameters typically have negligible impact on the user experience, since any un-acceptable latency jitter can be hidden with buffer management.
- the setting of these configuration parameters can have a negative effect on the user experience since they might affect the variation in latency and/or the maximum latency.
- some approaches for latency control may comprise controlling the setting of one or more network configuration parameter, at least for latency sensitive services, to provide latencies which are predictable (i.e., low variation in latency) and relatively low (i.e., low maximum latency).
- some approaches for latency control may comprise keeping latency predictable and relatively low (bounded and predictable), while (preferably) enabling maintenance of the communication connection through the communication network at a predictable throughput. Thereby, latency sensitive services can be satisfactorily supported in the communication network.
- some approaches for latency control may comprise, at least for latency sensitive services, informing the application about current and/or future conditions of the communication network that impact throughput and/or latency.
- Example adaptions by the application include lowering of a data rate (e.g., by lowering an encoder rate, such as a video encoding rate).
- Examples of already existing approaches for differentiating services in RAN include slicing, dedicated bearers, resource assignment differentiation, scheduling prioritization, etc.
- some legacy solutions allow an operator to assign more or less resources and/or set a priority for a bearer that transport a specific service type (e.g., voice over LTE, VoLTE).
- Such approaches may comprise assigning different service class identifiers (e.g., quality-of-service class identifier - QCI, 5QI, etc.) to different bearers based on which type of service is carried by the bearer.
- some legacy solutions allow an operator to enable different types of services based on which type of subscription is associated with a user device.
- FIG. 2 schematically illustrates a communication scenario with differentiated bearers.
- the user device 30a communicates with an application dwelling in the Internet 40A via the communication network 20', using a bearer 91.
- the user device 30b communicates with an application associated with a latency sensitive service via the communication network 20', using another bearer 92.
- the application dwells in a cloud edge 40B.
- the user device 30c communicates with an application dwelling in the Internet 40A using the bearer 91, as well as with an application associated with a latency sensitive service dwelling in the cloud edge 40B using the bearer 92.
- the different bearers 91, 92 may be differentiated to provide different latency characteristics as exemplified above, using any suitable latency approach (e.g., any of the alternative approaches for latency control described herein).
- Some embodiments of the approaches for latency control address the problems associated with latency sensitive services (i.e., services with a sensitive relationship between latency requirements on the communication between the end points and the internal latency performance of the communication network).
- a possible principle for alternative approaches for latency control is to improve the internal latency performance of the communication network (e.g., decreasing the maximum duration of transfer through the communication network, and/or decreasing the average duration of transfer through the communication network, and/or decreasing the variance of duration of transfer through the communication network, etc.). This may, for example, be achieved by avoiding unnecessary handovers and/or by controlling the setting of one or more network configuration parameter.
- Another possible principle for alternative approaches for latency control is to dynamically vary the utilization of the communication network by the service in view of the latency requirements on the communication between the end points. For example, temporarily (when the internal latency performance of the communication network is poor) lowering a communication rate that the service applies in the communication network may temporarily improve the internal latency performance of the communication network (e.g., due to less HARQ retransmissions, etc.) at the cost of reduced throughput. The latter may be mitigated by temporary buffer build-up (compare e.g., with 31 and 42 of Figure 1) within the boundaries set by latency requirements on the communication between the end points. This may, for example, be achieved by informing the application about current and/or future conditions of the communication network to allow the application to adjust its data rate.
- the above possible principles are used exclusively for latency sensitive services, or only for communication end points associated with a latency sensitive service.
- embodiments will be described where alternative approaches for latency control are provided. Some embodiments are particularly suitable for latency control in situations as that described in connection with Figure 1.
- some embodiments apply the principle to dynamically vary the utilization of the communication network by the service. This is a form of latency control is also termed latency management, and provides accommodation of latency variations of the communication network.
- Figure 3 illustrates an example method 100 according to some embodiments.
- the method is for latency control in a communication network (compare with communication networks 20 of Figure 1 and 20' of Figure 2).
- the latency control may take the form of accommodation of latency variations of the communication network.
- the communication network is typically configured to serve a plurality of users (compare with 30a, 30b, 30c of Figure 2).
- latency control may refer to one or more of: mitigation of latency spikes, reduction/decrease of latency variance/variation associated with the communication network for the user device, reduction/decrease of average latency associated with the communication network for the user device, reduction/decrease of the probability for (i.e., number - e.g., per time unit - of) latency events associated with the communication network for the user device that exceed a latency threshold value, reduction/decrease of a maximum latency associated with the communication network for the user device, or any other suitable change in latency behavior.
- a user device may, for example, comprise one of the communication end points 30, 40 of Figure 1, one of the user devices 30a, 30b, 30c of Figure 2, a user equipment (UE), a station (STA), or similar.
- UE user equipment
- STA station
- latency control may be for mitigation of latency variations and/or for providing predictable latency and/or for providing reliable communication.
- the latency control is performed under a throughput condition (e.g., that throughput should be kept at, or above, a minimum acceptable throughput).
- a throughput condition e.g., that throughput should be kept at, or above, a minimum acceptable throughput
- step 110 it is identified that a service is currently associated with a user device - the user device in turn being associated with (e.g., served by) the communication network - wherein the service has bounded deviation between a latency requirement of the service and an internal latency performance of the communication network.
- That a service is associated with a user device may, for example, include that part of a service application (e.g., an application client) is running on the user device.
- a service application e.g., an application client
- the service is the type of service elaborated on above - a service which relates to the communication network such that the latency requirements enforced by the service on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network.
- This type of service is also referred to herein as latency sensitive services.
- This type of relationship between the service and the communication network is referred to herein by specifying that the deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded.
- a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network falls within a bounding range.
- the bounding range may have any suitable value and/or may be dynamic or static.
- a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network are in a same order of magnitude.
- the same order of magnitude may be defined as not deviating more than a factor, e.g., 2, 5, or 10.
- one example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network are equal.
- one example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a required end-to-end round-trip-time of the service falls within a time range specified relative an internal round-trip-time of the communication network.
- the latency requirement parameter may, for example, refer to one or more of: latency variation, latency average, probability for latencies above a threshold value, maximum latency, or any other suitable latency metric.
- the internal latency performance parameter may, for example, be a corresponding parameter of the communication network (i.e., latency variation, latency average, probability for latencies above a threshold value, maximum latency, or any other suitable latency metric).
- the service might, for example, be a service which has a maximum allowable latency which is lower than that of mobile broadband (MBB) services and/or higher than that of ultra-reliable low latency communication (URLLC) services; or correspondingly for any other suitable latency requirement parameter.
- MBB mobile broadband
- URLLC ultra-reliable low latency communication
- the identification in step 110, that a service with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is currently associated with a user device may be performed in any suitable way. Some illustrative examples include detecting that a service class identifier is indicative of the service, detecting that a bearer dedicated for low latency requirements is assigned for the service, detecting that single network slice selection assistance information (S-NSSAI) is indicative of the service, and/or determining that a traffic pattern of the service matches a latency sensitive traffic pattern.
- S-NSSAI single network slice selection assistance information
- step 130 an upcoming reduction of communication network throughput for the user device is predicted and, in step 140, a scheduler is configured to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput. Thereby, latency control may be achieved.
- the scheduler being configured to temporarily increase resource allocation for the user device may be seen as applying dynamic resource allocation based on communication network throughput predictions.
- the scheduler may be configured to decrease resource allocation for the user device; i.e., declare the temporarily increased resource allocation for the user device to be completed.
- the temporary increased resource allocation may be on user device level (i.e., the increased resource allocation is applied for all services associated with the user device) or on service level (i.e., a user device associated with different services may have the increased resource allocation for some services and another - for example, a default - resource allocation for other services).
- the prediction and scheduler configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
- the scheduler configuration may be applied in a differentiated fashion, whereby user devices associated with a service of this type are subject to the scheduler configuration responsive to prediction of communication network throughput reduction while other user devices are not.
- the other user devices may apply default scheduler configuration.
- a communication network throughput reduction may refer to corresponding change in any suitable metric(s) related to communication network throughput.
- Some illustrative examples include a reduction in received signal quality, a reduction in received signal strength, a reduction in received signal strength indicator (RSSI), a reduction in reference signal received power (RSRP), a reduction in reference signal received quality (RSRQ), and a reduction in signal-to-interference ratio (SIR).
- the metric(s) related to communication network throughput is as measured by the user device on received signals.
- the change may be in terms of a relative metric (e.g., SIR) and/or in terms of an absolute metric (e.g., RSSI).
- predicting the upcoming reduction of communication network throughput for the user device may comprise predicting that a future throughput is lower than a throughput of the latency sensitive service.
- step 140 causes switching from a resource allocation to a second resource allocation, wherein the first resource allocation has less resources than the second resource allocation.
- the resource allocation increase of step 140 may be equal for different (e.g., all) types of reductions of communication network throughput of the user device, and/or may be different for different types of reductions of communication network throughput of the user device (e.g., depending on the expected duration of the reduced communication network throughput), and/or may be provided only for some types of reductions of communication network throughput of the user device.
- the type of reduction of communication network throughput may, for example, depend on one or more of: a current communication network throughput value at the time of prediction, a previous communication network throughput value at the time of prediction, a predicted communication network throughput value, an estimated time to the predicted communication network throughput value at the time of prediction, a length of the prediction time window, a difference between a current communication network throughput value at the time of prediction and a predicted communication network throughput value, etc.
- some embodiments suggest to cause a scheduler to increase resource allocation - in advance of an unfavorable situation - to avoid (or at least mitigate) congestion and thereby reduce latency peaks that are harmful for the end-to-end (E2E) experience.
- Configuring the scheduler to increase resource allocation may comprise and/or entail transmitting one or more messages to other network nodes and/or the user device, wherein the one or more messages are indicative of the increased resource allocation.
- step 130 may include predicting that a future communication network throughput value for the user device is lower than a rate of the service, or lower than a value depending on a rate of the service.
- step 130 may include predicting that a future communication network throughput value for the user device is lower than a current communication network throughput value for the user device.
- step 130 may include predicting that a future communication network throughput value for the user device is lower than a throughput threshold value (as illustrated by optional sub-step 131).
- the throughput threshold value may be static or dynamic.
- the throughput threshold value may depend on the rate of the service.
- the prediction of an upcoming reduction of communication network throughput in step 130 may comprise determining that a reduction of communication network throughput has already commenced for the user device and/or determining that a probability for future reduction of communication network throughput for the user device is above a reduction probability threshold value (as illustrated by optional sub-step 132).
- the reduction probability threshold value may be static or dynamic and may take any suitable value in the interval from zero to one.
- the prediction of an upcoming reduction of communication network throughput in step 130 may comprise predicting the upcoming reduction of communication network throughput in an upcoming time window.
- step 130 may comprise determining that a reduction of communication network throughput will, or may (with some probability), take place for the user device within the time window.
- the time window may be defined through one or more parameters, such as, for example, one or more of a start time, and end time, and a duration (i.e., length).
- the time window may be statically defined or may have varying parameters.
- the start time may be the time when the prediction is made, or a later point in time.
- the end time and/or the duration of the time window may be seen as representing an indication of how far into the future predictions need to be made.
- the duration of the time window may be based on how opportunistic/conservative the rate adaption should be (e.g., in terms of exploiting the maximum bitrate at any given time).
- the start time of the upcoming time window and/or the length of the upcoming time window may be based on one or more of: an expected duration of the communication network throughput reduction, a maximum acceptable latency of the latency sensitive service, and a buffering setting of the service.
- the duration of the time window may be increased.
- the duration of the time window may be set to accommodate a worst case reduced communication network throughput duration.
- the duration of the time window may be relatively large.
- the duration of the time window may be relatively large.
- the time window may be seen as a definition of how far into the future predictions are needed. If large latency margins are required/desired, the duration of the time window should typically be relatively long.
- the prediction of an upcoming reduction of communication network throughput in step 130 may be based on cell-specific prediction criteria and/or on network-general prediction criteria.
- the prediction of an upcoming reduction of communication network throughput in step 130 is based on communication network throughput statistics.
- Communication network throughput statistics e.g., statistics of communication network throughput reductions
- Each occurrence of communication network throughput reduction may be related to certain values of one or more parameters, and the statistics may comprise associations between the occurrences of communication network throughput reduction and corresponding parameter values. Based on such associations, the probability of communication network throughput reduction for a parameter setting of a current communication situation may be derived, and such probability may be used for the prediction step 130.
- Derivation of the probability of communication network throughput reduction for a parameter setting may be accomplished in any suitable way.
- a machine learning algorithm may use the statistics to derive the probability.
- the machine learning algorithm may apply self-training based on the statistics.
- the machine learning algorithm may be trained based on initial training data or other training data related to derivation of the probability of communication network throughput reduction.
- training entities may be implemented in a radio base station serving the user device or in another node (e.g., on a dedicated system outside the radio base station).
- Some example parameters that may be related to an occurrence of communication network throughput reduction include user device characteristics (e.g., capability category, UE category, subscription type, vendor, brand, original equipment manufacturer - OEM, etc.), a current time indication (e.g., indicating one or more of: time of day, day or week, date, season, etc.), current and/or previous location of the user device, current and/or previous signal quality of the user device, current and/or previous communication network throughput of the user device, current and/or previous timing advance (TA) of the user device, and load of serving cell and/or target cell.
- the statistics may indicate that a user device at a certain location has a corresponding probability of communication network throughput reduction within an upcoming time window.
- the statistics may indicate that a user device moving towards a certain location (e.g., from a specific direction) has a corresponding probability of communication network throughput reduction within an upcoming time window.
- the statistics may indicate that a user device moving with a certain speed has a corresponding probability of communication network throughput reduction within an upcoming time window.
- the statistics may indicate that a user device with a certain current communication network throughput drop has a corresponding probability of communication network throughput reduction within an upcoming time window.
- the statistics may indicate that a user device in a cell with certain load has a corresponding probability of communication network throughput reduction within an upcoming time window.
- Configuring the scheduler to temporarily increase resource allocation for the user device may, for example, comprise enforcing scheduling prioritization of the user device over one or more other user devices.
- Configuring the scheduler to temporarily increase resource allocation for the user device may be further in response to a traffic load that is less than a traffic load threshold.
- the resource allocation increase may be applied only in situations with relatively low traffic load.
- the scheduler may be configured to temporarily increase resource allocation for the user device when the cell load increases.
- a cell load increase may be seen as an indicator for communication network throughput reduction for the user device.
- Figure 3 has demonstrated that latency variations of a communication network may be accommodated by proactively causing a scheduler to increase resource allocation based on predicted reduction of communication network throughput.
- Various embodiments may be implemented entirely or partly in a network node such as a radio base station.
- various embodiments may be fully or partly implemented in a virtual environment and/or in a cloud environment. For example, training may be performed in a cloud environment and fed back to a radio base station that causes the scheduler configuration, and/or cloud-based services may be used for storing.
- Figure 4 schematically illustrates some example principles according to some embodiments.
- Part (a) of Figure 4 shows a user device (e.g., a UE) 400 associated with a latency sensitive service, a radio base station 410, and a communication network throughput reduction prediction algorithm 420.
- the communication network throughput reduction prediction algorithm 420 may be trained based on one or more optimization targets 430 (e.g., accuracy of throughput prediction during reduction).
- the base station 410 may be instructed 422 to cause the scheduler to increase resource allocation for the user device.
- Part (b) of Figure 4 illustrates example training of the communication network throughput reduction prediction algorithm 420 using reinforcement learning and one or more optimization targets 430.
- An interpreter 424 receives observations 421 (parameter setting; e.g., network conditions and/or performance metrics) for the user device.
- the interpreter 424 informs an agent 425 about a current state based on the observations and provides the agent with a corresponding reward for decisions taken by the agent based on the optimization target 430.
- the agent learns about patterns that improve the service quality for a current state (i.e., patterns of communication network throughput reduction occurrence in relation to observations).
- the agent may learn to recognize/predict that a communication network throughput reduction will happen in the near future. Consequently, the base station can be instructed 422 to cause the scheduler to increase resource allocation for the user device proactively.
- Part (c) of Figure 4 illustrates example training of the communication network throughput reduction prediction algorithm 420 using one or more optimization targets 430 and supervised learning of a neural network 427.
- observations 421 are input to the neural network 427, as well as to a classification module 426 and a training algorithm 428.
- the training algorithm 428 Based on the observations and a previous error, the training algorithm 428 provides weights to the neural network 427.
- the output of the neural network is compared, in an error calculator 429, with a desired output provided by the classification module 426, and the error is provided to the training algorithm.
- the training may be iterated until a convergence is achieved.
- observations 421 are input to the neural network 427, the output of which can be used to instruct 422 the base station to cause the scheduler to increase resource allocation for the user device proactively.
- Figure 5 illustrates example principles according to some embodiments.
- a scenario is depicted using throughput as a function of time.
- available bitrate is indicated as 503 (without application of embodiments) and 504 (with application of embodiments).
- duration 510 When a communication network throughput reduction is caused, the available bitrate decreases for the user device, as illustrated by duration 510.
- the bitrate required by the application is shown at 508.
- the scheduler can be configured to temporarily increase resource allocation, responsive to a prediction of the communication network throughput reduction (e.g., at time 501, predicting communication network throughput reduction within a time window 520). Thereby, more resources are available for the latency sensitive service at the start of the communication network throughput reduction 510.
- prediction of a communication network throughput reduction may comprise one or more of: prediction of a communication network throughput value 509, prediction of a communication network throughput being below a threshold value, prediction of decreasing communication network throughput, etc.
- the available bitrate 503 would drop substantially during the communication network throughput reduction duration 510.
- a consequence of this scenario is illustrated by application bitrate curve 505. Since the application bitrate 505 is considerably higher than the available bitrate 503 during the communication network throughput reduction duration 510, there will be queue build-up during the communication network throughput reduction, in turn resulting in latency peaks with potentially severe application impact. Eventually, this will be noticed by the application (e.g., due to lack of feedback from the application client and/or due to the queue build-up) and the application bitrate 505 will need to be lowered substantially to recover from the communication network throughput reduction.
- application of embodiments presented herein entails that the available bitrate 504 would drop less during the communication network throughput reduction duration 510, due to that the scheduler is proactively configured to temporarily increase resource allocation.
- a consequence of this scenario is illustrated by application bitrate curve 506. Since the application bitrate 506 is rather close to the available bitrate 504 during the communication network throughput reduction duration 510, there will be less queue build-up during the communication network throughput reduction, in turn resulting in better latency behavior. The application bitrate 506 may still need to be lowered to recover from the communication network throughput reduction, but the lowering is generally much less pronounced.
- the latency sensitive application may consume less resources on average even if embodiment are applied which increase resource allocation temporarily. This is due to decreased resource allocation (e.g., implemented by a more restrictive priority policy for the service) being applied in situations where the communication network throughput is good enough (e.g., when the available bitrate 504 is larger than the required application bitrate 508).
- This principle may be applied generally, as suitable.
- some embodiments may manifest themselves by the following rule for controlling the scheduler: increase resource allocation at time 501 when the predicted throughput 509 is lower than the required application bitrate 508, and apply a default resource allocation for latency sensitive services at time 501 when the predicted throughput 509 is higher than the required application bitrate 508.
- the default resource allocation for latency sensitive services may be lower than a default resource allocation for other types of services.
- Figure 6 schematically illustrates an example apparatus 610 for latency control in a communication network according to some embodiments.
- the apparatus 610 may, for example, be comprisable (e.g., comprised) in a network node (e.g., a radio access node such as a base station, or a central processing node).
- the apparatus 610 may be configured to cause execution of one or more of the method steps described herein (e.g., in connection with the method 100 of Figure 3).
- the apparatus comprises a controller (CNTR; e.g., controlling circuitry or a control module) 600.
- CNTR controlling circuitry or a control module
- the controller 600 is configured to cause identification that a service is currently associated with a user device associated with the communication network, with bounded deviation between a latency requirement of the service and an internal latency performance of the communication network (compare with step 110 of Figure 3).
- the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) an identifier (ID; e.g., identifying circuitry or an identification module) 601.
- ID e.g., identifying circuitry or an identification module
- the identifier may be configured to identify that a service is currently associated with a user device associated with the communication network, with bounded deviation between a latency requirement of the service and an internal latency performance of the communication network.
- the controller 600 is also configured to cause prediction of an upcoming reduction of communication network throughput for the user device (compare with step 130 of Figure 3).
- the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a predictor (PD; e.g., predicting circuitry or a prediction module) 602.
- the predictor may be configured to predict the upcoming reduction of communication network throughput for the user device using any suitable approach.
- the prediction of the upcoming reduction of communication network throughput for the user device may be based on communication network throughput statistics as explained above in connection to Figure 3.
- the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a statistics database.
- the database may be co-located with the controller as illustrated by the database (DB) 603 of Figure 6.
- the database may be located apart from the controller 600, but accessible by the controller, e.g., via an apparatus interface (I/O) 640.
- I/O apparatus interface
- the predictor may be configured to predict the upcoming reduction of communication network throughput for the user device using any communication network throughput statistics (e.g., based on machine-learning, or other analysis of communication network throughput statistics).
- any communication network throughput statistics e.g., based on machine-learning, or other analysis of communication network throughput statistics.
- the controller 600 is also configured to cause configuration of a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput (compare with step 140 of Figure 3).
- the scheduler may be co-located with the controller as illustrated by the scheduler (SCH) 604 of Figure 6. Alternatively or additionally, the scheduler may be located apart from the controller 600, but accessible by the controller, e.g., via an apparatus interface (I/O) 640.
- I/O apparatus interface
- Configuring the scheduler to increase resource allocation may comprise and/or entail transmission of one or more messages to the user device, wherein the one or more messages are indicative of the increase resource allocation.
- the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a transmitter (TX; e.g., transmitting circuitry or a transmission module) 630 and/or an apparatus interface (I/O; e.g., interface circuitry or an interface module) 640.
- TX transmitting circuitry or a transmission module
- I/O apparatus interface
- the transmitter may be configured to transmit message(s) to the user device (e.g., when the apparatus is comprisable in a radio access node).
- the interface may be configured to transfer message(s) to a radio access node for transmission to the user device (e.g., when the apparatus is comprisable in a central processing node).
- the controlling circuitry may be configured to cause prediction and configuration (of the scheduler) only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
- the described embodiments and their equivalents may be realized in software or hardware or a combination thereof.
- the embodiments may be performed by general purpose circuitry. Examples of general purpose circuitry include digital signal processors (DSP), central processing units (CPU), co-processor units, field programmable gate arrays (FPGA) and other programmable hardware.
- DSP digital signal processors
- CPU central processing units
- FPGA field programmable gate arrays
- the embodiments may be performed by specialized circuitry, such as application specific integrated circuits (ASIC).
- ASIC application specific integrated circuits
- the general purpose circuitry and/or the specialized circuitry may, for example, be associated with or comprised in an apparatus such as a network node.
- Embodiments may appear within an electronic apparatus (such as a network node) comprising arrangements, circuitry, and/or logic according to any of the embodiments described herein.
- an electronic apparatus such as a network node
- an electronic apparatus may be configured to perform methods according to any of the embodiments described herein.
- a computer program product comprises a tangible, or non tangible, computer readable medium such as, for example a universal serial bus (USB) memory, a plug-in card, an embedded drive or a read only memory (ROM).
- Figure 7 illustrates an example computer readable medium in the form of a compact disc (CD) ROM 700.
- the computer readable medium has stored thereon a computer program comprising program instructions.
- the computer program is loadable into a data processor (PROC; e.g., data processing circuitry or a data processing unit) 720, which may, for example, be comprised in a network node 710.
- PROC data processor
- the computer program may be stored in a memory (MEM) 730 associated with or comprised in the data processor.
- the computer program may, when loaded into and run by the data processor, cause execution of method steps according to, for example, any of the methods as illustrated in Figure 3 or otherwise described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Quality & Reliability (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
A method for accommodation of latency variation in a communication network is disclosed. The method comprises identifying that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded, predicting an upcoming reduction of communication network throughput for the user device, and configuring a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput. In some embodiments, the prediction and configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network. In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future communication network throughput value for the user device is lower than a throughput threshold value. In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises determining that a probability for upcoming reduction of communication network throughput for the user device is larger than a reduction probability threshold value. Corresponding apparatus, network node and computer program product are also disclosed.
Description
LATENCY CONTROL FOR A COMMUNICATION NETWORK
TECHNICAL FIELD
The present disclosure relates generally to the field of wireless communication. More particularly, it relates to latency control in wireless communication scenarios.
BACKGROUND
Different forms of latency control are generally applied in wireless communication scenarios. In some situations, existing approaches for latency control do not provide for desirable performance.
Therefore, there is a need for alternative approaches for latency control.
SUMMARY
It should be emphasized that the term "comprises/comprising" (replaceable by "includes/including") when used in this specification is taken to specify the presence of stated features, integers, steps, or components, but does not preclude the presence or addition of one or more other features, integers, steps, components, or groups thereof. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
Generally, when an arrangement is referred to herein, it is to be understood as a physical product; e.g., an apparatus. The physical product may comprise one or more parts, such as controlling circuitry in the form of one or more controllers, one or more processors, or the like.
It is an object of some embodiments to solve or mitigate, alleviate, or eliminate at least some of the above or other disadvantages.
A first aspect is a method for latency control in a communication network. The method comprises identifying that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the
service and an internal latency performance of the communication network is bounded, predicting an upcoming reduction of communication network throughput for the user device, and configuring a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
In some embodiments, the prediction and configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future communication network throughput value for the user device is lower than a throughput threshold value.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises determining that a probability for upcoming reduction of communication network throughput for the user device is larger than a reduction probability threshold value.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises predicting the upcoming reduction of communication network throughput in an upcoming time window.
In some embodiments, a start time of the upcoming time window and/or a length of the upcoming time window is based on one or more of: an expected duration of the communication network throughput reduction, a maximum acceptable latency of the service, and a buffering setting of the service.
In some embodiments, configuring the scheduler to temporarily increase resource allocation for the user device comprises enforcing scheduling prioritization of the user device over one or more other user devices.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future throughput is lower than a throughput of the latency sensitive service.
In some embodiments, configuring the scheduler to temporarily increase resource allocation for the user device is further in response to a traffic load that is less than a traffic load threshold.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device is based on one or more cell-specific prediction criteria and/or on one or more network-general prediction criteria.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device is based on one or more of: a capability category of the user device, a subscription type of the user device, an original equipment manufacturer (OEM) of the user device, a current time indication, a current and/or previous location of the user device, and a current and/or previous signal quality of the user device a current and/or previous communication network throughput of the user device, a current and/or previous timing advance (TA) of the user device, and a load of a serving cell and/or a target cell.
In some embodiments, predicting the upcoming reduction of communication network throughput for the user device is based on communication network throughput statistics.
In some embodiments, identifying that a service is currently associated with a user device, wherein the deviation between the latency requirement of the service and the internal latency performance of the communication network is bounded, comprises one or more of: detecting that a service class identifier is indicative of the service, detecting that a bearer dedicated for low latency requirements is assigned for the service, and determining that a traffic pattern of the service matches a latency sensitive traffic pattern.
In some embodiments, the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network comprises one or more of: a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network not exceeding a bounding threshold, a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being in a same order of magnitude, a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being equal, and a required
end-to-end round-trip-time of the service falling within a time range specified relative an internal round-trip-time of the communication network.
In some embodiments, the service has a maximum allowable latency which is lower than that of mobile broadband (MBB) services and/or higher than that of ultra-reliable low latency communication (URLLC) services.
In some embodiments, the latency control comprises one or more of: decrease of latency variance associated with the communication network for the user device, decrease of a maximum latency associated with the communication network for the user device, decrease of a number of latency events associated with the communication network for the user device, that exceed a latency threshold value, and decrease of an average latency associated with the communication network for the user device.
A second aspect is a computer program product comprising a non-transitory computer readable medium, having thereon a computer program comprising program instructions. The computer program is loadable into a data processing unit and configured to cause execution of the method according to the first aspect when the computer program is run by the data processing unit.
A third aspect is an apparatus for latency control in a communication network. The apparatus comprises controlling circuitry configured to cause identification that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded, prediction of an upcoming reduction of communication network throughput for the user device, and configuration of a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
A fourth aspect is a network node comprising the apparatus of the third aspect.
In some embodiments, any of the above aspects may additionally have features identical with or corresponding to any of the various features as explained above for any of the other aspects.
An advantage of some embodiments is that alternative approaches for latency control are provided. Generally, the alternative approaches for latency control may be used instead of, or together with, other approaches for latency control, as suitable.
An advantage of some embodiments is that the probability of occurrences with relatively large latency (e.g., latency spikes) may be reduced.
An advantage of some embodiments is that the average latency may be reduced.
An advantage of some embodiments is that the latency variance may be reduced.
An advantage of some embodiments is that accommodation of latency variations of a communication network is enabled, under the constraint of latency requirements on communication between end points of a service.
An advantage of some embodiments is that efficient resource allocation is enabled.
An advantage of some embodiments is that robust resource allocation is enabled.
BRIEF DESCRIPTION OF THE DRAWINGS
Further objects, features and advantages will appear from the following detailed description of embodiments, with reference being made to the accompanying drawings. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the example embodiments.
Figure 1 is a schematic block diagram illustrates an example communication scenario according to some embodiments;
Figure 2 is a schematic block diagram illustrates an example communication scenario according to some embodiments;
Figure 3 is a flowchart illustrating example method steps according to some embodiments;
Figure 4 is a schematic drawing illustrating example principles according to some embodiments;
Figure 5 is a plot diagram illustrating example principles according to some embodiments;
Figure 6 is a schematic block diagram illustrating an example apparatus according to some embodiments; and
Figure 7 is a schematic drawing illustrating an example computer readable medium according to some embodiments.
DETAILED DESCRIPTION
As already mentioned above, it should be emphasized that the term "comprises/comprising" (replaceable by "includes/including") when used in this specification is taken to specify the presence of stated features, integers, steps, or components, but does not preclude the presence or addition of one or more other features, integers, steps, components, or groups thereof. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
Embodiments of the present disclosure will be described and exemplified more fully hereinafter with reference to the accompanying drawings. The solutions disclosed herein can, however, be realized in many different forms and should not be construed as being limited to the embodiments set forth herein.
As mentioned before, different forms of latency control are generally applied in wireless communication scenarios. Figure 1 schematically illustrates a communication scenario 10 for demonstrating a type of situation where latency control may be challenging.
The communication scenario 10 comprises two communication end points 30, 40 and a communication network 20. The communication network 20 is for carrying information (e.g., data and/or control information) from end point 30 to end point 40 as illustrated by 33, 23, and 43 and/or from end point 40 to end point 30 as illustrated by 44, 24, and 34.
The end points 30, 40 may be any suitable communication end points. One example of a communication end point pair is an application client-server pair.
Depending on the type of service the end points 30, 40 are engaged in, there may be different latency requirements on the communication between the end points 30, 40.
Generally, latency of communication between the end points 30, 40 may be defined as one or more of: a time for transfer of information from end point 30 to end point 40 (possibly defined
as a time between information entering a transmission buffer 31 associated with the end point 30 and the same information being dispatched from a reception buffer 41 associated with the end point 40), a time for transfer of information from end point 40 to end point 30 (possibly defined as a time between information entering a transmission buffer 42 associated with the end point 40 and the same information being dispatched from a reception buffer 32 associated with the end point 30), a time from issuing of first information at end point 30 (possibly defined as a time when the first information enters a transmission buffer 31 associated with the end point 30) to reception of second information at end point 30 (possibly defined as a time when the second information is dispatched from a reception buffer 32 associated with the end point 30) wherein the second information is issued by end point 40 in response to reception of the first information, and a time from issuing of first information at end point 40 (possibly defined as a time when the first information enters a transmission buffer 42 associated with the end point 40) to reception of second information at end point 40 (possibly defined as a time when the second information is dispatched from a reception buffer 41 associated with the end point 40) wherein the second information is issued by end point 30 in response to reception of the first information.
Alternatively or additionally, and generally, latency of communication between end points may be characterized by one or more of: an average duration of end point to end point transfer, a minimum duration of end point to end point transfer, a maximum duration of end point to end point transfer, a variance of the duration of end point to end point transfer, and a probability that duration of end point to end point transfer exceeds a duration threshold.
Generally, end point to end point transfer may refer to a one way transfer or to a round-trip- time (RTT).
The latency requirements on the communication between the end points may be defined according to any of the above, or other suitable, definitions and characterization. For example, a specific service may require that the time from issuing of first information at end point 30 to reception of second information at end point 30 (wherein the second information is issued by end point 40 in response to reception of the first information) is below a maximum duration value and/or has a variance below a maximum variance value.
The communication network 20 may be any suitable communication network. One example of a communication network is any wireless communication network operating in accordance with a standard advocated by the third generation partnership project (3GPP); e.g., the universal mobile telecommunication system (UMTS), UMTS long term evolution (LTE), or a fifth generation (5G) system. The communication network may, for example, comprise a radio access network (RAN) and/or a core network (CN).
The communication network 20 typically has an internal latency performance, schematically illustrated by 25.
The internal latency performance 25 of the communication network 20 determines (e.g., limits) how quickly information delivered to the communication network at 36 can be transferred through the communication network over 23 and provided at 46 and/or how quickly information delivered to the communication network at 47 can be transferred through the communication network over 24 and provided at 37.
The internal latency performance 25 of the communication network 20 may be characterized in terms of the duration (delay) of the transfer over 23 and/or 24. For example, the internal latency performance 25 of the communication network 20 may be characterized by one or more of: an average duration of transfer through the communication network, a minimum duration of transfer through the communication network, a maximum duration of transfer through the communication network, a variance of the duration of transfer through the communication network, and a probability that duration of transfer through the communication network exceeds a duration threshold.
Generally, transfer through the communication network may refer to a one way transfer or to a round-trip-time (RTT).
The internal latency performance 25 of the communication network 20 may be caused by one or more of various (standardized or non-standardized) settings and limitations of the communication network. Some example settings and limitations of a communication network that inherently introduce latency include - but are not limited to - standardized time domain dimensions of communication resources (e.g., time duration of one or more units for communication), scheduling principles, protocols (e.g., retransmission protocols such as hybrid
automatic repeat request - HARQ), and response requirements (e.g., for acknowledgement - ACK).
When the latency requirements on the communication between the end points are easily accommodated by the internal latency performance of the communication network, the end- to-end communication scenario is unproblematic from a latency perspective.
Such situations may, for example, occur when an average duration of transfer through the communication network is much lower than a required average duration of end point to end point transfer, when a maximum duration of transfer through the communication network is much lower than a required maximum (or average) duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is much lower than a required maximum variance of duration of end point to end point transfer.
When the latency requirements on the communication between the end points are impossible to fully accommodate by the internal latency performance of the communication network, the end-to-end communication scenario is infeasible from a latency perspective.
Such situations may, for example, occur when an average duration of transfer through the communication network is much higher than a required average duration of end point to end point transfer, when a minimum duration of transfer through the communication network is higher than a required minimum (or average) duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is higher than a required maximum variance of duration of end point to end point transfer.
These problems may be solved by application of a different communication network, or a specifically designed communication type within the communication network, to accommodate the latency requirements on the communication between the end points. Embodiments presented herein are particularly applicable in situations which are neither of the above, i.e., situations when the latency requirements on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network. Then, the end-to-end communication scenario is feasible, but problematic (e.g., presenting challenges), from a latency perspective. This may be
seen as the end-to-end communication scenario comprising a service which is latency sensitive (i.e., with latency requirements on the communication between the end points) in relation to the internal latency performance of the communication network (i.e., latency sensitive service).
Such situations may, for example, occur when an average duration of transfer through the communication network is similar to a required average duration of end point to end point transfer, when a maximum duration of transfer through the communication network is similar to a required maximum duration of end point to end point transfer, and/or when a variance of duration of transfer through the communication network is similar to a required maximum variance of duration of end point to end point transfer.
Generally, a latency sensitive service associated with a user of a communication network may be defined as a service with latency requirements on the communication between the end points which are similar to the internal latency performance of the communication network.
For example, a latency sensitive service associated with a user of a communication network may be defined as a service with one or more latency requirement parameter value (e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.) for the communication between the end points being in the same order of magnitude as the value of a corresponding parameter of the internal latency performance of the communication network.
Alternatively or additionally, a latency sensitive service associated with a user of a communication network may be defined as a service with one or more latency requirement parameter value (e.g., average duration of transfer, maximum duration of transfer, variance of transfer duration, etc.) for the communication between the end points deviating from the value of a corresponding, or otherwise relevant, parameter of the internal latency performance of the communication network by less than a threshold value.
Alternatively or additionally, a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of maximum duration of transfer for the communication between the end points which is lower than a maximum duration of transfer through the communication network.
Alternatively or additionally, a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of average duration of transfer for the communication between the end points deviates from an average duration of transfer through the communication network by less than a threshold value.
Alternatively or additionally, a latency sensitive service associated with a user of a communication network may be defined as a service with a requirement of variance of duration of transfer for the communication between the end points which is lower than a value based on a variance of duration of transfer through the communication network (e.g., lower than the variance of duration of transfer through the communication network, or lower than the variance of duration of transfer through the communication network plus or minus a bias value).
The problems associated with latency sensitive services may be solved in the same way as situations where the end-to-end communication scenario is infeasible from a latency perspective, i.e., by application of a different communication network, or a specifically designed communication type within the communication network, to more easily accommodate the latency requirements on the communication between the end points. However, application of a communication network (or a specifically designed communication type within a communication network) which accommodates strict latency requirements on the communication between the end points is typically inefficient in terms of throughput and/or capacity. For example, increasing the amount of allocated communication resources is one approach that is helpful to accommodate strict latency requirements on the communication between the end points, but has a negative impact on overall throughput of the communication network.
Therefore, there is a need for alternative approaches for latency control, which preferably address the problems associated with latency sensitive services (i.e., services with a sensitive relationship between latency requirements on the communication between the end points and the internal latency performance of the communication network).
A more detailed context will now be described, in relation to which embodiments may be particularly applicable. It should be noted that the following context is merely an illustrative example and not to be construed as limiting.
Some typical existing wireless communication networks (e.g., 3GPP-based networks supporting fourth generation, 4G, and earlier releases of the communication standard) are mainly optimized for mobile broadband (MBB) services and voice services. Generally, MBB traffic is not particularly latency sensitive but can be very throughput demanding. For example, for streaming services latency is typically handled by using large buffers which will efficiently hide latency jitter caused by latency events in the communication network, and thereby provide good end user experience. This exemplifies situations when the latency requirements on the communication between the end points are easily accommodated by the internal latency performance of the communication network, and the end-to-end communication scenario is unproblematic from a latency perspective.
In later releases of 4G, and especially in 5G, services of other types than MBB and voice have come into focus. One example is ultra-reliable low latency communication (URLLC) services. URLLC may be particularly suitable for industrial applications. Within 3GPP standardization, features are developed to support these new URLLC services and use cases. This exemplifies situations when the latency requirements on the communication between the end points are impossible to fully accommodate by the internal latency performance of the communication network, the end-to-end communication scenario is infeasible from a latency perspective, and a specifically designed communication type within the communication network is applied to accommodate the latency requirements on the communication between the end points.
Embodiments presented herein are particularly applicable in situations which are neither of the above (MBB, voice, and URLLC), i.e., situations when the latency requirements on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network (referred to herein as latency sensitive services). In some embodiments, a relatively high throughput is also required (which is typically not the case for services requiring a specifically designed communication type, e.g., URLLC).
Some typical example services where embodiments may be particularly applicable - e.g., in the context of a 3GPP-based communication network - are gaming applications (gaming with or without rendering, and including multi-user gaming), augmented reality (AR), virtual reality (VR), and tele-operated vehicle control (e.g., driving).
Generally, the latency through the radio network (RAN), the core network (CN), and all the way to the communication end points (e.g., application client and application server) needs to be considered in view of latency requirements on the communication between the end points. One approach to reduce the impact of CN latency and/or of latency between the communication network and the application server, is to apply an edge cloud deployment of the application.
For situations when the latency requirements on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network, some example latency requirements include a maximum round-trip-time (RTT) for communication between end points (end-to-end, E2E, RTT) in any of the ranges 10-100 ms, 30-100 ms, 30-50 ms, and 80-100 ms, and/or some example throughput requirements include a throughput in the range 5-10 Mbps or there over; up to 400 Mbps (e.g., for VR streaming applications).
For situations when the latency requirements on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network, it may be further beneficial to consider reliability of the communication (e.g., measured as the probability of delivering traffic within a specified time duration, i.e., fulfilling the latency requirement). The reliability is tightly coupled with the latency requirements (without any latency requirement, the traffic can always be delivered, e.g., by using sufficiently many retransmissions). Thus, reliability is a relevant metric when a communication network is tuned for latency sensitive traffic.
Thus, some communication networks are typically dimensioned and configured to provide services (e.g., for MBB traffic) with high throughput and relatively relaxed latency requirements. Although latency is typically considered in such communication networks (e.g., in relation to transmission control protocol, TCP, throughput and ramp-up times), predictable latency (i.e., low latency variance) is typically not required. One explanation to the latter is that the timing requirements in some human-machine interaction (e.g., web-browsing and video streaming) is quite relaxed and rather large latency variations can be hidden with buffers.
For latency sensitive services, however, extensive use of buffers is not possible due to the nature of the applications (e.g., quick reaction times required for gaming, fast control response required for vehicle tele-operation, etc.). Typically, a latency spike will have negative impact on the application experience/performance for latency sensitive services. Some example events in a communication network that may cause latency spikes include handovers, slow fading dips, and fast fading dips.
In association with some typical communication networks (e.g., SGPP-based networks for 4G and 5G), efforts are made to reduce overall latency (e.g., reducing the average latency). This, however, does not exclude a relatively large maximum latency and/or a relatively large latency variation (which may result in latency spikes, for example).
Regarding handover, some typical communication networks (e.g., SGPP-based networks for 4G and 5G) apply a handover mechanism where service by one cell is released before service setup towards a target cell is completed. This mechanism causes a brief communication interruption during the handover procedure. The interruption may, for example, be in the range 30-60 ms, or considerably longer (e.g., up to 100 ms, or up to several hundreds of ms, such as 200 ms, 500 ms, or 900 ms).
For MBB services and voice services, these handover interruptions typically do not negatively affect the quality of the services, since the latency caused by the interrupt can be hidden with buffer management. For latency sensitive services, however, the length of these handover interruptions may be in the same order of magnitude as the latency requirements of the service, and a handover can negatively affect the quality of the service.
Hence, some approaches for latency control may comprise avoiding unnecessary handovers, at least for latency sensitive services, while performing handovers that are necessary to maintain connection between the communication network and a device operating at the communication end point.
Alternatively or additionally, some approaches for latency control may comprise controlling the setting of one or more network configuration parameter, at least for latency sensitive services.
Generally, there are many network configuration parameters that potentially impact the latency of a communication network. Examples include: the scheduling request periodicity (for UE initiated transmissions), the coding and modulation selected for HARQ retransmissions, the maximum number of HARQ retransmissions, timer settings in medium access control (MAC), timer settings in radio link control (RLC).
For MBB services and voice services, the setting of these configuration parameters typically have negligible impact on the user experience, since any un-acceptable latency jitter can be hidden with buffer management. For latency sensitive services, however, the setting of these configuration parameters can have a negative effect on the user experience since they might affect the variation in latency and/or the maximum latency.
Hence, some approaches for latency control may comprise controlling the setting of one or more network configuration parameter, at least for latency sensitive services, to provide latencies which are predictable (i.e., low variation in latency) and relatively low (i.e., low maximum latency). Generally, some approaches for latency control may comprise keeping latency predictable and relatively low (bounded and predictable), while (preferably) enabling maintenance of the communication connection through the communication network at a predictable throughput. Thereby, latency sensitive services can be satisfactorily supported in the communication network. Alternatively or additionally, some approaches for latency control may comprise, at least for latency sensitive services, informing the application about current and/or future conditions of the communication network that impact throughput and/or latency.
For example, when the application receives information that the throughput of the communication network throughput is decreasing, or predicted to decrease in a near future, the application can adapt to this situation. Example adaptions by the application include lowering of a data rate (e.g., by lowering an encoder rate, such as a video encoding rate).
Some communication networks already have approaches for differentiating services in RAN, which may be used for differentiation also in relation to latency sensitive services as defined
herein. In some embodiments, such approaches may be combined with the alternative approaches for latency control presented herein.
Examples of already existing approaches for differentiating services in RAN include slicing, dedicated bearers, resource assignment differentiation, scheduling prioritization, etc. For example, some legacy solutions allow an operator to assign more or less resources and/or set a priority for a bearer that transport a specific service type (e.g., voice over LTE, VoLTE). Such approaches may comprise assigning different service class identifiers (e.g., quality-of-service class identifier - QCI, 5QI, etc.) to different bearers based on which type of service is carried by the bearer. Alternatively or additionally, some legacy solutions allow an operator to enable different types of services based on which type of subscription is associated with a user device.
Figure 2 schematically illustrates a communication scenario with differentiated bearers. The user device 30a communicates with an application dwelling in the Internet 40A via the communication network 20', using a bearer 91. The user device 30b communicates with an application associated with a latency sensitive service via the communication network 20', using another bearer 92. To reduce end-to-end latency, the application dwells in a cloud edge 40B. The user device 30c communicates with an application dwelling in the Internet 40A using the bearer 91, as well as with an application associated with a latency sensitive service dwelling in the cloud edge 40B using the bearer 92.
The different bearers 91, 92 may be differentiated to provide different latency characteristics as exemplified above, using any suitable latency approach (e.g., any of the alternative approaches for latency control described herein).
As mentioned above, alternative approaches for latency control are provided by this disclosure. Some embodiments of the approaches for latency control address the problems associated with latency sensitive services (i.e., services with a sensitive relationship between latency requirements on the communication between the end points and the internal latency performance of the communication network).
A possible principle for alternative approaches for latency control is to improve the internal latency performance of the communication network (e.g., decreasing the maximum duration
of transfer through the communication network, and/or decreasing the average duration of transfer through the communication network, and/or decreasing the variance of duration of transfer through the communication network, etc.). This may, for example, be achieved by avoiding unnecessary handovers and/or by controlling the setting of one or more network configuration parameter.
Another possible principle for alternative approaches for latency control is to dynamically vary the utilization of the communication network by the service in view of the latency requirements on the communication between the end points. For example, temporarily (when the internal latency performance of the communication network is poor) lowering a communication rate that the service applies in the communication network may temporarily improve the internal latency performance of the communication network (e.g., due to less HARQ retransmissions, etc.) at the cost of reduced throughput. The latter may be mitigated by temporary buffer build-up (compare e.g., with 31 and 42 of Figure 1) within the boundaries set by latency requirements on the communication between the end points. This may, for example, be achieved by informing the application about current and/or future conditions of the communication network to allow the application to adjust its data rate.
The above possible principles may be used alone or in combination.
In some embodiments, the above possible principles are used exclusively for latency sensitive services, or only for communication end points associated with a latency sensitive service. In the following, embodiments will be described where alternative approaches for latency control are provided. Some embodiments are particularly suitable for latency control in situations as that described in connection with Figure 1. Furthermore, some embodiments apply the principle to dynamically vary the utilization of the communication network by the service. This is a form of latency control is also termed latency management, and provides accommodation of latency variations of the communication network.
Figure 3 illustrates an example method 100 according to some embodiments. The method is for latency control in a communication network (compare with communication networks 20 of Figure 1 and 20' of Figure 2). The latency control may take the form of accommodation of
latency variations of the communication network. The communication network is typically configured to serve a plurality of users (compare with 30a, 30b, 30c of Figure 2).
Generally, latency control may refer to one or more of: mitigation of latency spikes, reduction/decrease of latency variance/variation associated with the communication network for the user device, reduction/decrease of average latency associated with the communication network for the user device, reduction/decrease of the probability for (i.e., number - e.g., per time unit - of) latency events associated with the communication network for the user device that exceed a latency threshold value, reduction/decrease of a maximum latency associated with the communication network for the user device, or any other suitable change in latency behavior.
A user device may, for example, comprise one of the communication end points 30, 40 of Figure 1, one of the user devices 30a, 30b, 30c of Figure 2, a user equipment (UE), a station (STA), or similar.
Also generally, latency control may be for mitigation of latency variations and/or for providing predictable latency and/or for providing reliable communication.
In typical embodiments, the latency control is performed under a throughput condition (e.g., that throughput should be kept at, or above, a minimum acceptable throughput).
In step 110, it is identified that a service is currently associated with a user device - the user device in turn being associated with (e.g., served by) the communication network - wherein the service has bounded deviation between a latency requirement of the service and an internal latency performance of the communication network.
That a service is associated with a user device may, for example, include that part of a service application (e.g., an application client) is running on the user device.
The service is the type of service elaborated on above - a service which relates to the communication network such that the latency requirements enforced by the service on the communication between the end points are not impossible, but not easy either, to accommodate by the internal latency performance of the communication network. This type of service is also referred to herein as latency sensitive services. This type of relationship between the service and the communication network is referred to herein by specifying that
the deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded.
One example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network falls within a bounding range. The bounding range may have any suitable value and/or may be dynamic or static.
Alternatively or additionally, one example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network are in a same order of magnitude. For example, the same order of magnitude may be defined as not deviating more than a factor, e.g., 2, 5, or 10.
Alternatively or additionally, one example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network are equal.
Alternatively or additionally, one example that substantiates the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is that a required end-to-end round-trip-time of the service falls within a time range specified relative an internal round-trip-time of the communication network.
The latency requirement parameter may, for example, refer to one or more of: latency variation, latency average, probability for latencies above a threshold value, maximum latency, or any other suitable latency metric. The internal latency performance parameter may, for example, be a corresponding parameter of the communication network (i.e., latency variation, latency average, probability for latencies above a threshold value, maximum latency, or any other suitable latency metric).
For BGPP-based communication networks, the service might, for example, be a service which has a maximum allowable latency which is lower than that of mobile broadband (MBB) services and/or higher than that of ultra-reliable low latency communication (URLLC) services; or correspondingly for any other suitable latency requirement parameter.
The identification in step 110, that a service with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network is currently associated with a user device, may be performed in any suitable way. Some illustrative examples include detecting that a service class identifier is indicative of the service, detecting that a bearer dedicated for low latency requirements is assigned for the service, detecting that single network slice selection assistance information (S-NSSAI) is indicative of the service, and/or determining that a traffic pattern of the service matches a latency sensitive traffic pattern.
In step 130, an upcoming reduction of communication network throughput for the user device is predicted and, in step 140, a scheduler is configured to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput. Thereby, latency control may be achieved.
The scheduler being configured to temporarily increase resource allocation for the user device may be seen as applying dynamic resource allocation based on communication network throughput predictions. When the communication network throughput increases again after a reduction of communication network throughput, the scheduler may be configured to decrease resource allocation for the user device; i.e., declare the temporarily increased resource allocation for the user device to be completed.
Generally, the temporary increased resource allocation may be on user device level (i.e., the increased resource allocation is applied for all services associated with the user device) or on service level (i.e., a user device associated with different services may have the increased resource allocation for some services and another - for example, a default - resource allocation for other services).
In some embodiments, the prediction and scheduler configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of
the service and the internal latency performance of the communication network. Thus, the scheduler configuration may be applied in a differentiated fashion, whereby user devices associated with a service of this type are subject to the scheduler configuration responsive to prediction of communication network throughput reduction while other user devices are not. Hence, the other user devices may apply default scheduler configuration.
Generally, a communication network throughput reduction may refer to corresponding change in any suitable metric(s) related to communication network throughput. Some illustrative examples include a reduction in received signal quality, a reduction in received signal strength, a reduction in received signal strength indicator (RSSI), a reduction in reference signal received power (RSRP), a reduction in reference signal received quality (RSRQ), and a reduction in signal-to-interference ratio (SIR). Typically, the metric(s) related to communication network throughput is as measured by the user device on received signals. The change may be in terms of a relative metric (e.g., SIR) and/or in terms of an absolute metric (e.g., RSSI). For example, predicting the upcoming reduction of communication network throughput for the user device may comprise predicting that a future throughput is lower than a throughput of the latency sensitive service.
In a typical example, step 140 causes switching from a resource allocation to a second resource allocation, wherein the first resource allocation has less resources than the second resource allocation.
Generally, the resource allocation increase of step 140 may be equal for different (e.g., all) types of reductions of communication network throughput of the user device, and/or may be different for different types of reductions of communication network throughput of the user device (e.g., depending on the expected duration of the reduced communication network throughput), and/or may be provided only for some types of reductions of communication network throughput of the user device. The type of reduction of communication network throughput may, for example, depend on one or more of: a current communication network throughput value at the time of prediction, a previous communication network throughput value at the time of prediction, a predicted communication network throughput value, an estimated time to the predicted communication network throughput value at the time of prediction, a length of the prediction time window, a difference between a current
communication network throughput value at the time of prediction and a predicted communication network throughput value, etc.
Thus, some embodiments suggest to cause a scheduler to increase resource allocation - in advance of an unfavorable situation - to avoid (or at least mitigate) congestion and thereby reduce latency peaks that are harmful for the end-to-end (E2E) experience.
Configuring the scheduler to increase resource allocation may comprise and/or entail transmitting one or more messages to other network nodes and/or the user device, wherein the one or more messages are indicative of the increased resource allocation.
In some embodiments, step 130 may include predicting that a future communication network throughput value for the user device is lower than a rate of the service, or lower than a value depending on a rate of the service.
In some embodiments, step 130 may include predicting that a future communication network throughput value for the user device is lower than a current communication network throughput value for the user device. For example, step 130 may include predicting that a future communication network throughput value for the user device is lower than a throughput threshold value (as illustrated by optional sub-step 131). The throughput threshold value may be static or dynamic. For example, the throughput threshold value may depend on the rate of the service.
The prediction of an upcoming reduction of communication network throughput in step 130 may comprise determining that a reduction of communication network throughput has already commenced for the user device and/or determining that a probability for future reduction of communication network throughput for the user device is above a reduction probability threshold value (as illustrated by optional sub-step 132). The reduction probability threshold value may be static or dynamic and may take any suitable value in the interval from zero to one.
Alternatively or additionally, the prediction of an upcoming reduction of communication network throughput in step 130 may comprise predicting the upcoming reduction of communication network throughput in an upcoming time window. Thus, step 130 may
comprise determining that a reduction of communication network throughput will, or may (with some probability), take place for the user device within the time window.
The time window may be defined through one or more parameters, such as, for example, one or more of a start time, and end time, and a duration (i.e., length). The time window may be statically defined or may have varying parameters. Generally, the start time may be the time when the prediction is made, or a later point in time. Also generally, the end time and/or the duration of the time window may be seen as representing an indication of how far into the future predictions need to be made.
Alternatively or additionally, the duration of the time window may be based on how opportunistic/conservative the rate adaption should be (e.g., in terms of exploiting the maximum bitrate at any given time).
The start time of the upcoming time window and/or the length of the upcoming time window may be based on one or more of: an expected duration of the communication network throughput reduction, a maximum acceptable latency of the latency sensitive service, and a buffering setting of the service.
For example, when the duration of the reduced communication network throughput is expected to be relatively long, the duration of the time window may be increased. Alternatively, the duration of the time window may be set to accommodate a worst case reduced communication network throughput duration.
Alternatively or additionally, when the maximum acceptable latency of the latency sensitive service is relatively small, the duration of the time window may be relatively large.
Alternatively or additionally, when the buffering setting of the latency sensitive service is relatively sensitive to buffer build-up, the duration of the time window may be relatively large.
The time window may be seen as a definition of how far into the future predictions are needed. If large latency margins are required/desired, the duration of the time window should typically be relatively long.
Generally, the prediction of an upcoming reduction of communication network throughput in step 130 may be based on cell-specific prediction criteria and/or on network-general prediction criteria.
In some embodiments, the prediction of an upcoming reduction of communication network throughput in step 130 is based on communication network throughput statistics. Communication network throughput statistics (e.g., statistics of communication network throughput reductions) may be collected, maintained, and updated over time. Each occurrence of communication network throughput reduction may be related to certain values of one or more parameters, and the statistics may comprise associations between the occurrences of communication network throughput reduction and corresponding parameter values. Based on such associations, the probability of communication network throughput reduction for a parameter setting of a current communication situation may be derived, and such probability may be used for the prediction step 130.
Derivation of the probability of communication network throughput reduction for a parameter setting may be accomplished in any suitable way. For example, a machine learning algorithm may use the statistics to derive the probability. For example, the machine learning algorithm may apply self-training based on the statistics. Alternatively or additionally, the machine learning algorithm may be trained based on initial training data or other training data related to derivation of the probability of communication network throughput reduction.
Generally, training entities may be implemented in a radio base station serving the user device or in another node (e.g., on a dedicated system outside the radio base station).
Some example parameters that may be related to an occurrence of communication network throughput reduction include user device characteristics (e.g., capability category, UE category, subscription type, vendor, brand, original equipment manufacturer - OEM, etc.), a current time indication (e.g., indicating one or more of: time of day, day or week, date, season, etc.), current and/or previous location of the user device, current and/or previous signal quality of the user device, current and/or previous communication network throughput of the user device, current and/or previous timing advance (TA) of the user device, and load of serving cell and/or target cell.
As a typical example of the prediction of step 130, the statistics may indicate that a user device at a certain location has a corresponding probability of communication network throughput reduction within an upcoming time window. Alternatively or additionally, the statistics may indicate that a user device moving towards a certain location (e.g., from a specific direction) has a corresponding probability of communication network throughput reduction within an upcoming time window. Alternatively or additionally, the statistics may indicate that a user device moving with a certain speed has a corresponding probability of communication network throughput reduction within an upcoming time window. Alternatively or additionally, the statistics may indicate that a user device with a certain current communication network throughput drop has a corresponding probability of communication network throughput reduction within an upcoming time window. Alternatively or additionally, the statistics may indicate that a user device in a cell with certain load has a corresponding probability of communication network throughput reduction within an upcoming time window.
Configuring the scheduler to temporarily increase resource allocation for the user device may, for example, comprise enforcing scheduling prioritization of the user device over one or more other user devices.
Configuring the scheduler to temporarily increase resource allocation for the user device may be further in response to a traffic load that is less than a traffic load threshold. Thus, the resource allocation increase may be applied only in situations with relatively low traffic load.
In some embodiments, the scheduler may be configured to temporarily increase resource allocation for the user device when the cell load increases. Thus, a cell load increase may be seen as an indicator for communication network throughput reduction for the user device.
In conclusion, Figure 3 has demonstrated that latency variations of a communication network may be accommodated by proactively causing a scheduler to increase resource allocation based on predicted reduction of communication network throughput.
Various embodiments may be implemented entirely or partly in a network node such as a radio base station. Alternatively or additionally, various embodiments may be fully or partly implemented in a virtual environment and/or in a cloud environment. For example, training
may be performed in a cloud environment and fed back to a radio base station that causes the scheduler configuration, and/or cloud-based services may be used for storing.
Figure 4 schematically illustrates some example principles according to some embodiments.
Part (a) of Figure 4 shows a user device (e.g., a UE) 400 associated with a latency sensitive service, a radio base station 410, and a communication network throughput reduction prediction algorithm 420. For example, the communication network throughput reduction prediction algorithm 420 may be trained based on one or more optimization targets 430 (e.g., accuracy of throughput prediction during reduction). When the communication network throughput reduction prediction algorithm 420 predicts an upcoming reduction of communication network throughput for the user device 400 based on observations 421 at the base station 410, the base station 410 may be instructed 422 to cause the scheduler to increase resource allocation for the user device.
Part (b) of Figure 4 illustrates example training of the communication network throughput reduction prediction algorithm 420 using reinforcement learning and one or more optimization targets 430.
An interpreter 424 receives observations 421 (parameter setting; e.g., network conditions and/or performance metrics) for the user device. The interpreter 424 informs an agent 425 about a current state based on the observations and provides the agent with a corresponding reward for decisions taken by the agent based on the optimization target 430.
Over time, the agent learns about patterns that improve the service quality for a current state (i.e., patterns of communication network throughput reduction occurrence in relation to observations). Thus, based on state and rewards, the agent may learn to recognize/predict that a communication network throughput reduction will happen in the near future. Consequently, the base station can be instructed 422 to cause the scheduler to increase resource allocation for the user device proactively.
Part (c) of Figure 4 illustrates example training of the communication network throughput reduction prediction algorithm 420 using one or more optimization targets 430 and supervised learning of a neural network 427. During training, observations 421 are input to the neural network 427, as well as to a classification module 426 and a training algorithm 428. Based on
the observations and a previous error, the training algorithm 428 provides weights to the neural network 427. The output of the neural network is compared, in an error calculator 429, with a desired output provided by the classification module 426, and the error is provided to the training algorithm. The training may be iterated until a convergence is achieved. During prediction, observations 421 are input to the neural network 427, the output of which can be used to instruct 422 the base station to cause the scheduler to increase resource allocation for the user device proactively.
Figure 5 illustrates example principles according to some embodiments. In Figure 5, a scenario is depicted using throughput as a function of time. In the throughput plot, available bitrate is indicated as 503 (without application of embodiments) and 504 (with application of embodiments). When a communication network throughput reduction is caused, the available bitrate decreases for the user device, as illustrated by duration 510. The bitrate required by the application is shown at 508.
When a latency sensitive service is associated with the user device, this communication network throughput reduction may negatively impact the end-to-end experience of the service. Applying embodiments of the invention, the scheduler can be configured to temporarily increase resource allocation, responsive to a prediction of the communication network throughput reduction (e.g., at time 501, predicting communication network throughput reduction within a time window 520). Thereby, more resources are available for the latency sensitive service at the start of the communication network throughput reduction 510.
Generally, prediction of a communication network throughput reduction may comprise one or more of: prediction of a communication network throughput value 509, prediction of a communication network throughput being below a threshold value, prediction of decreasing communication network throughput, etc.
If the scheduler was not proactively configured to temporarily increase resource allocation, the available bitrate 503 would drop substantially during the communication network throughput reduction duration 510. A consequence of this scenario is illustrated by application bitrate curve 505. Since the application bitrate 505 is considerably higher than the available
bitrate 503 during the communication network throughput reduction duration 510, there will be queue build-up during the communication network throughput reduction, in turn resulting in latency peaks with potentially severe application impact. Eventually, this will be noticed by the application (e.g., due to lack of feedback from the application client and/or due to the queue build-up) and the application bitrate 505 will need to be lowered substantially to recover from the communication network throughput reduction.
On the other hand, application of embodiments presented herein entails that the available bitrate 504 would drop less during the communication network throughput reduction duration 510, due to that the scheduler is proactively configured to temporarily increase resource allocation. A consequence of this scenario is illustrated by application bitrate curve 506. Since the application bitrate 506 is rather close to the available bitrate 504 during the communication network throughput reduction duration 510, there will be less queue build-up during the communication network throughput reduction, in turn resulting in better latency behavior. The application bitrate 506 may still need to be lowered to recover from the communication network throughput reduction, but the lowering is generally much less pronounced.
In the example of Figure 5, the latency sensitive application may consume less resources on average even if embodiment are applied which increase resource allocation temporarily. This is due to decreased resource allocation (e.g., implemented by a more restrictive priority policy for the service) being applied in situations where the communication network throughput is good enough (e.g., when the available bitrate 504 is larger than the required application bitrate 508). This principle may be applied generally, as suitable.
In an example implementation, some embodiments may manifest themselves by the following rule for controlling the scheduler: increase resource allocation at time 501 when the predicted throughput 509 is lower than the required application bitrate 508, and apply a default resource allocation for latency sensitive services at time 501 when the predicted throughput 509 is higher than the required application bitrate 508. The default resource allocation for latency sensitive services may be lower than a default resource allocation for other types of services.
Figure 6 schematically illustrates an example apparatus 610 for latency control in a communication network according to some embodiments. The apparatus 610 may, for example, be comprisable (e.g., comprised) in a network node (e.g., a radio access node such as a base station, or a central processing node). Alternatively or additionally, the apparatus 610 may be configured to cause execution of one or more of the method steps described herein (e.g., in connection with the method 100 of Figure 3). The apparatus comprises a controller (CNTR; e.g., controlling circuitry or a control module) 600.
The controller 600 is configured to cause identification that a service is currently associated with a user device associated with the communication network, with bounded deviation between a latency requirement of the service and an internal latency performance of the communication network (compare with step 110 of Figure 3).
To this end, the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) an identifier (ID; e.g., identifying circuitry or an identification module) 601. The identifier may be configured to identify that a service is currently associated with a user device associated with the communication network, with bounded deviation between a latency requirement of the service and an internal latency performance of the communication network.
The controller 600 is also configured to cause prediction of an upcoming reduction of communication network throughput for the user device (compare with step 130 of Figure 3).
To this end, the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a predictor (PD; e.g., predicting circuitry or a prediction module) 602. The predictor may be configured to predict the upcoming reduction of communication network throughput for the user device using any suitable approach.
The prediction of the upcoming reduction of communication network throughput for the user device may be based on communication network throughput statistics as explained above in connection to Figure 3.
To this end, the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a statistics database.
The database may be co-located with the controller as illustrated by the database (DB) 603 of Figure 6. Alternatively or additionally (e.g., for a distributed database), the database may be located apart from the controller 600, but accessible by the controller, e.g., via an apparatus interface (I/O) 640.
The predictor may be configured to predict the upcoming reduction of communication network throughput for the user device using any communication network throughput statistics (e.g., based on machine-learning, or other analysis of communication network throughput statistics).
The controller 600 is also configured to cause configuration of a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput (compare with step 140 of Figure 3).
The scheduler may be co-located with the controller as illustrated by the scheduler (SCH) 604 of Figure 6. Alternatively or additionally, the scheduler may be located apart from the controller 600, but accessible by the controller, e.g., via an apparatus interface (I/O) 640.
Configuring the scheduler to increase resource allocation may comprise and/or entail transmission of one or more messages to the user device, wherein the one or more messages are indicative of the increase resource allocation.
To this end, the controller 600 may comprise or be otherwise associated with (e.g., connectable, or connected, to) a transmitter (TX; e.g., transmitting circuitry or a transmission module) 630 and/or an apparatus interface (I/O; e.g., interface circuitry or an interface module) 640. The transmitter may be configured to transmit message(s) to the user device (e.g., when the apparatus is comprisable in a radio access node). The interface may be configured to transfer message(s) to a radio access node for transmission to the user device (e.g., when the apparatus is comprisable in a central processing node).
In some embodiments, the controlling circuitry may be configured to cause prediction and configuration (of the scheduler) only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
The described embodiments and their equivalents may be realized in software or hardware or a combination thereof. The embodiments may be performed by general purpose circuitry. Examples of general purpose circuitry include digital signal processors (DSP), central processing units (CPU), co-processor units, field programmable gate arrays (FPGA) and other programmable hardware. Alternatively or additionally, the embodiments may be performed by specialized circuitry, such as application specific integrated circuits (ASIC). The general purpose circuitry and/or the specialized circuitry may, for example, be associated with or comprised in an apparatus such as a network node.
Embodiments may appear within an electronic apparatus (such as a network node) comprising arrangements, circuitry, and/or logic according to any of the embodiments described herein. Alternatively or additionally, an electronic apparatus (such as a network node) may be configured to perform methods according to any of the embodiments described herein.
According to some embodiments, a computer program product comprises a tangible, or non tangible, computer readable medium such as, for example a universal serial bus (USB) memory, a plug-in card, an embedded drive or a read only memory (ROM). Figure 7 illustrates an example computer readable medium in the form of a compact disc (CD) ROM 700. The computer readable medium has stored thereon a computer program comprising program instructions. The computer program is loadable into a data processor (PROC; e.g., data processing circuitry or a data processing unit) 720, which may, for example, be comprised in a network node 710. When loaded into the data processor, the computer program may be stored in a memory (MEM) 730 associated with or comprised in the data processor. According to some embodiments, the computer program may, when loaded into and run by the data processor, cause execution of method steps according to, for example, any of the methods as illustrated in Figure 3 or otherwise described herein.
Generally, all terms used herein are to be interpreted according to their ordinary meaning in the relevant technical field, unless a different meaning is clearly given and/or is implied from the context in which it is used.
Reference has been made herein to various embodiments. However, a person skilled in the art would recognize numerous variations to the described embodiments that would still fall within the scope of the claims.
For example, the method embodiments described herein discloses example methods through steps being performed in a certain order. However, it is recognized that these sequences of events may take place in another order without departing from the scope of the claims. Furthermore, some method steps may be performed in parallel even though they have been described as being performed in sequence. Thus, the steps of any methods disclosed herein do not have to be performed in the exact order disclosed, unless a step is explicitly described as following or preceding another step and/or where it is implicit that a step must follow or precede another step.
In the same manner, it should be noted that in the description of embodiments, the partition of functional blocks into particular units is by no means intended as limiting. Contrarily, these partitions are merely examples. Functional blocks described herein as one unit may be split into two or more units. Furthermore, functional blocks described herein as being implemented as two or more units may be merged into fewer (e.g. a single) unit.
Any feature of any of the embodiments disclosed herein may be applied to any other embodiment, wherever suitable. Likewise, any advantage of any of the embodiments may apply to any other embodiments, and vice versa.
Hence, it should be understood that the details of the described embodiments are merely examples brought forward for illustrative purposes, and that all variations that fall within the scope of the claims are intended to be embraced therein.
Claims
1. A method for latency control in a communication network, the method comprising: identifying (110) that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded; predicting (130) an upcoming reduction of communication network throughput for the user device; and configuring (140) a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
2. The method of claim 1, wherein the prediction and configuration are performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network.
3. The method of any of claims 1 through 2, wherein predicting the upcoming reduction of communication network throughput for the user device comprises predicting (131) that a future communication network throughput value for the user device is lower than a throughput threshold value.
4. The method of any of claims 1 through 3, wherein predicting the upcoming reduction of communication network throughput for the user device comprises determining (132) that a probability for upcoming communication network throughput reduction for the user device is larger than a reduction probability threshold value.
5. The method of any of claims 1 through 4, wherein predicting the upcoming reduction of communication network throughput for the user device comprises predicting the upcoming reduction of communication network throughput in an upcoming time window (520).
6. The method of claim 5 wherein a start time of the upcoming time window and/or a length of the upcoming time window is based on one or more of: an expected duration of the
reduction of communication network throughput, a maximum acceptable latency of the service, and a buffering setting of the service.
7. The method of any of claims 1 through 6, wherein configuring the scheduler to temporarily increase resource allocation for the user device comprises enforcing scheduling prioritization of the user device over one or more other user devices.
8. The method of any of claims 1 through 7, wherein predicting the upcoming reduction of communication network throughput for the user device comprises predicting that a future throughput is lower than a throughput of the latency sensitive service.
9. The method of any of claims 1 through 8, wherein configuring the scheduler to temporarily increase resource allocation for the user device is further in response to a traffic load that is less than a traffic load threshold.
10. The method of any of claims 1 through 9, wherein predicting the upcoming reduction of communication network throughput for the user device is based on one or more cell- specific prediction criteria and/or on one or more network-general prediction criteria.
11. The method of any of claims 1 through 10, wherein predicting the upcoming reduction of communication network throughput for the user device is based on one or more of: a capability category of the user device; a subscription type of the user device; an original equipment manufacturer, OEM, of the user device; a current time indication; a current and/or previous location of the user device; a current and/or previous signal quality of the user device; a current and/or previous communication network throughput of the user device; a current and/or previous timing advance, TA, of the user device; and a load of a serving cell and/or a target cell.
12. The method of any of claims 1 through 11, wherein predicting the upcoming reduction of communication network throughput for the user device is based on communication network throughput statistics.
13. The method of any of claims 1 through 12, wherein identifying that a service is currently associated with a user device, wherein the deviation between the latency requirement of the service and the internal latency performance of the communication network is bounded, comprises one or more of: detecting that a service class identifier is indicative of the service; detecting that a bearer dedicated for low latency requirements is assigned for the service; and determining that a traffic pattern of the service matches a latency sensitive traffic pattern.
14. The method of any of claims 1 through 13, wherein the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network comprises one or more of: a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network not exceeding a bounding threshold; a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being in a same order of magnitude; a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being equal; and a required end-to-end round-trip-time of the service falling within a time range specified relative an internal round-trip-time of the communication network.
15. The method of any of claims 1 through 14, wherein the service has a maximum allowable latency which is lower than that of mobile broadband, MBB, services and/or higher than that of ultra-reliable low latency communication, URLLC, services.
16. The method of any of claims 1 through 15, wherein the latency control comprises one or more of: decrease of latency variance associated with the communication network for the user device; decrease of a maximum latency associated with the communication network for the user device; decrease of a number of latency events associated with the communication network for the user device, that exceed a latency threshold value; and decrease of an average latency associated with the communication network for the user device.
17. A computer program product comprising a non-transitory computer readable medium
(700), having thereon a computer program comprising program instructions, the computer program being loadable into a data processing unit and configured to cause execution of the method according to any of claims 1 through 16 when the computer program is run by the data processing unit.
18. An apparatus for latency control in a communication network, the apparatus comprising controlling circuitry (600) configured to cause: identification that a service is currently associated with a user device associated with the communication network, wherein a deviation between a latency requirement of the service and an internal latency performance of the communication network is bounded; prediction of an upcoming reduction of communication network throughput for the user device; and configuration of a scheduler to temporarily increase resource allocation for the user device in response to predicting the reduction of communication network throughput.
19. The apparatus of claim 18, wherein the controlling circuitry is configured to cause prediction and configuration to be performed only for user devices associated with services with bounded deviation between the latency requirement of the service and the internal latency performance of the communication network. 20. The apparatus of any of claims 18 through 19, wherein the controlling circuitry is configured to cause prediction of the upcoming reduction of communication network throughput for the user device by causing prediction that a future communication network throughput value for the user device is lower than a throughput threshold value. 21. The apparatus of any of claims 18 through 20, wherein the controlling circuitry is configured to cause prediction of the upcoming reduction of communication network throughput for the user device by causing determination that a probability for upcoming reduction of communication network throughput for the user device is larger than a reduction probability threshold value. 22. The apparatus of any of claims 18 through 21, wherein the controlling circuitry is configured to cause prediction of the upcoming reduction of communication network throughput for the user device by causing prediction of the upcoming reduction of communication network throughput in an upcoming time window.
23. The apparatus of claim 22, wherein a start time of the upcoming time window and/or a length of the upcoming time window is based on one or more of: an expected duration of the reduction of communication network throughput, a maximum acceptable latency of the service, and a buffering setting of the service.
24. The apparatus of any of claims 18 through 23, wherein the controlling circuitry is further configured to cause configuration of the scheduler to temporarily increase resource allocation for the user device by causing enforcement of scheduling prioritization of the user device over one or more other user devices.
25. The apparatus of any of claims 18 through 24, wherein the controlling circuitry is configured to cause prediction of the upcoming reduction of communication network
throughput for the user device by causing prediction that a future throughput is lower than a throughput of the latency sensitive service.
26. The apparatus of any of claims 18 through 25, wherein the controlling circuitry is configured to cause configuration of the scheduler to temporarily increase resource allocation for the user device further in response to a traffic load that is less than a traffic load threshold.
27. The apparatus of any of claims 18 through 26, wherein prediction of the upcoming reduction of communication network throughput for the user device is based on one or more cell-specific prediction criteria and/or on one or more network-general prediction criteria.
28. The apparatus of any of claims 18 through 27, wherein prediction of the upcoming reduction of communication network throughput for the user device is based on one or more of: a capability category of the user device; a subscription type of the user device; an original equipment manufacturer, OEM, of the user device; a current time indication; a current and/or previous location of the user device; a current and/or previous signal quality of the user device; a current and/or previous communication network throughput of the user device; a current and/or previous timing advance, TA, of the user device; and a load of a serving cell and/or a target cell.
29. The apparatus of any of claims 18 through 28, wherein prediction of the upcoming reduction of communication network throughput for the user device is based on communication network throughput statistics.
30. The apparatus of any of claims 18 through 29, wherein the controlling circuitry is configured to cause identification that a service is currently associated with a user device, wherein the deviation between the latency requirement of the service and the internal latency performance of the communication network is bounded, by causing one or more of: detection that a service class identifier profile of the user is indicative of the service; detection that an Internet user authorization, IUA, profile of the user is indicative of the service; detection that a bearer dedicated for low latency requirements is assigned for the user; and determination that a traffic pattern of the user matches a latency sensitive traffic pattern.
31. The apparatus of any of claims 18 through 30, wherein the bounded deviation between the latency requirement of the service and the internal latency performance of the communication network comprises one or more of: a ratio between a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network falling within a bounding range; a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being in a same order of magnitude; a latency requirement parameter value of the service and an internal latency performance parameter value of the communication network being equal; and a required end-to-end round-trip-time of the service falling within a time range specified relative an internal round-trip-time of the communication network.
32. The apparatus of any of claims 18 through 31, wherein the service has a maximum allowable latency which is lower than that of mobile broadband, MBB, services and/or higher than that of ultra-reliable low latency communication, URLLC, services.
33. The apparatus of any of claims 18 through 32, wherein the latency control comprises one or more of: decrease of latency variance associated with the communication network for the user device; decrease of a maximum latency associated with the communication network for the user device; decrease of a number of latency events associated with the communication network for the user device, that exceed a latency threshold value; and decrease of an average latency associated with the communication network for the user device.
34. A network node comprising the apparatus of any of claims 18 through 33.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2020/068580 WO2022002395A1 (en) | 2020-07-01 | 2020-07-01 | Latency control for a communication network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2020/068580 WO2022002395A1 (en) | 2020-07-01 | 2020-07-01 | Latency control for a communication network |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022002395A1 true WO2022002395A1 (en) | 2022-01-06 |
Family
ID=71452238
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2020/068580 WO2022002395A1 (en) | 2020-07-01 | 2020-07-01 | Latency control for a communication network |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2022002395A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110044262A1 (en) * | 2009-08-24 | 2011-02-24 | Clear Wireless, Llc | Apparatus and method for scheduler implementation for best effort (be) prioritization and anti-starvation |
US20190014050A1 (en) * | 2017-07-07 | 2019-01-10 | Qualcomm Incorporated | Apparatus and method for adaptive de-jitter buffer |
US20200015121A1 (en) * | 2018-07-09 | 2020-01-09 | Uhana, Inc. | Systems and methods for mobile network guidance for over-the-top applications |
-
2020
- 2020-07-01 WO PCT/EP2020/068580 patent/WO2022002395A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110044262A1 (en) * | 2009-08-24 | 2011-02-24 | Clear Wireless, Llc | Apparatus and method for scheduler implementation for best effort (be) prioritization and anti-starvation |
US20190014050A1 (en) * | 2017-07-07 | 2019-01-10 | Qualcomm Incorporated | Apparatus and method for adaptive de-jitter buffer |
US20200015121A1 (en) * | 2018-07-09 | 2020-01-09 | Uhana, Inc. | Systems and methods for mobile network guidance for over-the-top applications |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20220182185A1 (en) | Systems and Methods to Reduce Consecutive Packet Loss for Delay Critical Traffic | |
JP5044012B2 (en) | Method and apparatus for dynamic adjustment of uplink transmission time | |
EP2525620B1 (en) | Method, device and communication system for dormancy timer control | |
JP2005522934A (en) | Node B and RNC actions during HSDPA serving cell change | |
EP1774804A1 (en) | Method for controlling communication between two nodes of a communication system | |
US20230275842A1 (en) | Accommodation of latency variations of a communication network | |
CN101420924A (en) | Method and apparatus for Node B controlled scheduling in soft handover | |
KR101882425B1 (en) | A scheduling device and method using virtual slotframe in industrial wireless sensor network | |
CN105766020B (en) | Method for controlling transmission power and transmission delay and communication terminal using the same | |
WO2022002393A1 (en) | Accommodation of latency variations of a communication network | |
CN102457914A (en) | Method and device for transmitting uplink data and uplink scheduling information | |
EP2954733A1 (en) | Transmission power regulation | |
US20100035623A1 (en) | Method and apparatus for controlling quality of service in mobile communication system | |
WO2009157826A1 (en) | Congestion control in a wireless communication network | |
EP4241432B1 (en) | Handling of communication interruption | |
WO2022002395A1 (en) | Latency control for a communication network | |
EP2925048B1 (en) | Method for sharing wireless access network, sending end, and receiving end | |
CN103384387B (en) | Scheduling method for handover control and system, user terminal, the network equipment | |
US20230247500A1 (en) | Latency control for a communication network | |
US20160057804A1 (en) | Optimizing Channel State Switch based on the Traffic Volume Indicator (TVI) Values Associated with Throughputs on the Communication Links | |
WO2015044776A2 (en) | Method and system configuring a base station to trigger hs-dpcch generation | |
US20230269193A1 (en) | Latency control for a communication network | |
RU2632930C1 (en) | Methods, wireless device, base radio station and second network node for controlling unidirectional eps-channel | |
US10206124B1 (en) | Method and apparatus for bidirectional modem | |
JPWO2006006208A1 (en) | Radio base station |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20736327 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20736327 Country of ref document: EP Kind code of ref document: A1 |