WO2019006649A1 - Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle - Google Patents

Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle Download PDF

Info

Publication number
WO2019006649A1
WO2019006649A1 PCT/CN2017/091603 CN2017091603W WO2019006649A1 WO 2019006649 A1 WO2019006649 A1 WO 2019006649A1 CN 2017091603 W CN2017091603 W CN 2017091603W WO 2019006649 A1 WO2019006649 A1 WO 2019006649A1
Authority
WO
WIPO (PCT)
Prior art keywords
scaling
data
network function
time period
capacity
Prior art date
Application number
PCT/CN2017/091603
Other languages
English (en)
Inventor
Huoming DONG
Wei Huang
Original Assignee
Telefonaktiebolaget Lm Ericsson (Publ)
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Telefonaktiebolaget Lm Ericsson (Publ) filed Critical Telefonaktiebolaget Lm Ericsson (Publ)
Priority to PCT/CN2017/091603 priority Critical patent/WO2019006649A1/fr
Publication of WO2019006649A1 publication Critical patent/WO2019006649A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/147Network analysis or design for predicting network behaviour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/10Machine learning using kernel methods, e.g. support vector machines [SVM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/01Dynamic search techniques; Heuristics; Dynamic trees; Branch-and-bound
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0896Bandwidth or capacity management, i.e. automatically increasing or decreasing capacities
    • H04L41/0897Bandwidth or capacity management, i.e. automatically increasing or decreasing capacities by horizontal or vertical scaling of resources, or by migrating entities, e.g. virtual resources or entities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/16Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using machine learning or artificial intelligence

Definitions

  • Embodiments of the present disclosure generally relate to the field of communications, and more particularly, to a method and device for network function capacity and scaling management.
  • NFs network functions
  • These NFs may communicate with surrounding peers to process and transport control signaling and/or user plane data through standardized communication and network protocols.
  • NFV network function virtualization
  • VNF virtualized network functions
  • scalability of a native NF or VNF is a very important and key feature for service providers.
  • scaling in/out is considered as a basic feature and is deemed as an enabler for quickly modifying capacity of products/services on demand in order to meet increasingly business needs.
  • WO2017/020298A1 invents a method to automatically manage peer connections in cloud, which could reduce manual work for operators and avoid human faults.
  • VNF capacity scaling is only reactive based approaches, e.g., depending on threshold alarms and key performance indicators (KPIs) .
  • KPIs key performance indicators
  • the reactive based approach has some disadvantages, such as has a time lag and is not real-time solution, because when a scaling indicator is triggered and sent to a VNF manager and/or when a scaling operation is started by the VNF manager, possibly the indicated system situation (such as overloaded or underloaded) has sustained for a long time or has been over already.
  • This time lag could heavily impact on system characteristics (such as overloaded) or system resources (such as underloaded) , and this scenario could happen frequently in a dynamic cloud environment.
  • the time delay could be produced due to analysis and synthesis of different level system alarms from network and processing time of getting system KPIs (such as counters and statistics) from network. For example, there are 5-15 minutes in most of the systems.
  • VNF scaling out is performed when the system load reaches 70%instead of the real overload indication like 90%. This might help in one side, but on the other hand, this also leads to VNF overprovisioning and resource/energy inefficiency.
  • embodiments of the present disclosure provide a solution for network function capacity and scaling management.
  • a method for network function capacity and scaling management includes: retrieving characteristics and configuration data from a scaling trigger handler; obtaining capacity prediction information of a next time period by using the characteristics and configuration data of one or more past time periods based on machine learning; and sending the capacity prediction information of the next time period to a scaling decision maker.
  • the method is implemented in a virtualized network function capacity prediction engine which enables a predictive based scaling instead of a reactive based scaling.
  • the scaling trigger handler is configured or implemented in an element management system or a network management system
  • the scaling decision maker is configured or implemented in a virtualized network function manager
  • the virtualized network function capacity prediction engine is configured or implemented in the virtualized network function manager or the element management system or the network management system.
  • the characteristics and configuration data comprises one or more of the following information: alarms, key performance indicators, system configurations, system running data, system resources; the time period is configurable and comprises one or more of the following time intervals: minutes, hours, days, weeks, months and years.
  • the virtualized network function capacity prediction engine regularly retrieves the characteristics and configuration data from the scaling trigger handler; and the scaling trigger handler regularly retrieves the characteristics and configuration data from the virtualized network functions.
  • the virtualized network function capacity prediction engine receives the characteristics and configuration data reported by the scaling trigger handler; and the scaling trigger handler receives the characteristics and configuration data reported by virtualized network functions.
  • the method comprises: training a machine learning model by using the characteristics and configuration data of one or more past time periods to discover inherent patterns from the characteristics and configuration data and to build an analytics model; predicting one or more capacities of the next time period using the built analytics model; and concluding resource needs of the next time period by comparing the predicted capacities and allocated resources of the next time period.
  • techniques applying the machine learning comprises applying supervised learning models or methods; one or more of following models or methods are used for the machine learning: regression analysis, supported vector machine, decision trees, random forest, artificial neural networks.
  • a fused model is used by combining backpropagation neural networks, radial basis functions and generalized regression neural networks.
  • the method further comprises: synthesizing multiple predicted capacities with different weights, and/or, combining multiple input data and/or output data with the different weights.
  • the different weights are calculated based on the following rules: the more aged the data is, the smaller value the weight has; and/or; the more similar the time period is, the bigger value the weight has.
  • the method further comprises: using actual capacity data of a current time period as the latest training data of the machine learning when the current time period is over.
  • the method further comprises: discarding the data of the most aged period; and revising a prediction model of the machine learning by using the latest training data.
  • a system for network function capacity and scaling management comprises a scaling trigger handler and a scaling decision maker; the system further comprises: a capacity prediction engine configured to perform a method according the first aspect.
  • system further comprises an element management system or a network management system, a virtualized network function manager, a virtualized infrastructure manager and a network function virtualization infrastructure.
  • the scaling trigger handler is configured or implemented in an element management system or a network management system
  • the scaling decision maker is configured or implemented in a virtualized network function manager
  • the capacity prediction engine is used for virtualized network function capacity and scaling management and is configured or implemented in the virtualized network function manager or the element management system or the network management system.
  • a device comprising a processor and a memory, wherein the memory containing instructions executable by the processor whereby the device is operative to perform a method according to the first aspect.
  • a predictive based approach is provided by using data insights through machine learning techniques, instead of the current reactive based approach. Therefore, scaling needs may be predicted by machine learning in advance, and thus scaling decision and action can be made at just right time with just right resources.
  • Fig. 1 is a schematic diagram which shows an example for VNF scaling
  • Fig. 2 is a schematic diagram which shows an example of reactive based VNF scaling
  • Fig. 3 is a flowchart which shows a method 300 for network function capacity and scaling management in accordance with an embodiment of the present disclosure
  • Fig. 4 is a schematic diagram which shows an example of predictive based VNF scaling in accordance with an embodiment of the present disclosure
  • Fig. 5 is a schematic diagram which shows a method of machine learning in accordance with an embodiment of the present disclosure
  • Fig. 6 is a schematic diagram which shows an example of VNF resource needs prediction results showing deviations of predicted needs and real needs in accordance with an embodiment of the present disclosure
  • Fig. 7 is a schematic diagram which shows an example of prediction accuracy and data size in accordance with an embodiment of the present disclosure
  • Fig. 8 shows a block diagram of an apparatus 800 for network function capacity and scaling management in accordance with an embodiment of the present disclosure
  • Fig. 9 is a simplified block diagram of a device that is suitable for implementing embodiments of the present disclosure.
  • references in the specification to “one embodiment, ” “an embodiment, ” “an example embodiment, ” and the like indicate that the embodiment described may include a particular feature, structure, or characteristic, but it is not necessary that every embodiment includes the particular feature, structure, or characteristic. Moreover, such phrases are not necessarily referring to the same embodiment. Further, when a particular feature, structure, or characteristic is described in connection with an embodiment, it is submitted that it is within the knowledge of one skilled in the art to affect such feature, structure, or characteristic in connection with other embodiments whether or not explicitly described.
  • first and second etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element, and similarly, a second element could be termed a first element, without departing from the scope of example embodiments.
  • the term “and/or” includes any and all combinations of one or more of the associated listed terms.
  • the network device may include processing circuitry, device readable medium, interface, user interface equipment, auxiliary equipment, power source, power delivery circuitry, and antenna. These components are depicted as single boxes located within a single larger box, and in some cases, contain additional boxes therein.
  • the network device may include multiple different physical components that make up a single illustrated component (e.g., interface includes ports or terminals for coupling wires for a wired connection and radio front end circuitry for a wireless connection) .
  • network device may be a virtual network node.
  • network node may be composed of multiple physically separate components (e.g., a NodeB component and a RNC component, a BTS component and a BSC component, etc. ) , which may each have their own respective components.
  • one or more of the separate components may be shared among several network nodes.
  • a single RNC may control multiple NodeB’s .
  • each unique NodeB and RNC pair may in some instances be considered a single separate network node.
  • network node may be configured to support multiple radio access technologies (RATs) .
  • RATs radio access technologies
  • some components may be duplicated (e.g., separate device readable medium for the different RATs) and some components may be reused (e.g., the same antenna may be shared by the RATs) .
  • IMS IP multimedia subsystem
  • SBC session border controller
  • MGw media gateway
  • Fig. 1 is a schematic diagram which shows an example for VNF scaling.
  • H. 248 protocol may be used and there may be some running instances, such as SBC 1, SBC 2, whil; and MGw 1, MGw 2, . «, at virtual machine (VM) level.
  • VM virtual machine
  • Fig. 2 is a schematic diagram which shows an example of reactive based VNF scaling.
  • modules or components such as an element management system (EMS) or a network management system (NMS) , a virtualized network function manager (VFN manager) , a virtualized infrastructure manager (VIM) and a network function virtualization infrastructure (NFVI) .
  • EMS element management system
  • NMS network management system
  • VNF manager virtualized network function manager
  • VIM virtualized infrastructure manager
  • NFVI network function virtualization infrastructure
  • Some VNFs may be provided for a plurality of resources, such as memory, central processing unit (CPU) , bandwidth, I/O, disk, and so on.
  • a scaling trigger handler may be configured or implemented in the EMS/NMS, and a scaling decision maker (SDM) may be configured or implemented in the VNF manager.
  • some blocks (or steps) may be performed for the VNF scaling.
  • VNF may report alarms and KPIs to EMS/NMS during a running period, at block 201.
  • a report will be generated periodically, such as 15 minutes in many of systems which is hard to be shorten as considering impact of characteristics and performance of the whole network.
  • the STH included in EMS/NMS may further analyze and synthesize the scaling related alarms and/or KPIs, and decide if a request for scaling should be triggered according to VNF business/application level knowledge. If a request for scaling should be triggered, the STH then sends the request to the VNF manager for further handling, at block 202.
  • the SDM included in the VNF manager may analyze the received request, together with other relevant VNF and system data, e.g. VNF life cycle management. The SDM may finally decide if a scaling action (for example, scaling in or scaling out) shall be performed on the VNF.
  • a scaling action for example, scaling in or scaling out
  • the VNF manager may notify the VNF to prepare the scaling, at block 203. For example, the VNF may be notified that a newly instance will be added later if scaling out is decided, or ongoing traffic or sessions will be released later if scaling in is decided.
  • the VNF manager may notify VIM to start scaling, at block 204.
  • the VIM may be notified to add VM resources or reduce VM resources.
  • the VIM may notify the NFVI at block 205.
  • the NFVI may perform VM instance (s) creation or removal and the VIM may receive an acknowledge from the NFVI.
  • the NFVI may notify the VNF a new status of VM resources, at block 206.
  • the VNF may conclude the scaling and send the result of scaling to the VNF manager, at block 207.
  • the VNF manager may update system data internally.
  • the reactive based approach has time lag and is not real-time solution, thus it could heavily impact on system characteristics and performance (such as overloaded) and/or waste system resources (such as underloaded) in a dynamic cloud environment.
  • some preventive measures adopted may cause overprotection. Therefore, the reactive based approach may lead to system characteristics decrease and/or resource/energy inefficiency.
  • triggering and decision for scaling are not easy to be made, as there are too many relevant alarms and KPIs at different system levels which need to be thoroughly analyzed and synthesized.
  • a lot of cloud parameters and uncertainties are introduced and need to be measured, which needs much efforts and makes that it is almost not possible to perform internal accurate measurements. All of those may lead to a wrong scaling decision by the reactive based approach.
  • machine learning is a method of data analysis that automates analytical model building. Using algorithms that iteratively learn from dataset, machine learning allows computers to find hidden patterns without being explicitly programmed. It’s a subfield of artificial intelligence (AI) . Machine learning has been widely applied in many areas.
  • AI artificial intelligence
  • a predictive based approach is provided for scaling and capacity management (such as for VNF) by using machine learning techniques.
  • the solution can be applied in any similar cloud scaling scenarios on both VM level and VNF level in networks.
  • the solution in this disclosure can also be applied/extended to the whole network level scaling, including radio/access network, core network, service network, etc., e.g. based on historical time period data of radio network characteristics, the machine learning based approach could provide predictive based core network scaling in the future time period through learning/training.
  • a method for network function capacity and scaling management is provided.
  • Fig. 3 is a flowchart which shows a method 300 for network function capacity and scaling management in accordance with an embodiment of the present disclosure, and illustrates the method for network function capacity and scaling management by taking a capacity prediction engine as an example.
  • the method 300 includes retrieving, by a capacity prediction engine, characteristics and configuration data from a scaling trigger handler (STH) , at block 301; obtaining, by the capacity prediction engine, capacity prediction information of a next time period by using the characteristics and configuration data of one or more past time periods based on machine learning, at block 302; and sending, by the capacity prediction engine, the capacity prediction information of the next time period to a scaling decision maker (SDM) .
  • STH scaling trigger handler
  • SDM scaling decision maker
  • the method may be implemented in a VNF capacity prediction engine (VCPE) which enables a predictive based scaling instead of a reactive based scaling.
  • VCPE VNF capacity prediction engine
  • the STH may be configured or implemented in an element management system (EMS) or a network management system (NMS)
  • the SDM may be configured or implemented in a virtualized network function manager (VNF manager)
  • the VCPE may be configured or implemented in the VNF manager or the EMS or the NMS.
  • the characteristics and configuration data may include one or more of the following information: alarms, key performance indicators, system configurations, system running data, system resources; however, it is not limited in this disclosure.
  • the time period is configurable and may include one or more of the following time intervals: minutes, hours, days, weeks, months and years. However, it is not limited in this disclosure. For example, other time interval (or time unit) may be adopted according to actual scenarios.
  • VCPE virtualized network function capacity prediction engine
  • Fig. 4 is a schematic diagram which shows an example of predictive based VNF scaling in accordance with an embodiment of the present disclosure.
  • modules or components such as an EMS/NMS, a VFN manager, a VIM and a NFVI.
  • Some VNFs may be provided for a plurality of resources, such as memory, central processing unit (CPU) , bandwidth, I/O, disk, and so on.
  • a STH may be configured or implemented in the EMS/NMS
  • a SDM may be configured or implemented in the VNF manager.
  • a VCPE may be configured or implemented in the VNF manager. However, it is not limited thereto.
  • the VCPE may also be configured or implemented in the EMS/NMS.
  • some blocks (or steps) may be performed for the VNF scaling.
  • VNF may report alarms and/or KPIs to EMS/NMS during a running period, at block 401.
  • a report will be generated periodically, such as 15 minutes in many of systems which is hard to be shorten as considering impact of characteristics and performance of the whole network.
  • the STH included in EMS/NMS may regularly retrieve the characteristics and configuration data from the VNF, at block 4021. For example, STH retrieves additional data required by the VCPE from VNF except that is reported and received from VNF already in block 401.
  • the VCPE may receive the characteristics and configuration data reported by the STH, at block 4022; and may also regularly retrieve the characteristics and configuration data from the STH, at clock 4023.
  • VCPE retrieves the data from STH regularly, e.g. per day time period, including alarms, KPIs, system configurations, system running data (resources) , etc. This data will be used as training (labelled) data for supervised machine learning by VCPE.
  • the VCPE may obtain capacity prediction information of a next time period by using the characteristics and configuration data of one or more past time periods based on machine learning.
  • the VCPE when the VCPE get the data (characteristics and/or configuration data) , it will firstly train a machine learning model, and after training it will perform future VNF capacity prediction (such as resource consumption prediction) for next time period. Then, the VCPE may compare the future system resource needs and the allocated resources at that time, to decide when the scaling is needed by pre-configured scaling policy, e.g. scaling out is performed if exceeding over 90%or scaling in is performed if dropping under 50%, during the next time period. The detail of the capacity prediction will be illustrated later.
  • VNF capacity prediction such as resource consumption prediction
  • the VCPE may send the capacity prediction information of the next time period to SDM, at block 4024.
  • the VCPE may send the scaling prediction data of the next time period to the SDM with time indications, thus SDM can plan the VM level or VNF level scaling in advance.
  • the SDM included in the VNF manager may analyze the information and may finally decide if a scaling action (for example, scaling in or scaling out) shall be performed on the VNF.
  • a scaling action for example, scaling in or scaling out
  • the VNF manager may notify the VNF to prepare the scaling, at block 403. For example, the VNF may be notified that a newly instance will be added later if scaling out is decided, or ongoing traffic or sessions will be released later if scaling in is decided.
  • the VNF manager may notify VIM to start scaling, at block 404.
  • the VIM may be notified to add VM resources or reduce VM resources.
  • the VIM may notify the NFVI at block 405.
  • the NFVI may perform VM instance (s) creation or removal and the VIM may receive an acknowledge from the NFVI.
  • the NFVI may notify the VNF a new status of VM resources, at block 406.
  • the VNF may conclude the scaling and send the result of scaling to the VNF manager, at block 407.
  • the VNF manager may update system data internally.
  • VNF scaling needs may be predicted by machine learning in advance, and thus VNF scaling decision and action can be made at just right time with just right resources.
  • Fig. 4 is only an example of the disclosure, but it is not limited thereto.
  • the order of operations at blocks may be adjusted and/or some blocks may be omitted.
  • some blocks not shown in Fig. 4 may be added.
  • Fig. 5 is a schematic diagram which shows a method of machine learning in accordance with an embodiment of the present disclosure.
  • the method 500 may include: training a machine learning model by using the characteristics and configuration data of one or more past time periods, at block 501. This block may be referred to as a training phase.
  • techniques applying the machine learning may include applying supervised learning models or methods; and one or more of following models or methods may be used for the machine learning: regression analysis (linear and/or non-linear) , supported vector machine, decision trees, random forest, Bayesian statistics, artificial neural networks (including deep neural networks) .
  • regression analysis linear and/or non-linear
  • supported vector machine decision trees
  • random forest random forest
  • Bayesian statistics artificial neural networks
  • the method 500 may include: predicting one or more capacities of the next time period using the built analytics model, at block 502. This block may be referred to as a predicting phase.
  • the future VNF capacity prediction of next time period may be performed by using the learnt patterns stored in the model automatically.
  • the method 500 may include: concluding resource needs of the next time period by comparing the predicted capacities and allocated resources of the next time period, at block 503. This block may be referred to as a synthesizing phase.
  • VCPE can conclude the VNF resource needs and/or scaling plan at a certain time, and then provides this data to SDM, which can make scaling plan in advance.
  • the time period (T1) for how long the subsequent VNF future capacity prediction is done may be configurable at VCPE. It could be hour (s) , day (s) , week (s) , month (s) , and so on. However, it is not limited and may be depending on different VNFs and deployments.
  • VCPE retrieves data from EMS/NMS, where the data is related to runtime VNF capacity, including KPIs (there should be no scaling related alarms if VCPE works well) , system configurations and runtime resource data. All data at a certain time actually represents one certain traffic model of system/network and corresponding needed system resource consumption. VCPE will then filter and shape the data as training data for machine learning, for example, removal of unused data, data consolidation per time granularity, data representation (e.g. numeralization) , data standardization and normalization, etc.
  • KPIs there should be no scaling related alarms if VCPE works well
  • All data at a certain time actually represents one certain traffic model of system/network and corresponding needed system resource consumption.
  • VCPE will then filter and shape the data as training data for machine learning, for example, removal of unused data, data consolidation per time granularity, data representation (e.g. numeralization) , data standardization and normalization, etc.
  • the time granularity (T0) could be seconds, minutes, hours, days and so on, and is configurable at VCPE. This also corresponds to the time granularity of the future VNF capacity prediction during next time period. Usually it’s not necessary to be too small for practical use at scaling and also not impact on system characteristics and performance, but T0 should be smaller than T1.
  • the method may further include: using actual capacity data of a current time period as the latest training data of the machine learning when the current time period is over; discarding the data of the most aged period; and revising a prediction model of the machine learning by using the latest training data.
  • VCPE may use the data of previous N time periods, where N is configurable at VCPE.
  • N is configurable at VCPE.
  • the actual VNF capacity data of current time period will be used as latest new training data, and also used to revise/update the prediction model, e.g. adjust model accordingly to fit better.
  • the data of most aged time period will be phased out and replaced by the latest data.
  • VCPE may predict VNF capacity of next time period based on each training dataset (N in total) from the previous N time periods.
  • the method may further include: synthesizing multiple predicted capacities with different weights, and/or, combining multiple input data and/or output data with the different weights.
  • the different weights may be calculated based on the following rules: the more aged the data is, the smaller value the weight has; and/or; the more similar the time period is, the bigger value the weight has.
  • the final prediction will be the synthesis of all N predictions with different weights.
  • the weights are calculated based on below rules: such as, the more aged data, the smaller weight value, and/or, the more similar the time period (traffic model) is, the bigger weight value it has, e.g. traffic model of Friday evenings can be considered similar and traffic model of workday can be seen different from that of weekend.
  • each training sample can contain more historic knowledge, e.g. one training sample data as follows,
  • the historical data will have different weights (principles as described above) to contribute combining the input sample.
  • model could predict several outputs over several future time periods, so the current system behavior and model prediction will have impact on a longer time period instead of only one future time period.
  • the weight setting principles as described above shall be utilized to make use all outputs to synthesize the final output for one specific future time period.
  • VNF scaling needs may be predicted by machine learning in advance, and thus VNF scaling decision and action can be made at just right time with just right resources.
  • SBC Session Border Controller
  • H. 248 protocol is used to communicate with each other.
  • SBC may be P-CSCF, IMS-ALG on UNI/PNI interface, or IBCF on NNI interface.
  • MGW may be C-BGF (IMS Access Gateway) on UNI/PNI interface and I-BGF (TrGW) on the NNI interface.
  • the input data (traffic model data) and output data (prediction data) are related to VNF capacity.
  • the input data could vary in different systems, but output data is similar, related to VNF capacity (resource) data. Take SBC and MGw as examples.
  • the input data may include: no. of registered users, session initiation protocol (SIP) call rate, transport protocols (UDP, TCP, TLS) , IP version (IPv4 or IPv6) , signaling interface (Gm, Rq, Rf, e2) , call hold time, call answer time, hypervisor type, CPU frequency, hyper-threading switch, etc.
  • the output data may include: memory usage (MB) , CPU usage (%) , Bandwidth usage.
  • the input data may include: subscribers, traffic per subscriber, MHT (mean hold time) , total sessions, traffic type (Access/Core/MSS/MRF/TDM) , audio codec (AMR-WB/AMR-NB/G. 711/G. 722/G. 729) , audio transcoding, hypervisor type, vSwitch variant, etc.
  • the output data may include: no. of vCPUs, no. of VMs, memory of VM, memory of VNF, disk size, bandwidth/packet rates, etc.
  • ANN artificial neural networks
  • the example model (it may be referred to as AlphaCANDI in this disclosure) is a fused model by using information fusion technology to combine three feedforward neural networks: BP (Backpropagation) neural networks, RBF (Radial Basis Functions) neural networks and GRNN (Generalized Regression Neural Networks) .
  • BP Backpropagation
  • RBF Random Basis Functions
  • GRNN Generalized Regression Neural Networks
  • MAE Mean Absolute Error
  • RMSE Root of Mean Square Error
  • BP network is a typical feedforward neural networks model, which uses BP (backpropagation) learning algorithm during training.
  • Backpropagation is a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent.
  • RBF network is an artificial neural network that uses Radial Basis Functions (RBF) as activation functions.
  • RBF Radial Basis Functions
  • the output of the network is a linear combination of radial basis functions of the inputs and neuron parameters.
  • RBF network does local approximation of non-linear mapping instead of global approximation, thus it only needs few training samples. It features fast learning (convergence) and can achieve better accuracy.
  • GRNN Generalized Regression Neural Network
  • RBF Regression Neural Network
  • GRNN is a variant of RBF network. It has a radial basis layer and a special linear layer. Unlike the standard RBF, GRNN is distinct in the calculation of the final outputs, i.e. a weighted sum of the radial basis functions and the training values by using normalization and dot product operation, and can be thought of as a normalized RBF network.
  • the fused model is only an example of the disclosure. However, it is not limited thereto, for example, other models may also be adopted according to actual scenarios.
  • nSBC native product
  • vSBC and vMGw virtualized products
  • Fig. 6 is a schematic diagram which shows an example of VNF resource needs prediction results showing deviations of predicted needs and real needs in accordance with an embodiment of the present disclosure.
  • Fig. 7 is a schematic diagram which shows an example of prediction accuracy and data size in accordance with an embodiment of the present disclosure.
  • cloud/virtualized environments are more complex but doesn’t bring any additional complexity and difficulty to machine learning based model while it does to today’s approaches.
  • the training time may be at sec/min level, while the prediction time may be at ms level. This can fit in VNF scaling prediction.
  • VNF resource usage is just at right time and no waste on system resource/energy.
  • machine learning based model may simplify complexity of cloud environments and traffic scenarios (complexity and scenario agnostic) , and provide good prediction accuracy and robustness of VNF resource needs in the future time period. This avoids incorrect VNF scaling decisions and operations, i.e. no scaling if actually needed or doing scaling if actually not needed. Therefore, it may reduce user complains and increase resource efficiency, and improve overall system characteristics and network performance.
  • the apparatus may be configured in a network device, and the same contents as those in the first aspect of embodiments are omitted.
  • Fig. 8 shows a block diagram of an apparatus 800 for network function capacity and scaling management in accordance with an embodiment of the present disclosure.
  • the apparatus 800 includes: a retrieving unit 801 configured to retrieve characteristics and configuration data from a scaling trigger handler; an obtaining unit 802 configured to obtain capacity prediction information of a next time period by using the characteristics and configuration data of one or more past time periods based on machine learning; and a sending unit 803 configured to send the capacity prediction information of the next time period to a scaling decision maker.
  • the apparatus 800 may be implemented in a virtualized network function capacity prediction engine which enables a predictive based scaling instead of a reactive based scaling.
  • the scaling trigger handler may be configured or implemented in an element management system or a network management system; the scaling decision maker may be configured or implemented in a virtualized network function manager; the virtualized network function capacity prediction engine may be configured or implemented in the virtualized network function manager or the element management system or the network management system.
  • the characteristics and configuration data may include one or more of the following information: alarms, key performance indicators, system configurations, system running data, system resources; the time period is configurable and may include one or more of the following time intervals: minutes, hours, days, weeks, months and years. However, it is not limited thereto.
  • the virtualized network function capacity prediction engine may regularly retrieve the characteristics and configuration data from the scaling trigger handler; and the scaling trigger handler may regularly retrieve the characteristics and configuration data from the virtualized network functions.
  • the virtualized network function capacity prediction engine may receive the characteristics and configuration data reported by the scaling trigger handler; and the scaling trigger handler may receive the characteristics and configuration data reported by virtualized network functions.
  • the obtaining unit 802 may be configured to train a machine learning model by using the characteristics and configuration data of one or more past time periods to discover inherent patterns from the characteristics and configuration data and to build an analytics model; predict one or more capacities of the next time period using the built analytics model; and conclude resource needs of the next time period by comparing the predicted capacities and allocated resources of the next time period.
  • techniques applying the machine learning may include applying supervised learning models or methods; one or more of following models or methods are used for the machine learning: regression analysis, supported vector machine, decision trees, random forest, artificial neural networks. However, it is not limited thereto.
  • a fused model may be used by combining backpropagation neural networks, radial basis functions and generalized regression neural networks.
  • backpropagation neural networks radial basis functions
  • generalized regression neural networks it is not limited thereto.
  • multiple predicted capacities may be synthesized with different weights, and/or, multiple input data and/or output data may be combined with the different weights.
  • the different weights may be calculated based on the following rules: the more aged the data is, the smaller value the weight has; and/or; the more similar the time period is, the bigger value the weight has. However, it is not limited thereto.
  • actual capacity data of a current time period may be used as the latest training data of the machine learning when the current time period is over.
  • the data of the most aged period may be discarded; and a prediction model of the machine learning may be revised by using the latest training data.
  • components included in the apparatus 800 correspond to the operations of the method 300 or 400. Therefore, all operations and features described above with reference to Fig. 3 or 4 are likewise applicable to the components included in the apparatus 800 and have similar effects. For the purpose of simplification, the details will be omitted.
  • the components included in the apparatus 800 may be implemented in various manners, including software, hardware, firmware, or any combination thereof.
  • one or more units may be implemented using software and/or firmware, for example, machine-executable instructions stored on the storage medium.
  • machine-executable instructions for example, machine-executable instructions stored on the storage medium.
  • parts or all of the components included in the apparatus 800 may be implemented, at least in part, by one or more hardware logic components.
  • FPGAs Field-programmable Gate Arrays
  • ASICs Application-specific Integrated Circuits
  • ASSPs Application-specific Standard Products
  • SOCs System-on-a-chip systems
  • CPLDs Complex Programmable Logic Devices
  • the apparatus 800 may be a part of a device. But it is not limited thereto, for example, the apparatus 800 may be the network device, other parts of the network device, such as transmitter and receiver, are omitted in the Fig. 8.
  • a predictive based approach is provided by using data insights through machine learning techniques, instead of the current reactive based approach.
  • Scaling needs may be predicted by machine learning in advance, and thus scaling decision and action can be made at just right time with just right resources.
  • a system includes a network device configured to perform a method for network function capacity and scaling management according to the first aspect of embodiments.
  • the system may include a scaling trigger handler, a scaling decision maker, and a capacity prediction engine configured to perform a method according to the first aspect of embodiments.
  • the system may further include an element management system or a network management system, a virtualized network function manager, a virtualized infrastructure manager and a network function virtualization infrastructure.
  • the scaling trigger handler is configured or implemented in an element management system or a network management system, the scaling decision maker is configured or implemented in a virtualized network function manager; the capacity prediction engine is used for virtualized network function capacity and scaling management and is configured or implemented in the virtualized network function manager or the element management system or the network management system.
  • a device is provided in an embodiment, and the same contents as those in the first aspect and the second aspect of embodiments are omitted.
  • Fig. 9 shows a simplified block diagram of a device 900 that is suitable for implementing embodiments of the present disclosure. It would be appreciated that the device 900 may be implemented as at least a part of, for example, the network device.
  • the device 900 includes a communicating means 930 and a processing means 950.
  • the processing means 950 includes a data processor (DP) 910, a memory (MEM) 920 coupled to the DP 910.
  • the communicating means 930 is coupled to the DP 910 in the processing means 950.
  • the MEM 920 stores a program (PROG) 940.
  • the communicating means 930 is for communications with other devices, which may be implemented as a transceiver for transmitting/receiving signals.
  • the memory 920 stores a plurality of instructions; and the processor 910 coupled to the memory 920 and configured to execute the instructions to: retrieve characteristics and configuration data from a scaling trigger handler; obtain capacity prediction information of a next time period by using the characteristics and configuration data of one or more past time periods based on machine learning; and send the capacity prediction information of the next time period to a scaling decision maker.
  • the PROG 940 is assumed to include program instructions that, when executed by the associated DP 910, enable the device 900 to operate in accordance with the embodiments of the present disclosure, as discussed herein with the methods 300 or 400.
  • the embodiments herein may be implemented by computer software executable by the DP 910 of the device 900, or by hardware, or by a combination of software and hardware.
  • a combination of the data processor 910 and MEM 920 may form processing means 950 adapted to implement various embodiments of the present disclosure.
  • the MEM 920 may be of any type suitable to the local technical environment and may be implemented using any suitable data storage technology, such as semiconductor based memory devices, magnetic memory devices and systems, optical memory devices and systems, fixed memory and removable memory, as non-limiting examples. While only one MEM is shown in the device 900, there may be several physically distinct memory modules in the device 900.
  • the DP 910 may be of any type suitable to the local technical environment, and may include one or more of general purpose computers, special purpose computers, microprocessors, digital signal processors (DSPs) and processors based on multicore processor architecture, as non-limiting examples.
  • the device 900 may have multiple processors, such as an application specific integrated circuit chip that is slaved in time to a clock which synchronizes the main processor.
  • various embodiments of the present disclosure may be implemented in hardware or special purpose circuits, software, logic or any combination thereof. Some aspects may be implemented in hardware, while other aspects may be implemented in firmware or software which may be executed by a controller, microprocessor or other computing devices. While various aspects of embodiments of the present disclosure are illustrated and described as block diagrams, flowcharts, or using some other pictorial representation, it will be appreciated that the blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non-limiting examples, hardware, software, firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof.
  • embodiments of the present disclosure can be described in the general context of machine-executable instructions, such as those included in program modules, being executed in a device on a target real or virtual processor.
  • program modules include routines, programs, libraries, objects, classes, components, data structures, or the like that perform particular tasks or implement particular abstract data types.
  • the functionality of the program modules may be combined or split between program modules as desired in various embodiments.
  • Machine-executable instructions for program modules may be executed within a local or distributed device. In a distributed device, program modules may be located in both local and remote storage media.
  • Program code for carrying out methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general-purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program codes, when executed by the processor or controller, cause the functions/operations specified in the flowcharts and/or block diagrams to be implemented.
  • the program code may execute entirely on a machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
  • the above program code may be embodied on a machine-readable medium, which may be any tangible medium that may contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • the machine-readable medium may include but not limited to an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • machine-readable storage medium More specific examples of the machine-readable storage medium would include an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random-access memory (RAM) , a read-only memory (ROM) , an erasable programmable read-only memory (EPROM or Flash memory) , an optical fiber, a portable compact disc read-only memory (CD-ROM) , an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
  • RAM random-access memory
  • ROM read-only memory
  • EPROM or Flash memory erasable programmable read-only memory
  • CD-ROM portable compact disc read-only memory
  • magnetic storage device or any suitable combination of the foregoing.
  • the device may be implemented in the general context of computer system-executable instructions, such as program modules, being executed by a computer system.
  • program modules may include routines, programs, objects, components, logic, data structures, and so on that perform particular tasks or implement particular abstract data types.
  • the device may be practiced in distributed cloud computing environments where tasks are performed by remote processing devices that are linked through a communications network.
  • program modules may be located in both local and remote computer system storage media including memory storage devices.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Databases & Information Systems (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

L'invention concerne un procédé et un dispositif pour une capacité de fonction réseau et une gestion d'échelle. Une approche fondée sur la prédiction faisant appel à des données percutantes par le biais de technique d'apprentissage machine remplace l'approche actuelle qui est fondée sur la réaction. Par conséquent, des besoins de mise à l'échelle peuvent être anticipés par l'apprentissage machine, ce qui permet aux décisions et aux actions de mise à l'échelle d'être réalisées exactement au bon moment et à l'aide des ressources exactement requises.
PCT/CN2017/091603 2017-07-04 2017-07-04 Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle WO2019006649A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/091603 WO2019006649A1 (fr) 2017-07-04 2017-07-04 Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/091603 WO2019006649A1 (fr) 2017-07-04 2017-07-04 Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle

Publications (1)

Publication Number Publication Date
WO2019006649A1 true WO2019006649A1 (fr) 2019-01-10

Family

ID=64949538

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/091603 WO2019006649A1 (fr) 2017-07-04 2017-07-04 Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle

Country Status (1)

Country Link
WO (1) WO2019006649A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2577055A (en) * 2018-09-11 2020-03-18 Samsung Electronics Co Ltd Improvements in and relating to telecommunication networks
EP4035451A4 (fr) * 2019-09-25 2022-09-21 Telefonaktiebolaget LM Ericsson (publ) Premier noeud, quatrième noeud et procédés mis en oeuvre au moyen de ces noeuds pour gérer le dimensionnement d'une tranche de réseau dans un réseau de communication
CN115473821A (zh) * 2021-06-11 2022-12-13 中国移动通信集团广东有限公司 网络容量预测方法、装置、电子设备及存储介质
US11558263B2 (en) 2020-07-10 2023-01-17 Hewlett Packard Enterprise Development Lp Network device association with network management system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105262664A (zh) * 2014-06-09 2016-01-20 诺基亚通信公司 用于在通信网络中使用的虚拟化网络功能的控制
EP3021521A1 (fr) * 2014-11-11 2016-05-18 Alcatel Lucent Procédé et système de mise à l'échelle, réseau de télécommunication et produit de programme informatique
US20170019302A1 (en) * 2015-07-13 2017-01-19 Telefonaktiebolaget L M Ericsson (Publ) Analytics-driven dynamic network design and configuration
US20170048308A1 (en) * 2015-08-13 2017-02-16 Saad Bin Qaisar System and Apparatus for Network Conscious Edge to Cloud Sensing, Analytics, Actuation and Virtualization
US20170126792A1 (en) * 2015-11-02 2017-05-04 Telefonaktiebolaget L M Ericsson (Publ) System and methods for intelligent service function placement and autoscale based on machine learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105262664A (zh) * 2014-06-09 2016-01-20 诺基亚通信公司 用于在通信网络中使用的虚拟化网络功能的控制
EP3021521A1 (fr) * 2014-11-11 2016-05-18 Alcatel Lucent Procédé et système de mise à l'échelle, réseau de télécommunication et produit de programme informatique
US20170019302A1 (en) * 2015-07-13 2017-01-19 Telefonaktiebolaget L M Ericsson (Publ) Analytics-driven dynamic network design and configuration
US20170048308A1 (en) * 2015-08-13 2017-02-16 Saad Bin Qaisar System and Apparatus for Network Conscious Edge to Cloud Sensing, Analytics, Actuation and Virtualization
US20170126792A1 (en) * 2015-11-02 2017-05-04 Telefonaktiebolaget L M Ericsson (Publ) System and methods for intelligent service function placement and autoscale based on machine learning

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2577055A (en) * 2018-09-11 2020-03-18 Samsung Electronics Co Ltd Improvements in and relating to telecommunication networks
GB2577055B (en) * 2018-09-11 2021-09-01 Samsung Electronics Co Ltd Improvements in and relating to telecommunication networks
EP4035451A4 (fr) * 2019-09-25 2022-09-21 Telefonaktiebolaget LM Ericsson (publ) Premier noeud, quatrième noeud et procédés mis en oeuvre au moyen de ces noeuds pour gérer le dimensionnement d'une tranche de réseau dans un réseau de communication
US11558263B2 (en) 2020-07-10 2023-01-17 Hewlett Packard Enterprise Development Lp Network device association with network management system
CN115473821A (zh) * 2021-06-11 2022-12-13 中国移动通信集团广东有限公司 网络容量预测方法、装置、电子设备及存储介质
CN115473821B (zh) * 2021-06-11 2023-09-08 中国移动通信集团广东有限公司 网络容量预测方法、装置、电子设备及存储介质

Similar Documents

Publication Publication Date Title
US20210326128A1 (en) Edge Computing Platform
US10956849B2 (en) Microservice auto-scaling for achieving service level agreements
US10671368B2 (en) Automatic creation of delivery pipelines
US11455573B2 (en) Data protection distributed learning
WO2019006649A1 (fr) Procédé et dispositif pour une capacité de fonction réseau et une gestion d'échelle
US10891560B2 (en) Supervised learning system training using chatbot interaction
US10756976B2 (en) Data network and execution environment replication for network automation and network applications
US20200019841A1 (en) Neural network model for predicting usage in a hyper-converged infrastructure
US11381463B2 (en) System and method for a generic key performance indicator platform
US11722371B2 (en) Utilizing unstructured data in self-organized networks
US10592385B1 (en) Performance engineering
US10587490B2 (en) Evaluating resource performance from misaligned cloud data
US11297564B2 (en) System and method for assigning dynamic operation of devices in a communication network
CN107277100B (zh) 用于近实时云基础设施策略实现和管理的系统和方法
US11310125B2 (en) AI-enabled adaptive TCA thresholding for SLA assurance
Dey et al. Implementing deep learning and inferencing on fog and edge computing systems
Volkov et al. SDN load prediction algorithm based on artificial intelligence
JP2017530482A (ja) 計算リソースの新たな構成を決定するための構成方法、機器、システム及びコンピュータ可読媒体
US11237881B2 (en) Message connector as a service to migrate streaming applications into cloud nativity
US20210064981A1 (en) Controlling performance of deployed deep learning models on resource constrained edge device via predictive models
US20220114019A1 (en) Distributed resource-aware training of machine learning pipelines
Lanciano et al. Predictive auto-scaling with OpenStack Monasca
US10715407B2 (en) Dispatcher for adaptive data collection
US10057327B2 (en) Controlled transfer of data over an elastic network
US11676574B2 (en) Duration based task monitoring of artificial intelligence voice response systems

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17917104

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17917104

Country of ref document: EP

Kind code of ref document: A1