GB2541939A - Evaluation of network condition - Google Patents

Evaluation of network condition Download PDF

Info

Publication number
GB2541939A
GB2541939A GB1515766.2A GB201515766A GB2541939A GB 2541939 A GB2541939 A GB 2541939A GB 201515766 A GB201515766 A GB 201515766A GB 2541939 A GB2541939 A GB 2541939A
Authority
GB
United Kingdom
Prior art keywords
reliability
measurements
monitoring
network
resource
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1515766.2A
Other versions
GB201515766D0 (en
GB2541939B (en
Inventor
Ntofon Okung
Shakya Siddhartha
Owusu Gilbert
Malpass Jonathan
Mccormick Alistair
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
British Telecommunications PLC
Original Assignee
British Telecommunications PLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by British Telecommunications PLC filed Critical British Telecommunications PLC
Priority to GB1515766.2A priority Critical patent/GB2541939B/en
Publication of GB201515766D0 publication Critical patent/GB201515766D0/en
Publication of GB2541939A publication Critical patent/GB2541939A/en
Application granted granted Critical
Publication of GB2541939B publication Critical patent/GB2541939B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/04Processing captured monitoring data, e.g. for logfile generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/142Network analysis or design using statistical or mathematical methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/5041Network service management, e.g. ensuring proper service fulfilment according to agreements characterised by the time relationship between creation and deployment of a service
    • H04L41/5051Service on demand, e.g. definition and deployment of services in real time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/508Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement
    • H04L41/5096Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement wherein the managed service relates to distributed or central networked applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/12Network monitoring probes

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Mathematical Physics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Pure & Applied Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Algebra (AREA)
  • Health & Medical Sciences (AREA)
  • Cardiology (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

Information received from disparate monitors 200-202 concurrently measuring a property of a resource 240 in network 24 is compared in a reliability computation engine 22. Active probe 200 actively polls live data related to performance, passive monitor 201 acts on live data without polling and estimation monitor 202 derives estimated network performance from modeling. Preferably the active probe 200 is selected as the reference measurement. A metric of the degree of similarity between the measurements is computed and a measure of reliability is determined. The difference values may be weighted according to consistency with previous measurements. The reliability information may be used by a provisioning engine 28 to select or reject individual resources for use in meeting service requirements on the basis of the reliability of the reports of their performance, as well as the reported performance itself. Monitors identified as unreliable can also be reported to a fault diagnosis function 27.

Description

EVALUATION OF NETWORK CONDITION
This invention relates to telecommunications, and in particular to the evaluation of the condition of a telecommunications network infrastructure.
In information/communications technology (ICT), utility computing relies on on-demand provisioning of resources to deliver software, platforms, and networked infrastructures to customers, typically on a pay-as-you-use model. This concept has been facilitated by the emergence of the cloud computing paradigm, and the fastest-growing segment of the public cloud services market deals with the on-demand provisioning of networked infrastructures, known as Infrastructure-as-a-Service (laaS). To successfully deliver Infrastructure-as-a-Service with high Quality-of-Service (QoS), infrastructure providers must deploy intelligent and cost-effective monitoring frameworks that ensure accurate and up-to-date visibility of operational properties of the resources in the underlying cloud infrastructure i.e. compute, storage, and network resources.
This invention applies to the deployment of disparate monitoring technologies and approaches in a networked infrastructure. This is a key research and innovation area as the Internet and other networked applications and services have become increasingly important in our everyday activities.
It is common practice for infrastructure providers to deploy a number of disparate monitoring technologies to observe the same properties of a single resource type. There are several reasons for this. Firstly, the underlying cloud infrastructure may have resources from multiple vendors with customized implementations of standardized monitoring protocols or with proprietary monitoring technologies. Secondly, it allows for a more robust monitoring framework to be deployed, such that the infrastructure provider always has a means to capture an up-to-date view of the operational properties of its resources. Thirdly, it ensures a cost-effective monitoring framework, as one monitoring technology could be more accurate compared to other deployed technologies, but also more operationally expensive, and therefore deployed less widely.
The deployment of a number of disparate monitoring technologies to observe the same properties of a single resource type could be based on a simultaneous configuration where each technology is operational at the same time. It could also be based on a complementary configuration where each technology is operational in different ways. Infrastructure providers tend to deploy a combination of active monitoring where probes actively poll resources for information, passive monitoring where information is obtained by monitoring the applications running on the resources in the infrastructure, and through estimations that are based on theoretical models. Infrastructure providers receive the most accurate information from active monitoring approaches; however, this is the most expensive approach, and also the most intrusive approach as the monitoring probes tend to compete with customers for resources e.g. monitoring traffic competing with customer traffic for capacity in network links.
This invention is particularly focused on determining the accuracy of information generated by disparate monitoring technologies that are observing the same operational properties of resources in the networked infrastructure. It ensures the service provisioning logic only attempts to deliver services from the pool of resources which have accurate and up-to-date information of operational properties.
It is known to synchronise the operation or deployment of active network probes to avoid interference between them, and to balance the use of active and passive monitoring in such a way that the intrusiveness of active monitoring is minimized, but the accuracy of measurements is not compromised.
However, these approaches are directed primarily to co-ordinating the number and activity of probes. There is therefore a requirement for infrastructure providers to validate the accuracy of the information received from active, passive, and estimation approaches. This becomes even more pertinent for deployments where these approaches are operating in a simultaneous configuration, as different measuring techniques may have different levels of accuracy or systematic errors.
The present invention is to validate the information that has been generated by disparate monitoring technologies, which have been deployed by an infrastructure provider to observe the same operational properties of resources. Instead of the measures of a given network element using the various techniques simply being collected and processed independently, they are collated into a single, more reliable measure. Our invention focuses on establishing the accuracy of the information generated by monitoring technologies and validating the trustworthiness of deployed monitoring probes.
The present invention is able to function in a network infrastructure that uses both active and passive monitoring, and also estimations based on theoretical models. It focuses on establishing the similarity between the information generated by disparate monitoring technologies as a means of validating the accuracy of the generated information and also as a means of updating the infrastructure provider’s view of reliable monitoring probes.
The present invention accordingly provides, in a first aspect, a networked component for comparing the information received from disparate monitoring technologies or approaches that are concurrently observing a property of a single resource in a network, to compute a metric of the degree of similarity between their measurements, and thus to determine a measure of the reliability of one or more of the individual monitors.
The invention can be integrated into any resource or service provisioning system such as a cloud service orchestrator, networked service broker, or network resource provisioning system. It exposes an intelligent framework that validates the accuracy of monitoring information by computing a metric called the Weighted Similarity Percentage (WSP); this metric is used to compare the information received from disparate monitoring technologies.
In this way the invention is enabled to determine the reliability of disparate monitoring technologies and also to validate the accuracy of information received from these technologies. The service provisioning logic uses the WSP to identify the most appropriate subset of resources in the networked infrastructure to deliver high-quality Infrastructure-as-a-Service to customers. The invention also exposes functionalities to periodically determine the trustworthiness of monitoring probes in the networked infrastructure. The state-of-the-art in the area of networked infrastructure monitoring is mainly focused on methods to optimize the operation of active monitoring of resources i.e. prevent collisions between active probes in the networked infrastructure, reduce the overhead of active probing by leveraging more sophisticated passive monitoring techniques, and methods for intelligently and dynamically distributing active probes across the cloud infrastructure.
The present invention accordingly provides, in a second aspect, a method for comparing the information received from disparate individual monitors that are concurrently measuring a predetermined property of a predetermined resource in a network, to compute a metric of the degree of similarity between their measurements, and thus to determine a measure of the reliability of one or more of the individual monitors. In particular, it may be used to select resources for use according to the predetermined properties, by selecting resources for which measurements have been made with a predetermined level of reliability.
The present invention accordingly provides, in a third aspect, a computer program element comprising computer program code to, when loaded into a computer system and executed thereon, cause the computer to perform the steps of the method set out above. A preferred embodiment of the present invention will now be described, by way of example only, with reference to the accompanying drawings, in which:
Figure 1 is a block diagram of a computer system suitable for the operation of embodiments of the present invention;
Figure 2 is a schematic diagram depicting a cloud networked infrastructure incorporating a network condition evaluation system for operation according to an embodiment of the invention
Figure 3 is a schematic diagram illustrating some of the functional elements of the network condition evaluator of Figure 2 in more detail.
Figure 4 is a flowchart giving a High-level representation of a service provisioning process operating according to an embodiment of the invention
Figure 5 is a flowchart giving a High-level representation of the process operated by the reliability Computation Engine as part of the process of Figure 4
Figure 6 is a flow chart illustrating a process for determining a consistency measure
Figure 7 is a flowchart illustrating a process for assigning a normalised weighting value
Figure 1 is a block diagram of a computer system suitable for the operation of embodiments of the present invention. A central processor unit (CPU) 102 is communicatively connected to a storage 104 and an input/output (I/O) interface 106 via a data bus 108. The storage 104 can be any read/write storage device such as a random access memory (RAM) or a non-volatile storage device. An example of a nonvolatile storage device includes a disk or tape storage device. The I/O interface 106 is an interface to devices for the input or output of data, or for both input and output of data. Examples of I/O devices connectable to I/O interface 106 include a keyboard, a mouse, a display (such as a monitor) and a network connection.
Figure 2 depicts a high-level architectural view of a networked infrastructure (indicated generally at 24), with individual components 240, 241, 242 ...........etc, each monitored by one or more probes or monitors 200, 201, 202, 203, 204 ............etc.
These probes use a variety of operating technologies (also hereinafter referred to as monitoring, or probe, technologies) and, in general, not all available probe technologies will be used at each element of the infrastructure. A Cloud Services Orchestrator 20 is used to provision computing, storage, and network resources for the delivery of Infrastructure-as-a-Service (laaS) in response to requests delivered through a customer interface 29 from individual customers 25. A network condition evaluation system 2 operable according to this embodiment of the invention is integrated into the Cloud Services Orchestrator 20. The network condition evaluation system 2 provides the Cloud Services Orchestrator 20 with a unified interface to various monitoring probes 200, 201, 202 etc., deployed in the networked infrastructure 24. Its main components are a Registry 21, a Reliability Computation Engine 22, and a Probe Trust Engine 23.
The Registry 21 provides persistent storage of information generated by the disparate monitoring technologies. This information is queried by the Reliability Computation Engine 22 for service provisioning, and by the Probe Trust Engine 23 for periodic validation of the monitoring probes.
The Reliability Computation Engine 22 contributes towards service provisioning by determining whether individual monitoring probes e.g 200, 201, 202, 203.....are generating accurate information about resource properties. It also houses logic that is used to generate the most appropriate subset of resources which the service provisioning logic considers for the delivery of end-to-end services that satisfy customer requirements.
Figure 3 illustrates the Network Condition Evaluation System 2 and, in particular, the Reliability Computation Engine 22 thereof in more detail. It should be noted that the various components depicted are functional elements, which may be embodied in software. The Reliability Computation Engine 22 comprises a number of measurement interfaces 220, 221, 222, 223, 224 etc., corresponding with the individual monitoring probes 200, 201, 202, 203, 204, etc. Although depicted as separate items, it will be appreciated that this is a functional distinction, and in practice communication with the different probes may be operated by fewer interfaces (or indeed just one) operating in a time division (or other multiple access) mode. It will be noted that some of the interfaces 220, 223 are two-way as they co-operate with active probes, that is to say the evaluator polls the relevant probes, whereas other interfaces 221 passively receive data from the network. Estimating inputs (e.g 222, 224) take readings from a network model 202, 204 associated with the respective network resources (as depicted schematically in Figure 2) but not necessarily located therein.
The various inputs are compared in a comparison engine 228, to identify differences between the measured properties. These values are then averaged in a weighted average calculator 229, using weightings retrieved (step 40) from the registry 21.
The Cloud Services Orchestrator 20 (shown in Figure 2) processes customer requests for Infrastructure-as-a-Service (laaS) received through an interface 29, identifies the resource types relevant to the request and determines the QoS requirements specified by the customer 24. For example, a customer request for Network-as-a-Service will require network resources such as routers and switches, and QoS requirements such as loss rate and latency on the network links interconnecting the routers and switches of the required connections. The resource types and QoS requirements for the customer request are forwarded to the network condition evaluation system 2. The network condition evaluation system uses the logic in the Reliability Computation Engine 22 to query the Registry 21 for information on the operational properties of resources matching the resource types that are specified in the customer request. This query is restricted to the properties which are specified by the customer in its QoS requirements e.g. loss rate and latency of the network links interconnecting routers and switches.
To facilitate service provisioning, the Reliability Computation Engine 22 identifies which resources have reliable information stored in the Registry 21, and computes a Weighted Similarity Percentage (WSP) which is an indication of the similarity between the information generated by disparate monitoring technologies 200, 201,202 that are observing the same property of a resource 240. High WSP values indicate the disparate monitoring technologies have generated similar observations and the Reliability Computation Engine considers such observations to be accurate. Any resources with low WSP values are considered to be ineligible for delivering the service to the customer. The resources that satisfy both reliability and QoS constraints are forwarded to the service provisioning logic 28 which attempts to set up an end-to-end service from this pool of resources. The acceptance limit of WSP values can be defined by the infrastructure provider to ensure it always delivers high quality services. It can also be specified by customers with other QoS requirements. For example, a customer request could specify a minimum WSP of 80% across the disparate monitoring technologies with QoS requirement of at least 20Mbps free capacity on network links interconnecting routers and switches.
The Probe Trust Engine 23 performs a periodical update (431, 531, 56) of the records of trustworthy monitoring probes in the networked infrastructure. This component of the network condition evaluation system 2 houses a training algorithm that is used to determine how often each monitoring probe has generated inaccurate information. This trend is used to dynamically optimize the computation of WSP values by assigning a low weighting to values delivered by unreliable monitoring probes until such probes become more trustworthy. When probes are identified as untrustworthy, the probe trust engine 23 can also generate reports 270 for transmission to a fault diagnosis processor 27 to analyse the nature of the anomalous measurements or prompt the infrastructure provider to investigate untrustworthy monitoring probes for possible faults or damages.
The flowchart depicted in Figure 4 presents a high-level view of the service provisioning procedure with the network condition evaluation system 2 co-operating with the service provisioning engine 28.
Customer requests are received at the interface 29 and forwarded to the provisioning engine 28, which processes the customer request and identifies the computing, storage and network resource types needed to deliver the requested service (step 31). The Cloud Services Orchestrator 20 then queries the Network Condition Evaluation System 2 for an up-to-date view of the compute, storage and network resources of the networked infrastructure 24 (step 32).
The network condition evaluator/evaluation system 2 then identifies a subset of the available resources that will meet the user’s requirements (step 33). This step is depicted in more detail in Figure 5.
In identifying which resources are suitable, a measure of the reliability of the measurements made by the probes of resources parameters, referred to herein as the Weighted Similarity Percentage (WSP), is computed. This calculation is performed as follows, using the following parameters and variables: T is the set of reporting periods during which monitoring probes generate and store information P is the set of disparate monitoring probes deployed in the networked infrastructure R is the set of compute, storage, and network resources in the networked infrastructure K is the subset of all resources i.e. κ c r that have satisfied the reliability acceptance constraint
SPiJrt is the 2-combination similarity percentage; it is a measure of the similarity between the information generated by a pair of disparate probes (ij) e P both monitoring a resource r eR at a reporting period teT wsprt is the weighted similarity percentage computed for the information received by the disparate monitoring technologies for resource refiat reporting period t e τ β is the sensitivity parameter; it is the upper limit requirement for the operational property specified in customer request as QoS requirement e.g. the maximum value for acceptable latency on network links is 5ms. ψ is the reliability acceptance limit; it is the lower limit requirement for acceptable WSP values 9irt is the information generated by monitoring probe i e P for resource r e R at reporting period ter nirt is the inverse of the information generated by monitoring probe i eP for resource ter at reporting period ter ωί;· is the accuracy weight for the pair of monitoring probes (/J) e P that is used in conjunction with the 2-combination similarity percentage SPijrt to compute the weighted similarity percentage parameter wsprt. This parameter is periodically updated by the Probe Trust Engine with stronger weights being assigned to pairs of monitoring probes (ij) e P that generate similar and accurate information about resources. A pseudocode representation of the algorithm that is implemented in the Reliability Computation Engine 22 is detailed in the following steps, depicted in Figure 5:
Step 331: Process the customer request received at time t to identify required resources r efi, set the sensitivity parameter β, and set the reliability acceptance limit ψ
Step 332: For each monitoring probe i e P and each resource specified in the customer request ter, retrieve from the Registry 21 the information generated for the operational property relevant to the requirements of the customer at reporting time t, i.e. set the parameter θιη
Step 333: For each parameter 0irt that is set in the preceding step, compute the inverse at reporting time t using the expressions in Equation 1 below. This property is the difference between the actual value and the maximum acceptable value for the property in question - in other words the margin by which the observed value falls within the acceptable limit. The purpose of this step is two-fold; it enables the invention to filter resources that do not satisfy customer QoS requirements (step 3330). It also further emphasizes any differences that exist between information generated by disparate monitoring probes; particularly as the order of magnitude of generated information approaches the sensitivity parameter β
(1)
Step 334: For each resource r e R and for each pair of disparate monitoring probes (ij) e P that have generated information for this resource, compute the 2-combination similarity percentage at reporting time t using the expressions in Equation 2
(2)
Step 335: For each resource r e R, compute the Weighted Similarity Percentage over the set of disparate monitoring probes at reporting time t using the expression in Equation 3; the parameter ωί; is periodically generated by the Probe Trust Engine 23 (step 50) using a training algorithm as will be described later with reference to Figure 7
(3)
Step 336: For each resource r eR, verify if the resource satisfies both reliability and QoS constraints
(4)
Step 337: If no resources satisfy the condition specified in Equation (4), the provisioning engine 28 is given a rejection instruction (step 3370). A simplified worked example of this process will be described later
If one or more resource does meet the requirements, the subset of eligible resources keK\K qr\s advised to the service provisioning logic 28 (step 3371) which attempts to deliver an end-to-end service to the customer from this pool of eligible resources (step 34)
The service provisioning logic 28 will then attempt to configure the selected resources 24 as required to fulfil the request (step 34). If it has received a rejection request 3370, or is otherwise unable to deliver the required service (step 35) it transmit a messages by way of the interface 29 to the requesting user 25, rejecting the request, or advising the user to modify the request, for example by reducing the bandwidth requirement, or specifying a different time for the service to be provided (step 350). If all the resources are available, the service can be delivered (step 351)
The accuracy weight parameter ωυ generated by the probe trust engine 23 for each pair of monitoring probes and provided to the service provisioning logic 28 is determined as follows, with reference to Figure 7, and using the following additional parameters: A is the training period: the Probe Trust Engine runs the training algorithm after every Λ reporting period has elapsed δ is the dissimilarity threshold, it is the maximum number of times the 2-combination similarity percentage SPijrt between monitoring probes i and j can be lower than the reliability acceptance limit ψ
Qij is a consistency indicator, indicative of the number of times the 2-combination similarity percentage SPiJrt between monitoring probes i and j is lower than the reliability acceptance limit ψ. This parameter is reset to zero after every Λ reporting period has elapsed
The pseudocode representation of the algorithm that is implemented in the Probe Trust Engine 23 to ensure an up-to-date view of monitoring probes generating information of the same operational property of a resource is detailed in the following steps, and illustrated in Figures 6 and 7.
The steps depicted in Figure 6 are performed periodically for each resource.
Step 41: Select the monitoring probe i e P that is expected to generate the most accurate information about the operational property of the resource. This monitoring probe is called the reference probe and in most cases is the probe that utilizes active monitoring
Step 42: For every reporting period teT within the interval between the penultimate and current training periods, compute the 2-combination similarity percentage SPijrt between the reference probe i and every other monitoring probe j e P that is generating information about the same operational property for the same resource
Step 43: For each pair of monitoring probes (ij) e P where i is the reference probe, check the number of times the 2-combination similarity percentage SPijrt is less than the reliability acceptance limit ψ i.e. set the consistency indicator ρ0·
The steps depicted in Figure 7 take place when a trustworthiness weighting is required. For each pair of monitoring probes (i,j) e P where i is the reference probe, retrieve the consistency indicator ρί;· (step 531) and check if it less than or equal to the dissimilarity threshold δ. (Step 54 )
If the parameter ρί7· is less than or equal to the dissimilarity threshold (step 541), the monitoring probe j is considered to have a high to medium accuracy depending on how close the value of parameter ρί7· is to the value of the dissimilarity threshold δ. If the value of the parameter ρί7· is greater than the value of the dissimilarity threshold δ, the monitoring probe is considered to have low accuracy and is therefore more untrustworthy.
Step 55: Generate the accuracy weight parameter ωί7· for each pair of monitoring probes (ij) e P where i is the reference probe. Weights are therefore assigned to pairs of monitoring probes (ij) e P that have generated the most similar information, and the most accurate in relation to the reference probe, within the training interval i.e. pairs of monitoring probes with the parameter ρί;· less than the dissimilarity threshold. The weightings are normalised (step 555) such that
(5)
These weightings can then be stored by the registry 21 (step 56), to be accessed by the provisioning process of Figure 6 as required (step 40). A simplified demonstration of the process implemented in the Probe Trust Engine 23 to determine the weightings to be applied to different probe measurements, and depicted in Figure 6 and 7 follows.
For the purposes of illustration, we assume the infrastructure provider has deployed six monitoring probes for monitoring the same properties of its resources, namely one active probe (A), four passive probes (P1, P2, P3, P4), and one estimation based probe (E). The active probe A is selected as the reference monitoring probe, as it is likely to be the most accurate. This is used to periodically update the view of the trustworthiness of the other monitoring probes in the infrastructure. Table B.1 shows the 2-combination similarity percentage for the information generated by the monitoring probes in 24 reporting periods. The Probe Trust Engine 23 is assumed to have the following configuration enabled for the training operation:
Retraining frequency every 12 reporting periods
Dissimilarity threshold <5 -> 6
Reliability acceptance limit ψ -> 75% or 0.75
Values below the reliability acceptance limit are highlighted in table B1
After the first twelve reporting periods T1 - T12, the Probe Trust Engine updates its view of the trustworthiness of the six monitoring probes using the two-combination similarity percentage values presented in Table B.1. (Step 43/431, Figure 6) For these periods, none of the pairs of monitoring probes exceeds the configured dissimilarity threshold value (in this example, the threshold value is six failures (out of twelve periods) to achieve the acceptance limit). The pair of monitoring probes (A-P1) generates the most values that are lower than the reliability acceptance limit. This causes the Probe Trust Engine to identify the monitoring probe Passive-1 as the most unreliable probe for the following reporting periods (T13 - T24).
The Probe Trust Engine now generates an accuracy weighting, as presented in Table B.2, to be used by the Reliability Computation Engine for service provisioning in the following reporting periods.
Table B.2 Accuracy weights generated by Probe Trust Engine Reporting period T1-T12 T13-T24
Active, Passive-1 (A-P1) 0.09 0.09
Active, Passive-2 (A-P2) 0.21 0.01
Active, Passive-3 (A-P3) 0.25 0.28
Active, Passive-4 (A,P4) 0.20 0.31
Active, Estimation (A-E) 0.25 0.31
The strength of the weights is dependent on the value of the parameter ρί;· for each pair of monitoring probes i.e. the number of times the two-combination similarity percentage between monitoring probes i and ; is lower than the reliability acceptance limit. Pairs of monitoring probes with lower values of the parameter ρί;· are considered to be more trustworthy and are assigned higher values for the accuracy weight parameter ωί7·.
For the reporting periods T1 - T12, the pairs of (A, P3) and (A, E) are assigned the strongest weights as these do not generate any two-combination similarity percentage values that are less than the reliability acceptance limit. The pairs of monitoring probes (A, P2) and (A, P4) generate one and three values of two-combination similarity percentage respectively that are less than the reliability acceptance limit.
During the next training period i.e. after reporting periods T13-T24, the Probe Trust Engine revaluates its view of the trustworthiness of the six monitoring probes. In this case, a pair of monitoring probes i.e. (A, P2) exceeds the dissimilarity threshold by generating ten values of the two-combination similarity percentage that are less than the reliability acceptance limit. As a result, the monitoring probe P2 is considered to be very untrustworthy by the Probe Trust Engine. The Probe Trust Engine generates the accuracy weights presented in this second reporting period for service provisioning in the following reporting periods i.e. T25 - T36. The pair of monitoring probes (A, P1) again generates six values that are lower than the reliability acceptance limit while the pair of monitoring probes (A, P3) generates one value lower than the reliability acceptance limit. The pairs of (A, P4) and (A-E) are assigned the strongest weights as these do not generate any two-combination similarity percentage values that are less than the reliability acceptance limit. The drastic change in the trustworthiness of the monitoring probe P2 could be due to a fault or damage and the probe trust engine 23 is arranged to detect such changes and report them to a fault diagnosis processor 27 (step 270). A simplified worked example of the process of Figure 4 to determine the similarity between probe readings will now be given.
Referring again to Figure 2, this depicts a simplified enterprise networked infrastructure with high performance routers 240-245 interconnected via high speed network links. Customers 25 request the delivery of Network-as-a-Service (NaaS) through a service provisioning system 28 co-operating with the network condition evaluation system 21
The details of the customer request are as follows:
Service requirement -> delivery of end-to-end connectivity between Node A (241) and Node D (245)
QoS requirement (sensitivity parameter β) maximum CPU utilization for each node in end-to-end path must be less than or equal to 75%
Reliability acceptance limit ψ -> 80%
It is assumed the infrastructure provider has deployed three monitoring technologies based on active and passive monitoring, and also theoretical estimations. The active monitoring probe is considered to be the most trustworthy and the accuracy weights ωί7· assigned for each pair of monitoring probes (i,j) e P are detailed below 0.8 for the pair of active-passive probes i.e. <U(acti„ejPassi1,e) 0.1 for the pair of active-estimation probes i.e. M(activeiestimation) 0.1 for the pair of passive-estimation probes i.e. o>(passiveiestimation)
Table A.1 shows the information generated by each monitoring probe for the nodes in the networked infrastructure. It also presents the result of the inverse computations using the customer specified sensitivity parameter/? = 75% (Step 333 of the Reliability Computation Engine process of Figure 5)
Table A.1 Information generated bv monitoring probes and inverse computations
\
Table A.2 Two-combination similarity percentage for pairs of monitoring probes and the weighted similarity percentages
Table A.2 shows the values obtained from the computation of the two-combination similarity percentage for the pairs of monitoring probes in the networked infrastructure using the inverse computations from Table A.1 (Step 334 of the process depicted in Figure 5). It also presents the weighted similarity percentage computed with the accuracy weights and two-combination similarity percentage for the pairs of monitoring probes (Figure 5, Step 335).
The invention considers the computed weighted similarity percentages presented in Table A.2 and generates the set of eligible resources for delivering the end to end service based on the customer specified reliability acceptance limit ψ = 80% i.e. 0.80 (Figure 5, Step 336).
The generated set of eligible resources is {Node A, Node B, Node D, Node E, Node F] as Node C does not satisfy the reliability acceptance constraint with a weighted similarity percentage value of 50%. The service provisioning logic attempts to deliver the end-to-end connectivity between nodes A and D by enabling only network links between nodes in this generated resource pool. The service provisioning logic assumed for this example is the selection of the shortest possible path between Nodes A and D. In this case, the invention will deliver NaaS to the customer with the network path Node A -Node E - Node F - Node D.
Insofar as embodiments of the invention described are implementable, at least in part, / using a software-controlled programmable processing device, such as a microprocessor, digital signal processor or other processing device, data processing apparatus or system, it will be appreciated that a computer program for configuring a programmable device, apparatus or system to implement the foregoing described methods is envisaged as an aspect of the present invention. The computer program may be embodied as source code or undergo compilation for implementation on a processing device, apparatus or system or may be embodied as object code, for example.
Suitably, the computer program is stored on a carrier medium in machine or device readable form, for example in solid-state memory, magnetic memory such as disk or tape, optically or magneto-optically readable memory such as compact disk or digital versatile disk etc., and the processing device utilises the program or a part thereof to configure it for operation. The computer program may be supplied from a remote source embodied in a communications medium such as an electronic signal, radio frequency carrier wave or optical carrier wave. Such carrier media are also envisaged as aspects of the present invention.
It will be understood by those skilled in the art that, although the present invention has been described in relation to the above described example embodiments, the invention is not limited thereto and that there are many possible variations and modifications which fall within the scope of the invention.
The scope of the present invention includes any novel features or combination of features disclosed herein. The applicant hereby gives notice that new claims may be formulated to such features or combination of features during prosecution of this application or of any such further applications derived therefrom. In particular, with reference to the appended claims, features from dependent claims may be combined with those of the independent claims and features from respective independent claims may be combined in any appropriate manner and not merely in the specific combinations enumerated in the claims.
Table B.1 Two-combination similarity percentage used bv Probe Trust Engine

Claims (14)

1. A method for comparing the information received from disparate individual monitors that are concurrently measuring a predetermined property of a predetermined resource in a network, to compute a metric of the degree of similarity between their measurements, and thus to determine a measure of the reliability of one or more of the individual monitors.
2. A method according to claim 1, wherein a reliability measurement is determined for a property of a resource by taking a reference measurement from a first monitor, taking one or more further measurements using further monitors, determining the difference between the reference measurement and each of the further measurements, and deriving a reliability value for the measured property of the resource from the said differences
3. A method according to Claim 2, wherein.the reliability value is derived from the difference values weighted according to a measure of consistency between a plurality of previous measurements
4. A method according to claim 3, wherein a training algorithm determines how often each monitoring probe has generated information that is inconsistent with measurements taken by other probes.
5. A method according to Claim 3 or Claim 4, wherein monitoring probes identified as providing such inconsistent data are reported to a fault diagnosis processor.
6. A method according to claim 1, in which resources are selected for use according to the measured properties, and wherein only measurements which are determined to have met a predetermined level of reliability are used for such selection.
7. A method according to any preceding claim, wherein the monitors include at least one active monitor, which actively polls resources for live data relating to their performance, and at least one passive monitor which acts on live data received from the network without such polling, wherein an active monitor is selected as the reference monitor.
8. A method according to any preceding claim, wherein the monitors include at least one live data monitor, which determines network performance from measurements received from network resources, and at least one estimation monitor which derives estimated network performance from network modelling data, wherein a live data monitor is selected as the reference monitor. o
9. A networked component (2) for comparing the information received from disparate monitoring technologies (200, 201, 202) or approaches that are concurrently observing a property of a resource (240) in a network (24), to compute a metric of the degree of similarity between their measurements, and thus to determine a measure of the reliability of one or more of the individual monitors.
10. A networked component according to claim 9, having a reliability computation engine (22) for determining a reliability measurement for a property of a resource, having measurement receivers (220, 221, 222, ...), by taking measurements from a plurality of monitors 200, 201,202..., a comparator (228) for determining the difference between a reference measurement and each of the further measurements, and a reliability value generator (229) for deriving a reliability value for the measured property of the resource from the said differences
11. A networked component according to claim 10, further comprising a consistency weighting generator (23) for generating a measure of consistency between a plurality of measurements
12. A networked component according to claim 11, wherein the weighting generator determines how often each of the monitors has generated information which is inconsistent with measurements taken by other monitors.
13. A networked component according to Claim 11 or Claim 12, further comprising an output 270 from the consistency weighting generator to a fault diagnosis processor.
14. A computer program element comprising computer program code to, when loaded into a computer system and executed thereon, cause the computer to perform the steps of a method as claimed in any of claims 1 to 7.
GB1515766.2A 2015-09-04 2015-09-04 Evaluation of network condition Active GB2541939B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB1515766.2A GB2541939B (en) 2015-09-04 2015-09-04 Evaluation of network condition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB1515766.2A GB2541939B (en) 2015-09-04 2015-09-04 Evaluation of network condition

Publications (3)

Publication Number Publication Date
GB201515766D0 GB201515766D0 (en) 2015-10-21
GB2541939A true GB2541939A (en) 2017-03-08
GB2541939B GB2541939B (en) 2018-04-18

Family

ID=54345832

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1515766.2A Active GB2541939B (en) 2015-09-04 2015-09-04 Evaluation of network condition

Country Status (1)

Country Link
GB (1) GB2541939B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5233542A (en) * 1989-08-29 1993-08-03 Nord-Micro Electronik Feinmechanik Ag Method and circuit configuration for forming an evaluation signal from a plurality of redundant measurement signals
US20060074496A1 (en) * 2004-09-30 2006-04-06 Taware Avinash V System and method for sensor validation and fusion
GB2428922A (en) * 2005-08-04 2007-02-07 Boeing Co Altitude measurement system for a ground effect vehicle
US20110213559A1 (en) * 2010-02-05 2011-09-01 Prima-Temp Inc. Multi-sensor patch and system
US20150148997A1 (en) * 2013-11-28 2015-05-28 Airbus Operations (S.A.S.) Method for fusing data from sensors using a consistency criterion

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5233542A (en) * 1989-08-29 1993-08-03 Nord-Micro Electronik Feinmechanik Ag Method and circuit configuration for forming an evaluation signal from a plurality of redundant measurement signals
US20060074496A1 (en) * 2004-09-30 2006-04-06 Taware Avinash V System and method for sensor validation and fusion
GB2428922A (en) * 2005-08-04 2007-02-07 Boeing Co Altitude measurement system for a ground effect vehicle
US20110213559A1 (en) * 2010-02-05 2011-09-01 Prima-Temp Inc. Multi-sensor patch and system
US20150148997A1 (en) * 2013-11-28 2015-05-28 Airbus Operations (S.A.S.) Method for fusing data from sensors using a consistency criterion

Also Published As

Publication number Publication date
GB201515766D0 (en) 2015-10-21
GB2541939B (en) 2018-04-18

Similar Documents

Publication Publication Date Title
US8630865B2 (en) Automated on-line business bandwidth planning methodology
US9729405B2 (en) Systems and methods for real-time service assurance
US10135698B2 (en) Resource budget determination for communications network
EP2409242B1 (en) Network status detection
CN112398680B (en) Fault delimiting method and equipment
US10153950B2 (en) Data communications performance monitoring
US8285841B2 (en) Service quality evaluator having adaptive evaluation criteria
EP3222004B1 (en) Diagnostic testing in networks
CN109005085A (en) A kind of service availability monitoring system, method, device and equipment
US20110075582A1 (en) Delay time measurement apparatus, storage medium storing delay time measurement program and network system
CN109617758B (en) Node network quality calculation method and device, server and computer storage medium
CN108924005A (en) Network detecting method, network detection device, medium and equipment
Clark et al. Secure monitoring of service level agreements
CN109921925A (en) A kind of dial testing method and device
US10320647B2 (en) Evaluation of network conditions
Gaglianese et al. Assessing and enhancing a Cloud-IoT monitoring service over federated testbeds
US20170302506A1 (en) Methods and apparatus for fault detection
GB2541939A (en) Evaluation of network condition
Rady Formal definition of service availability in cloud computing using OWL
Wagle Cloud service optimization method for multi-cloud brokering
US9311210B1 (en) Methods and apparatus for fault detection
Gol Mohammadi et al. Maintenance of Trustworthiness in a Cyber-Physical System and Preparation for Run-Time
Jeswani et al. Adaptive monitoring: A hybrid approach for monitoring using probing
JP2004241812A (en) Method for discriminating heavy load in packet switching network and apparatus thereof