WO2006121462A2 - Systems and methods for endoscope integrity testing - Google Patents

Systems and methods for endoscope integrity testing Download PDF

Info

Publication number
WO2006121462A2
WO2006121462A2 PCT/US2005/033418 US2005033418W WO2006121462A2 WO 2006121462 A2 WO2006121462 A2 WO 2006121462A2 US 2005033418 W US2005033418 W US 2005033418W WO 2006121462 A2 WO2006121462 A2 WO 2006121462A2
Authority
WO
WIPO (PCT)
Prior art keywords
network
routing
links
elements
network element
Prior art date
Application number
PCT/US2005/033418
Other languages
French (fr)
Other versions
WO2006121462A3 (en
WO2006121462A9 (en
Inventor
Melissa Kubach
Original Assignee
Verimetrix, Llc, A Missouri Limited Liability Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/123,336 external-priority patent/US20060252991A1/en
Priority claimed from US11/123,335 external-priority patent/US20060252990A1/en
Application filed by Verimetrix, Llc, A Missouri Limited Liability Corporation filed Critical Verimetrix, Llc, A Missouri Limited Liability Corporation
Priority to EP05801033.1A priority Critical patent/EP1887919B8/en
Priority to ES05801033.1T priority patent/ES2625927T3/en
Publication of WO2006121462A2 publication Critical patent/WO2006121462A2/en
Priority to AU2007240157A priority patent/AU2007240157A1/en
Publication of WO2006121462A9 publication Critical patent/WO2006121462A9/en
Publication of WO2006121462A3 publication Critical patent/WO2006121462A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/085Retrieval of network configuration; Tracking network configuration history
    • H04L41/0853Retrieval of network configuration; Tracking network configuration history by actively collecting configuration information or by backing up configuration information
    • H04L41/0856Retrieval of network configuration; Tracking network configuration history by actively collecting configuration information or by backing up configuration information by backing up or archiving configuration information
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01MTESTING STATIC OR DYNAMIC BALANCE OF MACHINES OR STRUCTURES; TESTING OF STRUCTURES OR APPARATUS, NOT OTHERWISE PROVIDED FOR
    • G01M3/00Investigating fluid-tightness of structures
    • G01M3/02Investigating fluid-tightness of structures by using fluid or vacuum
    • G01M3/26Investigating fluid-tightness of structures by using fluid or vacuum by measuring rate of loss or gain of fluid, e.g. by pressure-responsive devices, by flow detectors
    • G01M3/28Investigating fluid-tightness of structures by using fluid or vacuum by measuring rate of loss or gain of fluid, e.g. by pressure-responsive devices, by flow detectors for pipes, cables or tubes; for pipe joints or seals; for valves ; for welds
    • G01M3/2846Investigating fluid-tightness of structures by using fluid or vacuum by measuring rate of loss or gain of fluid, e.g. by pressure-responsive devices, by flow detectors for pipes, cables or tubes; for pipe joints or seals; for valves ; for welds for tubes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0806Configuration setting for initial configuration or provisioning, e.g. plug-and-play
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0805Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
    • H04L43/0817Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking functioning

Definitions

  • Communications networks such as next generation broadband networks
  • networks have become increasingly complex due to increased size, numerous intermixed technologies/protocols (e.g., ATM, Frame Relay, etc.), and the intermixing of equipment manufactured by numerous different vendors.
  • network configuration management systems that can provision virtual trunks and circuits within these networks are becoming increasingly important.
  • Such network configuration management systems function to determine the paths/routes between network equipment, herein referred to as network elements, and to communicate with those network elements to realize desired trunks or circuits that facilitate the transmission of traffic across the network.
  • network configuration management systems have traditionally determined the paths available by modeling portions of network elements as nodes on a graph and the links/interconnections between these portions as links between the nodes. More particularly, prior systems typically modeled every port of every network element as a node on the graph and modeled every physical link that interconnected these ports to one another as links that interconnected the nodes of the graph. The network model was then used to provision virtual trunks, which formed paths between network elements in the network. Once these virtual trunks were provisioned, virtual circuits could then be established across these trunks to support traffic flow from one point to another in the network.
  • FIG. 1 shows an exemplary prior art network configuration management system 102 and a network 110 managed by system 102.
  • the network configuration management system 102 functions to determine a preferred path between two points in a network (i.e., between two network elements) and for provisioning a communications connection across this path by communicating with the managed network 110.
  • Managed network 110 consists primarily of broadband network 112 which, in turn, consists of a plurality of network elements 114-118 interconnected by physical links and virtual trunks and circuits represented in FIG. 1 by links 120-124.
  • the network elements comprise varying technologies and protocols and may be manufactured by different vendors.
  • Managed network 110 further comprises network management systems, such as network management system (NMS) 126, and element management systems, such as element management system (EMS) 128.
  • NMS network management system
  • EMS element management system
  • NMSs and EMSs may function to control both the network elements and the links between those elements. However some may not control the links between the elements and, instead, only manage the network elements themselves.
  • an NMS such as NMS 126
  • NMS 126 may function to collectively manage a set of network elements 114 and the physical links 120 between them, thus forming a collectively managed sub-network having network elements 114. Accordingly, when network traffic arrives at an ingress port into one of the network elements 114, such as port 130, the NMS 126 determines a set of links and network element cross- connects to interconnect port 130 to an egress port, such as port 132.
  • the NMS 126 then provisions the network elements to realize this interconnection.
  • some management systems such as EMS 128, may only manage one or more network elements 118, but not the links 124 between them.
  • a higher layer entity such as the Network Configuration Management System 102, determines the links between network elements 118 required to create a path and then instructs the EMS to perform the necessary cross-connects within network elements 118 to realize the complete path.
  • FIG. 1 also shows how some network elements, such as network elements 116, are not managed by either an NMS or EMS.
  • a higher layer entity once again such as Network Configuration Management System 102, directly communicates with these elements to perform network configuration functions.
  • Network Configuration Management System 102 would configure any cross-connects within network elements 116 as well as any links between network elements.
  • the combination of Network Configuration Management System 102, NMS 126 and EMS 128 will collectively determine an appropriate network path across and between network elements 114, 116 and then provision virtual trunks and circuits across network 112.
  • a simplified routing graph was created by the network configuration management system whereby, instead of modeling each port of a network element as a node on a routing graph, an entire network element itself could be represented as one or more routing nodes or, in some cases, multiple network elements could be represented as a single routing node.
  • network elements 114 of FIG. 1 that are managed by NMS 126 are modeled as a single node 201.
  • network elements 118, which are managed by both EMS 128 and the Network Configuration Management System 102 are also modeled as a single routing node 204.
  • Network elements 116 are each modeled as individual routing nodes, since the Network Configuration Management System 102 manages both the network element and the link between the elements.
  • a network provisioning system for establishing a path between two networks wherein a common network device between those networks is modeled as a link between a first network element in one network and a second network element in a second network.
  • a network routing graph is created by an inventory subsystem in a routing manager by inventorying the physical network elements and links in the network. The inventory subsystem then models those elements/links as a plurality of virtual nodes and links between the nodes.
  • At least one common network device such as a digital cross connect located at a junction between the two networks, is modeled as a link instead of a node.
  • a routing engine then uses the network routing graph, including the link modeled from the common network device, to provision a path between the two networks.
  • route processing is reduced, resulting in a corresponding reduction in overhead and resources required to route network traffic from one node to another.
  • FIG. 1 shows a prior art managed broadband network and a network configuration management system for determining and provisioning routing paths within the network
  • FIG. 2 shows a network routing model whereby some network elements are combined and treated as single routing nodes;
  • FIG. 3 shows an illustrative network configuration management system
  • FIG. 4 shows a network routing model whereby Digital Cross Connect
  • DCSs are used to interconnect different network nodes.
  • FIG. 5 shows a network routing model in accordance with the principles of the present invention whereby DCSs are modeled as links.
  • FIG. 3 shows an illustrative network configuration management system, such as Network Configuration Management System (NCMS) 102 in FIGs. 1 and 2.
  • NCMS 102 determines preferred routing paths between two ports within the network by modeling the network paths as a plurality of routing nodes and links between the nodes, and for using these paths to provision virtual trunks and circuits within the networks.
  • NCMS 102 includes, among other components, a routing manager 304 and inventory database 322.
  • the routing manager 304 provides end-to-end connection management functions including the determination and provisioning of routing paths in broadband network 112 in FIG. 1.
  • routing manager 304 comprises an inventory subsystem 306, a routing engine 308 and a service activation system 310.
  • the routing manager 304 is connected to the various network elements via an element adapter 312 and connection 313. Broadly, the routing manager 304 maintains a topological graph comprising nodes and links that model the broadband network 112. This graph is used to determine and provision routing paths between, for example, two ports within the network. These paths are then used to provision virtual trunks and circuits.
  • the inventory subsystem 306 builds and maintains the topological graph in accordance with modeling methods such as those described above in association with the '187 application.
  • This graph is maintained, illustratively, in three database tables: routing link table 314, routing node table 316, and NMS/EMS table 318.
  • the routing engine 308 determines a routing path for traffic through the network using the network graph maintained by the inventory subsystem 306.
  • the service activation system 310 uses the determined routing path to provision the actual virtual trunk or virtual circuit. Specifically, the service activation system 310 activates the routing engine 308 to obtain a routing path given two endpoints and then invokes the element adapter 312 which interfaces with network elements, NMSs and EMSs to physically provision the determined path.
  • the element adapter 312 functions as an interface between the routing manager 304 and the NMSs 126, EMSs 128, and network elements 116 in managed broadband network 112.
  • the service activation system determines a routing path, it invokes the appropriate adapter(s) or adapter module(s) to communicate the required configuration settings to the management systems/elements 126, 128, and 116 to provision the determined path.
  • network traffic may be required to traverse multiple separate networks. These different networks may be interconnected with cross connects, such as digital cross connects (DCSs). As such, it is necessary for the NCMS 102 to also have available configuration and status information related to these DCSs. This configuration and status information is, illustratively, maintained in cross-connection status database 324. Thus, in provisioning the aforementioned path, service activation system 310 may also refer to cross-connection status database 324. [00018] The prior illustrative method described in the '187 application for using an NMS to simplify routing graphs is advantageous in many regards.
  • OSI Open System Interconnection
  • Layer 2 is also known as the data link layer and is the layer at which the physical medium is shared and where data link and media access are controlled.
  • OSI Open System Interconnection
  • Layer 2 is the layer at which network routing between media access control (MAC) addresses of individual hardware components is performed.
  • MAC media access control
  • DCSs or other similar devices, such as optical cross connect systems (OCSs).
  • OCSs optical cross connect systems
  • a DCS is any device that interconnects networks to facilitate traffic routing from one network to another or to link portions of networks using one protocol or traffic rate to another portion using a different protocol or rate.
  • DCSs are very well known in the art and serve to efficiently manage disparate traffic protocols and line speeds in telecommunications system central offices as well as remote field locations and other locations such as within hotels and even at user premises.
  • DCSs may be used in a variety of different applications.
  • DCSs may be used at a customer premises to interface with both voice protocol networks and a number of different data protocol networks in order to aggregate and cross connect these networks to a high-speed copper wire or optical fiber loop.
  • DCSs may be used in a digital loop carrier (DLC) capacity to aggregate networks using multiple protocols for transmission across a SONET ring network.
  • DLC digital loop carrier
  • DCSs may be used within, illustratively, a telecommunications central office in order to manage and cross connect channels from multiple SONET rings that are employed in an access network and/or a metro or inter-office network.
  • Other uses of DCS are well known and will be obvious to one skilled in the art.
  • FIG. 4 shows one illustrative routing map wherein DCSs are used to connect networks to facilitate traffic flow from one network to another.
  • FIG. 4 shows routing nodes 401-406, each of which represents, illustratively, a network, such as broadband network 112, or a portion of a network, such as the group of network elements 116 also in FIG. 1.
  • each of the routing nodes 401-406 illustratively has a plurality of network elements that are modeled, for routing purposes, as a single routing node with an ingress port and an egress port, such as ports 130 and 134, respectively, in FIG. 1.
  • the networks represented by each of routing nodes 401-406 may, for example, operate using a different protocol or speed and, therefore, DCSs, such as DCSs 407, 408 and 409, may be used to aggregate and/or disaggregate traffic in order to facilitate the transmission of that traffic between and over the different networks interconnected by the DCSs.
  • DCSs such as DCSs 407, 408 and 409
  • routing nodes 402 and 405 may represent well-known OC-3 networks operating at an illustrative speed of 155.52 Mb/s while the networks represented by nodes 403 and 406 may be well-known OC-12 networks operating at an illustrative 622.08 Mb/s rate.
  • DCSs 407 and 408 aggregate and/or disaggregate the data between the networks represented by nodes 402 and 403 and DCS 409 aggregates and disaggregates the traffic between the networks represented by nodes 405 and 406.
  • paths through DCSs 407-409 are provisioned in a relatively static manner. For example, a path from port 407A, associated with node 402, to port 407B, associated with node 403, is provisioned on DCS 407 in order to provide connectivity between the networks represented by nodes 402 and 403. Connections between ports 408A/408B and 409A/409B are similarly provisioned to connect nodes 402/403 and 405/406, respectively.
  • DCSs 407-409 function as common nodes between the respective networks. [00022]
  • a DCS such as any one of
  • DCSs 407-409 functions similarly in some respects to a network switch, such as a router or ATM switch.
  • a network switch such as a router or ATM switch.
  • routers/switches typically operate as a function at least in part of the signaling accompanying traffic transiting the network and, hence, such routers/switches are typically closely tied to specific services provided by a network service provider.
  • a DCS is typically not used for such purposes. Instead, a DCS is typically used for transmission management at a higher level of the network.
  • DCSs are typically used as an engineering and provisioning control mechanism at layer 1 in the network (i.e., the physical layer of the network).
  • DCSs are typically not used to dynamically alter switching over a short time period, as are routers and other types of switches. Additionally, DCSs are not typically controlled as a function of signaling from a customer but are, instead, controlled directly by, for example, engineers at the service provider. Also unlike simpler network switches, a typical DCS facilitates the provisioning of network paths and connections across the DCS that are typically constant over a period of hours to months.
  • the present inventor have further discovered that, in addition to simplifying routing decisions at layer 2 of a network, as described in the '187 application, it is desirable to also simplify the routing graph at layer 1 of the network. Specifically, instead of treating DCSs as a separate node (or multiple nodes corresponding the ports on the DCS) in the network, it is also desirable to model DCSs differently in order to further simplify the routing graphs/decisions. More particularly, in part since DCSs and other similar devices are relatively static in configuration, the present inventor have discovered that such devices may be treated as links, such as would be formed by a physical cable, instead of nodes that require processing as an affirmative routing hop.
  • a network configuration management system such as NCMS 126 in FIG. 1 will inventory the network elements and links in the network. Once these elements and links are defined, the NCMS generates a routing graph showing the network topology in terms of routing nodes and links to be used in provisioning trunks/circuits across the network.
  • FIG. 5 is a simplified representation of such a routing map. In particular, routing nodes 401-406 are as described above in association with FIG. 4.
  • Each of those routing nodes consists, for example, of a plurality of network elements that are modeled at a high level as a single routing node in order to decrease the processing overhead required to provision the aforementioned trunks/circuits.
  • those illustrative DCSs are modeled as links 501 , 502 and 503. Links 501 , 502 and 503 are used in the routing graph of FIG. 5 to represent DCSs 407, 408 and 409, respectively.
  • a cross-connect such as a DCS
  • a cross-connect is not modeled as one or more routing nodes having various links connecting ports to each other and to external ports on other network elements.
  • such a cross-connect is modeled as a separate link between network elements in one or more networks. Accordingly, the routing map is greatly simplified.
  • the NCMS will inventory the network elements and links between the elements, treating DCSs as links as described above. Specifically, this inventory is conducted by the inventory subsystem 306 of FIG. 3. As a part of this inventory, routing link table 314, routing node table 316, NMS/EMS table 318 and cross connection status database 324 are updated with information about the links, nodes and cross connections in and between the networks managed by the NCMS 102. Therefore, in this inventory, information concerning each DCS will be updated in the cross-connection status database and those same DCSs will be updated as links in the routing link table.
  • service activation system 310 invokes the routing engine 308 to provision a path
  • that engine will treat the DCSs as links to be provisioned and not one or more network nodes corresponding to the ports on the DCS.
  • configuration and status information related to that DCS is retrieved from cross connection status database 324 to identify how the path across the DCS should be provisioned to route the traffic to the appropriate destination.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Endoscopes (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

Systems and methods for determining the integrity of an endoscope. The systems and methods provide for an air enclosure which includes the internal area of the sleeve or tubing of the endoscope to be pressurized. Pressure measurements of the air enclosure are then taken both before and after the endoscope is manipulated by a user. The systems and methods may also or alternatively vent air from inside the endoscope to a humidity detector to detect the presence of fluid inside the endoscope.

Description

DETERMINING AND PROVISIONING PATHS IN A NETWORK
[0001] This application claims the benefit of U.S. Provisional Application No.
60/614,609, filed September 30, 2004, which is hereby incorporated herein by reference.
BACKGROUND OF THE INVENTION
[0002] Communications networks, such as next generation broadband networks, have become increasingly complex due to increased size, numerous intermixed technologies/protocols (e.g., ATM, Frame Relay, etc.), and the intermixing of equipment manufactured by numerous different vendors. As a result, network configuration management systems that can provision virtual trunks and circuits within these networks are becoming increasingly important. Such network configuration management systems function to determine the paths/routes between network equipment, herein referred to as network elements, and to communicate with those network elements to realize desired trunks or circuits that facilitate the transmission of traffic across the network.
[0003] In general, network configuration management systems have traditionally determined the paths available by modeling portions of network elements as nodes on a graph and the links/interconnections between these portions as links between the nodes. More particularly, prior systems typically modeled every port of every network element as a node on the graph and modeled every physical link that interconnected these ports to one another as links that interconnected the nodes of the graph. The network model was then used to provision virtual trunks, which formed paths between network elements in the network. Once these virtual trunks were provisioned, virtual circuits could then be established across these trunks to support traffic flow from one point to another in the network.
[0004] FIG. 1 shows an exemplary prior art network configuration management system 102 and a network 110 managed by system 102. The network configuration management system 102 functions to determine a preferred path between two points in a network (i.e., between two network elements) and for provisioning a communications connection across this path by communicating with the managed network 110. Managed network 110 consists primarily of broadband network 112 which, in turn, consists of a plurality of network elements 114-118 interconnected by physical links and virtual trunks and circuits represented in FIG. 1 by links 120-124. The network elements comprise varying technologies and protocols and may be manufactured by different vendors. Managed network 110 further comprises network management systems, such as network management system (NMS) 126, and element management systems, such as element management system (EMS) 128. These systems are typically provided by the network element manufacturers and typically function to perform the actual configuration and management of the individual network elements. [0005] NMSs and EMSs may function to control both the network elements and the links between those elements. However some may not control the links between the elements and, instead, only manage the network elements themselves. For example, an NMS, such as NMS 126, may function to collectively manage a set of network elements 114 and the physical links 120 between them, thus forming a collectively managed sub-network having network elements 114. Accordingly, when network traffic arrives at an ingress port into one of the network elements 114, such as port 130, the NMS 126 determines a set of links and network element cross- connects to interconnect port 130 to an egress port, such as port 132. The NMS 126 then provisions the network elements to realize this interconnection. In another example, some management systems, such as EMS 128, may only manage one or more network elements 118, but not the links 124 between them. Here, a higher layer entity, such as the Network Configuration Management System 102, determines the links between network elements 118 required to create a path and then instructs the EMS to perform the necessary cross-connects within network elements 118 to realize the complete path.
[0006] FIG. 1 also shows how some network elements, such as network elements 116, are not managed by either an NMS or EMS. Specifically, a higher layer entity, once again such as Network Configuration Management System 102, directly communicates with these elements to perform network configuration functions. In this case, Network Configuration Management System 102 would configure any cross-connects within network elements 116 as well as any links between network elements. Thus, as shown in FIG. 1 , to facilitate traffic flow across broadband network 112, for example from port 130 on network element 114 to network element 118, the combination of Network Configuration Management System 102, NMS 126 and EMS 128 will collectively determine an appropriate network path across and between network elements 114, 116 and then provision virtual trunks and circuits across network 112.
[0007] One difficulty with prior methods of using network configuration management systems, such as those described above, is that the modeling of the network elements, physical links, and virtual trunks and circuits results in very large, inefficient models that do not adapt well to diverse network elements and large networks. Specifically, such large models result in correspondingly large and complex network model graphs which, in turn, create performance and scalability issues due to the demanding processing requirements associated with such graphs. Therefore, in one prior attempt at solving this problem and to reduce the aforementioned disadvantages, a network model was created based on how the ingress and egress ports of each network element can be interconnected within themselves and to other network elements. Specifically, in this prior attempt, a simplified routing graph was created by the network configuration management system whereby, instead of modeling each port of a network element as a node on a routing graph, an entire network element itself could be represented as one or more routing nodes or, in some cases, multiple network elements could be represented as a single routing node. Referring to FIG. 2, for example, network elements 114 of FIG. 1 that are managed by NMS 126 are modeled as a single node 201. Additionally, network elements 118, which are managed by both EMS 128 and the Network Configuration Management System 102 are also modeled as a single routing node 204. Network elements 116 are each modeled as individual routing nodes, since the Network Configuration Management System 102 manages both the network element and the link between the elements. In such a model, therefore, the individual physical hardware links are not each modeled but, rather, one or more network elements are modeled as a single routing node based on how those network elements and the links between them are managed. Such an attempt is generally described in pending U.S. Patent Application Serial No. 10/118,187, filed April 8, 2002 and entitled "Determining and Provisioning Paths Within a Network of Communication Elements" (hereinafter referred to as the "'187 application"), which is hereby incorporated by reference herein in its entirety.
SUMMARY OF THE INVENTION
[0008] While the prior methods of creating network models for routing traffic across networks and between multiple networks are advantageous in many regards, as discussed above they are limited in certain regards. In particular, while processing associated with network routing can be greatly simplified using the prior methods, such processing can still be resource and overhead intensive. This is especially the case as networks using different speeds and/or protocols are being interconnected to provide new and more complex services to customers.
[0009] Accordingly, the present inventor has invented a network provisioning system for establishing a path between two networks wherein a common network device between those networks is modeled as a link between a first network element in one network and a second network element in a second network. In one embodiment, a network routing graph is created by an inventory subsystem in a routing manager by inventorying the physical network elements and links in the network. The inventory subsystem then models those elements/links as a plurality of virtual nodes and links between the nodes. At least one common network device, such as a digital cross connect located at a junction between the two networks, is modeled as a link instead of a node. A routing engine then uses the network routing graph, including the link modeled from the common network device, to provision a path between the two networks. Thus, since fewer nodes are represented in a network graph of the modeled network, route processing is reduced, resulting in a corresponding reduction in overhead and resources required to route network traffic from one node to another.
DESCRIPTION OF THE DRAWING
[00010] FIG. 1 shows a prior art managed broadband network and a network configuration management system for determining and provisioning routing paths within the network;
[00011] FIG. 2 shows a network routing model whereby some network elements are combined and treated as single routing nodes; [00012] FIG. 3 shows an illustrative network configuration management system;
[00013] FIG. 4 shows a network routing model whereby Digital Cross Connect
Systems (DCSs) are used to interconnect different network nodes; and
[00014] FIG. 5 shows a network routing model in accordance with the principles of the present invention whereby DCSs are modeled as links.
DETAILED DESCRIPTION OF THE INVENTION
[00015] FIG. 3 shows an illustrative network configuration management system, such as Network Configuration Management System (NCMS) 102 in FIGs. 1 and 2. As discussed above, NCMS 102 determines preferred routing paths between two ports within the network by modeling the network paths as a plurality of routing nodes and links between the nodes, and for using these paths to provision virtual trunks and circuits within the networks. To accomplish this function, NCMS 102 includes, among other components, a routing manager 304 and inventory database 322. The routing manager 304 provides end-to-end connection management functions including the determination and provisioning of routing paths in broadband network 112 in FIG. 1. In order to accomplish these functions, routing manager 304 comprises an inventory subsystem 306, a routing engine 308 and a service activation system 310. The routing manager 304 is connected to the various network elements via an element adapter 312 and connection 313. Broadly, the routing manager 304 maintains a topological graph comprising nodes and links that model the broadband network 112. This graph is used to determine and provision routing paths between, for example, two ports within the network. These paths are then used to provision virtual trunks and circuits.
[00016] The inventory subsystem 306 builds and maintains the topological graph in accordance with modeling methods such as those described above in association with the '187 application. This graph is maintained, illustratively, in three database tables: routing link table 314, routing node table 316, and NMS/EMS table 318. The routing engine 308 determines a routing path for traffic through the network using the network graph maintained by the inventory subsystem 306. The service activation system 310 then uses the determined routing path to provision the actual virtual trunk or virtual circuit. Specifically, the service activation system 310 activates the routing engine 308 to obtain a routing path given two endpoints and then invokes the element adapter 312 which interfaces with network elements, NMSs and EMSs to physically provision the determined path. As such, the element adapter 312 functions as an interface between the routing manager 304 and the NMSs 126, EMSs 128, and network elements 116 in managed broadband network 112. As one skilled in the art will recognize, there is typically a specific element adapter designed for use with NMSs, EMSs, and network elements manufactured by different manufacturers. As such, a network management system may have multiple element adapters, or different modules in one element adapter. Accordingly, once the service activation system determines a routing path, it invokes the appropriate adapter(s) or adapter module(s) to communicate the required configuration settings to the management systems/elements 126, 128, and 116 to provision the determined path. [00017] As one skilled in the art will recognize, and as is further discussed herein below, network traffic may be required to traverse multiple separate networks. These different networks may be interconnected with cross connects, such as digital cross connects (DCSs). As such, it is necessary for the NCMS 102 to also have available configuration and status information related to these DCSs. This configuration and status information is, illustratively, maintained in cross-connection status database 324. Thus, in provisioning the aforementioned path, service activation system 310 may also refer to cross-connection status database 324. [00018] The prior illustrative method described in the '187 application for using an NMS to simplify routing graphs is advantageous in many regards. By eliminating the need to inventory individual ports and by reducing the number of nodes necessary to consider in routing network traffic from one point to another, the processing overhead and timeliness associated with making routing decisions is greatly reduced. Additionally, such an approach adds considerable flexibility in designing and maintaining routing graphs. Specifically, as described in that application, instead of inventorying and maintaining a database of each port in a network and the interconnections between those ports, it is only necessary to inventory the routing nodes and the links between the routing nodes that, for example, may consist of several network elements.
[00019] As one skilled in the art will recognize, the method described in the
'187 application is primarily focused on network routing at layer 2 of the network. As is well understood, networks have been modeled as operating at different layers. One model for such network layers is known as the Open System Interconnection (OSI) model, which generally defines 7 different layers in the network. Layer 2 is also known as the data link layer and is the layer at which the physical medium is shared and where data link and media access are controlled. For example, in Ethernet networks, layer 2 is the layer at which network routing between media access control (MAC) addresses of individual hardware components is performed. [00020] The above-described network model at layer 2 of a network is primarily useful within a single network. However, with increasingly complex and large networks it has become necessary to cross network boundaries in order to route network traffic from one destination to another. In many cases, the different networks rely on different protocols, operate at different speeds and may even operate using a different physical medium (e.g., copper vs. optical fiber). In order to interconnect such networks, DCSs or other similar devices, such as optical cross connect systems (OCSs), are used. As used herein, a DCS is any device that interconnects networks to facilitate traffic routing from one network to another or to link portions of networks using one protocol or traffic rate to another portion using a different protocol or rate. Such DCSs are very well known in the art and serve to efficiently manage disparate traffic protocols and line speeds in telecommunications system central offices as well as remote field locations and other locations such as within hotels and even at user premises. Such DCSs may be used in a variety of different applications. For example, DCSs may be used at a customer premises to interface with both voice protocol networks and a number of different data protocol networks in order to aggregate and cross connect these networks to a high-speed copper wire or optical fiber loop. Additionally, DCSs may be used in a digital loop carrier (DLC) capacity to aggregate networks using multiple protocols for transmission across a SONET ring network. In another common implementation, such DCSs may be used within, illustratively, a telecommunications central office in order to manage and cross connect channels from multiple SONET rings that are employed in an access network and/or a metro or inter-office network. Other uses of DCS are well known and will be obvious to one skilled in the art. [00021] FIG. 4 shows one illustrative routing map wherein DCSs are used to connect networks to facilitate traffic flow from one network to another. In particular, FIG. 4 shows routing nodes 401-406, each of which represents, illustratively, a network, such as broadband network 112, or a portion of a network, such as the group of network elements 116 also in FIG. 1. As such, each of the routing nodes 401-406 illustratively has a plurality of network elements that are modeled, for routing purposes, as a single routing node with an ingress port and an egress port, such as ports 130 and 134, respectively, in FIG. 1. The networks represented by each of routing nodes 401-406 may, for example, operate using a different protocol or speed and, therefore, DCSs, such as DCSs 407, 408 and 409, may be used to aggregate and/or disaggregate traffic in order to facilitate the transmission of that traffic between and over the different networks interconnected by the DCSs. For example, routing nodes 402 and 405 may represent well-known OC-3 networks operating at an illustrative speed of 155.52 Mb/s while the networks represented by nodes 403 and 406 may be well-known OC-12 networks operating at an illustrative 622.08 Mb/s rate. DCSs 407 and 408 aggregate and/or disaggregate the data between the networks represented by nodes 402 and 403 and DCS 409 aggregates and disaggregates the traffic between the networks represented by nodes 405 and 406. Typically, paths through DCSs 407-409 are provisioned in a relatively static manner. For example, a path from port 407A, associated with node 402, to port 407B, associated with node 403, is provisioned on DCS 407 in order to provide connectivity between the networks represented by nodes 402 and 403. Connections between ports 408A/408B and 409A/409B are similarly provisioned to connect nodes 402/403 and 405/406, respectively. Thus, one skilled in the art will recognize that DCSs 407-409 function as common nodes between the respective networks. [00022] As one skilled in the art will recognize, a DCS, such as any one of
DCSs 407-409, functions similarly in some respects to a network switch, such as a router or ATM switch. However, such routers/switches typically operate as a function at least in part of the signaling accompanying traffic transiting the network and, hence, such routers/switches are typically closely tied to specific services provided by a network service provider. A DCS is typically not used for such purposes. Instead, a DCS is typically used for transmission management at a higher level of the network. Specifically, unlike most telecom services where switch control is an inherent element of the service provided to customers and is closely tied to the protocol used at layer 2 of the network, DCSs are typically used as an engineering and provisioning control mechanism at layer 1 in the network (i.e., the physical layer of the network). As such, DCSs are typically not used to dynamically alter switching over a short time period, as are routers and other types of switches. Additionally, DCSs are not typically controlled as a function of signaling from a customer but are, instead, controlled directly by, for example, engineers at the service provider. Also unlike simpler network switches, a typical DCS facilitates the provisioning of network paths and connections across the DCS that are typically constant over a period of hours to months.
[00023] As service providers, such as telecommunication service providers, strive to provide more advanced features to consumers, interconnections and junctions between networks, such as those created by DCSs 407-409 and other similar devices, become greater in number and grow in importance. These interconnection devices must be taken into account when developing a network routing strategy. Traditionally, in making routing decisions, the network configuration management system modeled devices such as DCSs as one or more separate routing nodes. The present inventor have discovered, however, that such a modeling of DCSs increases the routing processing required due to a larger number of "hops" necessary to traverse nodes in the routing model. This increases both the time and overhead necessary to, for example, generate the aforementioned routing graphs. Therefore, the present inventor have further discovered that, in addition to simplifying routing decisions at layer 2 of a network, as described in the '187 application, it is desirable to also simplify the routing graph at layer 1 of the network. Specifically, instead of treating DCSs as a separate node (or multiple nodes corresponding the ports on the DCS) in the network, it is also desirable to model DCSs differently in order to further simplify the routing graphs/decisions. More particularly, in part since DCSs and other similar devices are relatively static in configuration, the present inventor have discovered that such devices may be treated as links, such as would be formed by a physical cable, instead of nodes that require processing as an affirmative routing hop.
[00024] As described above in association with the '187 application, prior to provisioning network paths in a network, such as network 112 in FIG. 1 , a network configuration management system, such as NCMS 126 in FIG. 1 will inventory the network elements and links in the network. Once these elements and links are defined, the NCMS generates a routing graph showing the network topology in terms of routing nodes and links to be used in provisioning trunks/circuits across the network. FIG. 5 is a simplified representation of such a routing map. In particular, routing nodes 401-406 are as described above in association with FIG. 4. Each of those routing nodes consists, for example, of a plurality of network elements that are modeled at a high level as a single routing node in order to decrease the processing overhead required to provision the aforementioned trunks/circuits. However, instead of modeling the ports of DCSs 407-409 of FIG. 4 as individual nodes, or as multiple nodes, those illustrative DCSs are modeled as links 501 , 502 and 503. Links 501 , 502 and 503 are used in the routing graph of FIG. 5 to represent DCSs 407, 408 and 409, respectively. Accordingly, in accordance with the principles of the present invention, a cross-connect, such as a DCS, is not modeled as one or more routing nodes having various links connecting ports to each other and to external ports on other network elements. Instead, such a cross-connect is modeled as a separate link between network elements in one or more networks. Accordingly, the routing map is greatly simplified.
[00025] One skilled in the art will recognize that, as DCSs or other network components are added or deleted, the NCMS will inventory the network elements and links between the elements, treating DCSs as links as described above. Specifically, this inventory is conducted by the inventory subsystem 306 of FIG. 3. As a part of this inventory, routing link table 314, routing node table 316, NMS/EMS table 318 and cross connection status database 324 are updated with information about the links, nodes and cross connections in and between the networks managed by the NCMS 102. Therefore, in this inventory, information concerning each DCS will be updated in the cross-connection status database and those same DCSs will be updated as links in the routing link table. As a result, when service activation system 310 invokes the routing engine 308 to provision a path, that engine will treat the DCSs as links to be provisioned and not one or more network nodes corresponding to the ports on the DCS. When network traffic traverses a particular DCS, configuration and status information related to that DCS is retrieved from cross connection status database 324 to identify how the path across the DCS should be provisioned to route the traffic to the appropriate destination. [00026] One skilled in the art will recognize that many variations are possible and that any or all of these embodiments described herein above may be combined in order to create a border element function that is decentralized depending upon the needs of a particular network in order to add flexibility to network design and to reduce routing management overhead costs. The foregoing Detailed Description is to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined from the Detailed Description, but rather from the claims as interpreted according to the full breadth permitted by the patent laws. It is to be understood that the embodiments shown and described herein are only illustrative of the principles of the present invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention. Those skilled in the art could implement various other feature combinations without departing from the scope and spirit of the invention.

Claims

CLAIMS:
1. A network provisioning system for establishing a path between at least a first network element and at least a second network element, said at least a first network element and said at least a second network element being interconnected through a common network device, said system comprising: an inventory subsystem for modeling each of said at least a first network element and said at least a second network element as one or more routing nodes in a graph, said graph comprising a plurality of routing nodes and a plurality of links interconnecting said plurality of routing nodes; and a routing engine adapted to use said graph for determining a path between said at least a first network element and said at least a second network element, wherein the inventory subsystem models said common network device as a link between said at least a first network element and said at least a second network element.
2. The network provisioning system of claim 1 wherein said common network device comprises a digital cross connect system.
3. The network provisioning system of claim 1 wherein said at least a first network element is in a first network and said at least a second network element is in a second network.
4. A routing manager for provisioning paths for network traffic between a plurality of network elements in one or more networks wherein at least a first common network device is disposed between a first network element in said plurality of network elements and a second network element in said plurality of network elements, said routing manager comprising: means for creating a graph of routing nodes and links, said routing nodes representing one or more network elements in said plurality of network elements and said links representing interconnections between said routing nodes; and means for modeling said at least a first common network device as a link between those routing nodes representing said first network element and said second network element.
5. The routing manager of claim 4 wherein said at least a first common network device comprises a digital cross connect system.
6. The routing manager of claim 4 wherein said first network element is an element in a first network and said second network element is an element in a second network.
7. A method for routing network traffic between a first network and a second network, each of said first and second networks comprising a plurality of network elements, said plurality of network elements connected by a digital cross connect, said method comprising the steps of: determining the interconnections created by said digital cross connect system between at least two network elements in said plurality of network elements; and representing each of said interconnections as a link between said at least two network elements.
8. The method of claim 7 further comprising storing the status of each of said interconnections in a cross connection status database.
9. The method of claim 7 further comprising provisioning a path between said at least two network elements.
10. A method for determining a path between a first network element in a first network and a second network element in a second network, said first network connected to said second network via a common network device, each of said networks comprising a plurality of network elements and a plurality of network links, said method comprising the steps of: modeling said plurality of network elements in said first and second networks as one or more routing nodes; modeling said network links as routing links, said routing links interconnecting said routing nodes; and modeling said common network device as a routing link connecting a first routing node in said first network to a second routing node in said second network.
11. The method of claim 10 further comprising: storing said routing links in a routing link table.
12. The method of claim 10 further comprising the step of: determining a network path between said first routing node and said second routing node using said routing link.
13. The method of claim 12 further comprising the step of: provisioning said network path to interconnect a first network element in said first routing node with a second network element in said second network element.
14. The method of claim 10 further comprising: maintaining the status of said routing links, said status indicating whether a cross-connection using at least one of said routing links was successfully or unsuccessfully provisioned.
PCT/US2005/033418 2005-05-06 2005-09-16 Systems and methods for endoscope integrity testing WO2006121462A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP05801033.1A EP1887919B8 (en) 2005-05-06 2005-09-16 Systems and methods for endoscope integrity testing
ES05801033.1T ES2625927T3 (en) 2005-05-06 2005-09-16 Systems and methods to check the integrity of endoscopes
AU2007240157A AU2007240157A1 (en) 2005-05-06 2007-12-06 Systems and methods for endoscope integrity testing

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US11/123,336 2005-05-06
US11/123,336 US20060252991A1 (en) 2005-05-06 2005-05-06 Systems and methods for endoscope integrity testing
US11/123,335 2005-05-06
US11/123,335 US20060252990A1 (en) 2005-05-06 2005-05-06 Systems and methods for endoscope integrity testing

Related Child Applications (1)

Application Number Title Priority Date Filing Date
AU2007240157A Division AU2007240157A1 (en) 2005-05-06 2007-12-06 Systems and methods for endoscope integrity testing

Publications (3)

Publication Number Publication Date
WO2006121462A2 true WO2006121462A2 (en) 2006-11-16
WO2006121462A9 WO2006121462A9 (en) 2008-02-14
WO2006121462A3 WO2006121462A3 (en) 2009-04-16

Family

ID=37397011

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/033418 WO2006121462A2 (en) 2005-05-06 2005-09-16 Systems and methods for endoscope integrity testing

Country Status (3)

Country Link
EP (1) EP1887919B8 (en)
ES (1) ES2625927T3 (en)
WO (1) WO2006121462A2 (en)

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4103146C1 (en) * 1991-02-02 1992-03-26 Richard Wolf Gmbh, 7134 Knittlingen, De
US5317896A (en) * 1992-03-13 1994-06-07 American Sterilizer Company Method of detecting liquid in a sterilization system
US5788688A (en) * 1992-11-05 1998-08-04 Bauer Laboratories, Inc. Surgeon's command and control
US5537880A (en) * 1995-06-07 1996-07-23 Abbott Laboratories Automatic pipetting apparatus with leak detection and method of detecting a leak
FR2784032B1 (en) * 1998-10-02 2000-11-17 Gerard Iffrig DECONTAMINATION PAIL OF FLEXIBLE ENDOSCOPES
US6112578A (en) * 1999-04-26 2000-09-05 Daimlerchrysler Corporation Multiple cavity leak test process
US6408682B2 (en) * 2000-02-07 2002-06-25 Steris Inc. Leak detection method for endoscopes

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of EP1887919A4 *

Also Published As

Publication number Publication date
EP1887919B1 (en) 2017-02-22
EP1887919A4 (en) 2011-09-07
EP1887919A2 (en) 2008-02-20
WO2006121462A3 (en) 2009-04-16
WO2006121462A9 (en) 2008-02-14
ES2625927T3 (en) 2017-07-21
EP1887919B8 (en) 2017-04-19

Similar Documents

Publication Publication Date Title
CA2581734C (en) Determining and provisioning paths in a network
US7289456B2 (en) Determining and provisioning paths within a network of communication elements
US9130878B2 (en) Traffic switching in hybrid fiber coaxial (HFC) network
Labourdette et al. Branch-exchange sequences for reconfiguration of lightwave networks
US5513171A (en) Arrangement for dynamically deriving a telephone network management database from telephone network data
US8402120B1 (en) System and method for locating and configuring network device
US20060013149A1 (en) Suprvisory channel in an optical network system
US20040042416A1 (en) Virtual Local Area Network auto-discovery methods
US20070121664A1 (en) Method and system for double data rate transmission
US7414985B1 (en) Link aggregation
US20070121619A1 (en) Communications distribution system
JP4381639B2 (en) Apparatus and method in a switched telecommunications system
US7706290B2 (en) Object-based operation and maintenance (OAM) systems and related methods and computer program products
JP4205953B2 (en) Improvements in and related to communication networks
CA2374064C (en) Path discovery in a distributed network management architecture
Ellanti et al. Next generation transport networks: data, management, and control planes
KR101160402B1 (en) Dual oss management of an ethernet access network
US20070121628A1 (en) System and method for source specific multicast
WO2006121462A2 (en) Systems and methods for endoscope integrity testing
WO2005018174A1 (en) Multiple services provisioning in a packet forwarding device with logical ports
Jajszczyk The ASON approach to the control plane for optical networks
Kasu et al. Spanning Tree Protocol
EP1045536A1 (en) Optical fiber communication network device
Suzuki et al. A management design for a LAN-like optical access network based on STM-PDS
Cao et al. Optical internetworking models and standards directions

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

NENP Non-entry into the national phase

Ref country code: RU

REEP Request for entry into the european phase

Ref document number: 2005801033

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2005801033

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2005801033

Country of ref document: EP