EP0724795A1 - Apparatus and method for determining network topology - Google Patents

Apparatus and method for determining network topology

Info

Publication number
EP0724795A1
EP0724795A1 EP94929738A EP94929738A EP0724795A1 EP 0724795 A1 EP0724795 A1 EP 0724795A1 EP 94929738 A EP94929738 A EP 94929738A EP 94929738 A EP94929738 A EP 94929738A EP 0724795 A1 EP0724795 A1 EP 0724795A1
Authority
EP
European Patent Office
Prior art keywords
data
network
devices
relay devices
port
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP94929738A
Other languages
German (de)
French (fr)
Inventor
Timothy L. Orr
Eric W. Gray
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cabletron Systems Inc
Original Assignee
Cabletron Systems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cabletron Systems Inc filed Critical Cabletron Systems Inc
Publication of EP0724795A1 publication Critical patent/EP0724795A1/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/02Topology update or discovery
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • H04L41/122Discovery or management of network topologies of virtualised topologies, e.g. software-defined networks [SDN] or network function virtualisation [NFV]

Definitions

  • This invention relates to a system for determining the topology of a computer network which includes data-relay devices and node devices, the topology being determined based on a comparison of source addresses "heard" by the various data-relay devices.
  • Computer networks are widely used to provide increased computing power, sharing of resources and communication between users.
  • Computer systems and computer system components are interconnected to form a network.
  • Networks may include a number of computer devices within a room, building or site that are interconnected by a high speed local data link such as local area network (LAN), token ring, Ethernet, or the like.
  • Local networks in different locations may be interconnected by techniques such as packet switching, microwave links and satellite links to form a world-wide network.
  • a network may include several hundred or more interconnected devices.
  • Network management systems have been utilized in the past in attempts to address such issues.
  • Prior art network management systems typically operated by remote access to and monitoring of information from network devices.
  • the network management system collected large volumes of information which required evaluation by a network administrator.
  • Prior art network management systems place a tremendous burden on the network administrator. He/she must be a networking expert in order to understand the implications of a change in a network device parameter.
  • the administrator must also understand the topology of each section of the network in order to understand what may have caused the change.
  • the administrator must sift. through reams of information and false alarms in order.to determine the cause of a problem.
  • An important aspect of any network management system is its ability to accurately represent interactions within the network and between network devices. Toward this end it is crucial that any network management system be provided with some means to determine the topology of the network.
  • a data-relay device e.g., a bridge
  • a list of network addresses "heard" by each port of a data-relay device is compiled for each data-relay device in a computer network.
  • the network includes a plurality of data-relay devices and node devices interconnected by a data bus.
  • the node devices e.g., work stations or disk units
  • the data-relay devices also transmit their own source address with replies sent to the network management system.
  • Each of the data-relay devices acquires and maintains a source address table which lists the addresses heard by each port of the data-relay device. These lists are then compared to determine whether there is a direct or transitive connection between select ports on different data-relay devices and the resulting connections are used to define a topology showing the interconnections between the devices in the network.
  • a node device z that is not in the set of addresses heard by X must be in some set X where q does not equal i. If port Y. is connected to port X. , then z must be in set Y. and in no other set Y such that r does not equal j. Thus, to determine which port of Y is connected to port X. we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. If enough of the set X ' ⁇ and Y 's are empty, there may be no ports remaining and the technique fails. Otherwise, one is left with one port Y which is the port that is connected to X ⁇ .
  • the apparatus of the present invention is a system for use with a digital computer network and preferably with a network management system.
  • the network management system includes a virtual network having models representing network devices, each model containing network data relating to the corresponding network device and means for processing the network data to provide user information.
  • the system includes means for transferring network data from the network devices to the corresponding models, and means for supplying user information from the virtual network to a user.
  • a model of each data-relay device includes as network data a list of source addresses heard by each port of the data-relay device and the processing means are inference handlers which compare the sets of addresses of the various data-relay devices to determine the topology of the computer network.
  • the determined network topology is used by the management system for controlling traffic on the network, optimizing network resources, security, isolation of network faults, and the like.
  • the determined topology may be provided to the user by a video display unit connected to the computer.
  • Fig. 1 is a schematic diagram showing an example of a network
  • Fig. 2 is a schematic diagram showing the partition imposed by data-relay device A in the network of Fig. 1;
  • Fig. 3 is a schematic diagram showing the partition imposed by data-relay device B in the network of Fig. 1;
  • Fig. 4 is a schematic diagram showing the partition imposed by data-relay device C in the network of Fig. 1;
  • Fig. 5 is a schematic diagram showing the combined partitions of the data-relay devices A, B and C in the network of Fig. 1;
  • Fig..6 is a schematic diagram showing trial connections between the data-relay devices of the network of Fig. 1, determined according to the present invention
  • Fig. 7a is a schematic diagram showing the final connections between the data-relay devices in the network of Fig. 1, determined according to the present invention, and Fig. 7b is the same as Fig. 7a but further includes the node devices between the data-relay devices;
  • Fig. 8 shows an example of the source address tables for each port of the data-relay devices in the network of Fig. 1;
  • Fig. 9 is a flow chart showing a method of determining a connection between data-relay devices according to the present invention.
  • Fig. 10 is a flow chart showing a further method of determining connections between data-relay devices according to the present invention.
  • Fig. 11a is a flow chart showing a method of making the trial connections and final connections (see Figs. 6-7) according to the present invention to determine the topology of the network of Fig. 1;
  • Figs, lib and lie are schematic diagrams showing the elimination of redundant connections in two sample networks;
  • Fig. 12 is a block diagram of a network management system which may utilize the present invention.
  • Fig. 13 is a block diagram showing an example of a network connected to the network management system of Fig. 13;
  • Fig. 14 is a schematic diagram showing the structure of models and the relations between models in the network management system of Fig. 13 which utilizes the present invention.
  • the present invention is an apparatus and method for determining the topology of a computer network consisting of data-relay devices and node devices, which topology may be computed automatically by obtaining and processing information from the network devices.
  • a data-relay device is a network element that provides electrical and/or logical isolation of local area network segments.
  • a data-relay device forwards network packets from one port of the device to other ports of the device, amplifying and retiming the electrical signals to (a) increase the maximum cable length, (b) increase the number of permissible nodes on the network, and/or (c) increase the number of network segments.
  • Examples of data-relay devices include bridges, repeaters, hubs and the like. Bridges are defined in IEEE 802.1(d) specification, and repeaters are defined in IEEE 802.1(e) specification. In addition to the above functions, bridges also provide a storage and forward mechanism with packet filtering, to reduce the network load. Hubs are bridges but forward all traffic without filtering. For use in this invention, the hub must be an "intelligent" hub with a source address table for network management as described below.
  • each data-relay device on the network has a source address table, which is a list of network addresses "heard" by each port of the device.
  • a source address table which is a list of network addresses "heard" by each port of the device.
  • Entries in the source address table are usually "aged,” so that an entry exists in the table for a finite period of time unless refreshed by another packet being received from the same source address on the same port. This maintains the currency of the table, since a device which is physically removed from the network will eventually disappear from the source address table.
  • the data-relay device will not receive a packet from that node for entry on the source address table. However, this can be corrected for by periodically poling all of the node devices on the network to confirm the existence of all devices on the network.
  • data-relay devices In order to prevent loops in the network, and thus duplication of packets, data-relay devices must not be connected redundantly. Some devices, such as IEEE 802.1(d) compliant bridges, use a spanning tree algorithm to automatically detect and disable redundant devices. In performing their primary function of packet forwarding, data-relay devices are essentially transparent to the network; They do not transmit packets with their own addresses. However, data-relay devices do transmit responses to network management requests with their own address in the source address field of the packet. This feature is important to the computation of the network topology according to the present invention since, otherwise, the data-relay devices would not appear in each other's source address tables.
  • Some data-relay devices perform directed transmissions of management packets. Thus, instead of the management response being transmitted on all ports of the device, it is transmitted only on the port where the management request was received. This is usually evident in bridges where the management packet itself is subject to a filtering process.
  • the address of the device may or may not appear in the source address tables of other data-relay devices, depending on the relative placement of the data-relay devices and the management station. This fact is taken into account in accordance with the present invention when computing the network topology form the source address tables as described below.
  • a sample network is shown in Fig. 1.
  • the network includes three data-relay devices, 10, 12 and 14, which have been identified as devices A, B and C. each with two numbered ports.
  • the data-relay devices separate the network into three sections or partitions, with end-node devices d-k and management station m at various locations on the network.
  • the partitions generated by devices A, B and C are shown in Figs. 2, 3 and 4, respectively.
  • the combination of partitions is shown in Fig. 5.
  • partition 9 includes node devices d, e and m connected to port 1 of data-relay device A; • partition 11 includes node devices f and g and between port 2 of data-relay device A and port 1 of data-relay device C;
  • partition 13 includes node devices h and i between port 2 of data-relay device C and port 1 of data-relay device B;
  • partition 15 includes node devices j and k connected to port 2 of data-relay device B.
  • the source address table for each data-relay device reflects the partitioning.
  • Each port "hears" all the devices contained in its partition.
  • the sets of devices heard by the various ports are:
  • A2 (B, C, f, g, h, i, j, k)
  • the data-relay devices may not be heard by other data-relay devices due to directed transmission of management packets. For example, note that Bl does not hear device C, since in this example C is directing its management traffic toward the management station m. However, since in this example A does not direct its management traffic, Bl does hear A.
  • a transitive connection is defined as a connection between two nodes that crosses one or more data-relay devices. For example, in Fig. 1 port A2 is transitively connected to port Bl via device C.
  • a direct connection has no data-relay device intervening.
  • port A2 is directly connected to port Cl.
  • a first method of the present invention for determining connections between the data-relay devices is illustrated by the flow chart of Fig. 9, wherein the address tables for a pair of data-relay devices X and Y, such as devices A and C in Fig. 8.
  • the address tables for a pair of data-relay devices X and Y such as devices A and C in Fig. 8.
  • C is an element of A2
  • A is an element of Cl
  • A2 is connected to Cl (see boxes around relevant entries in the address tables of Fig. 8).
  • a node device z that is not in the set X. must be in some set X such that q does not equal i. If port Y. is connected to port X., then z must be in the set Y. , and in no other set Y such that r does not equal j . In other words, for all q not equal to i, there is some intersection of the sets Xq and Yr if p c ort Xl- is connected to port Y . If there is no intersection between X and Y (i.e., the null set), then port X. is not connected to port Y .
  • step 300 of Fig. 10 we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. Thus, in step 300 of Fig. 10, we consider whether Y hears X . If the answer is, "yes” then there is a connection between X. and Y (step 301); if the answer is, "no” we check whether this is the last port on Y (step 302) and, if not, we increment to the next port on Y (step 303), After checking all of the ports of Y, if the intersection is the null set, then no connection can be determined between X. and any port on Y (the method fails)(step 304).
  • connections, or edges between the two ports include both transitive connections as well as direct connections. If a transitive connection exists, then the direct connection is redundant and should be eliminated. For example, as shown in Fig. 7, because there is a transitive connection between A2 and Bl, the direct connection between A2 and Bl has been eliminated.
  • step 320 we first form trial connections between X. and Y. , as shown in Fig. 6.
  • step 321 we check whether X. is directly connected to Y. (step 321). If there is no direct connection, we check whether this is the last port on X (step 322) and either increment to the next port on X (step 323) or, if this is the last port on X, we increment to the next device (step 324). If we find that X. is directly connected to Y. , we then check whether X. is transitively connected to Y. (step 325). If the answer is, "yes" we remove the direct connection between X. and Y. (step 326) .
  • the topology of Fig. 7 can be further defined to include the node devices in each network segment.
  • Each node device belongs to a segment if its address appears in the intersection of the port sets X. of each data-relay port which is directly connected to the segment.
  • node devices h and i appear in the address tables for each ports C2 and Bl and, therefore, nodes h and i should be included as part of the connection 22 between devices A and C.
  • Fig. 7b shows a topology further including the node devices.
  • the method and apparatus for determining network topology form a part of a network management system, or provide network topology information to a network management system, such as that described in copending and commonly assigned U.S. Application Serial No. 07/583,509, filed September 17, 1990, entitled, NETWORK MANAGEMENT SYSTEM USING MODEL BASED INTELLIGENCE, by R. Dev et al; that application was also filed as PCT/US91/06725 on September 17, 1991 and published as Int'l Publ. No. WO92/05485 on April 2, 1992; that application is hereby incorporated by reference in its entirety.
  • the present invention may also be used in combination with other network management systems, or in any application where it is desired to obtain the network topology.
  • Fig. 12 shows a block diagram of the system.
  • the major components of the network management system are a user interface 410, a virtual network machine 412, and a device communication manager 414.
  • the user interface 410 which may include a video display screen, keyboard, mouse and printer, provides all interaction with the user.
  • the user interface controls the screen, keyboard, mouse and printer and provides the user with different views of the network that is being managed.
  • the user interface receives network information from the virtual network machine 412.
  • the virtual network machine 412 contains a software representation of the network being managed, including models that represent the devices and other entities associated with the network, and relations between the models.
  • the virtual network machine 412 is associated with a database manager 416 which manages the storage and retrieval of disk-based data. Such data includes configuration data, an event log, statistics, history and current state information.
  • the device communication manager 414 is connected to a network 418 and handles communication between the virtual network machine 412 and network devices.
  • the data received from the network devices is provided by the device communication manager to the virtual network machine 412.
  • the device communication manager 414 converts generic requests from the virtual network machine 412 to the required network management protocol for communicating with each network device.
  • Existing network management protocols include Simple Network Management Protocol (SNMP), Internet Control Message Protocol (ICMP) and many proprietary network management protocols. Certain types of network devices are designed to communicate with a network management system using one of these protocols.
  • a view personality module 420 connected to the user interface 410 contains a collection of data modules which permit the user interface to provide different views of the network.
  • a device personality module 422 connected to the virtual network machine 412 contains a collection of data modules which permit devices and other network entities to be configured and managed with the network management system.
  • a protocol personality module 424 connected to the device communication manager contains a collection of data modules which permit communication with all devices that communicate using the network management protocols specified by the module 424.
  • the personality modules 420, 422 and 424 provide a system that is highly flexible and user configurable. By altering the personality module 420, the user can specify customized views or displays. By changing the device personality module 422, the user can add new types of network devices to the system. Similarly, by changing the protocol personality module 424, the network management system can operate with new or different network management protocols.
  • the personality modules permit the system to be reconfigured and customized without changing the basic control code of the system.
  • the hardware for supporting the system of Fig. 12 is typically a workstation such as a Sun Model 3 or 4, or a 386 PC compatible computer running Unix. A minimum of 8 megabytes of memory is required with a display device which supports a minimum of 640 x 680 pixels x 256 color resolution.
  • the basic software includes a Unix release that supports sockets, X-windows and Open Software Foundation Motif 1.0.
  • the network management system is implemented using the C++ programming language, but could be implemented in other object-oriented languages such as Eiffel, Smalltalk, ADA, or the like.
  • the virtual network machine 412 and the device communication manager 414 may be run on a separate computer from the user interface 410 for increased operating speed.
  • the network includes workstations 430, 431, 432, 433 and disk units 434 and 435 interconnected by a data bus 436.
  • Workstations 430 and 431 and disk unit 434 are located in a room 438, and workstations 432 and 433 and disk unit 435 are located in a room 440.
  • the rooms 438 and 440 are located within a building 442.
  • Network devices 444, 445 and 446 are interconnected by a data bus 447 and are located in a building 448 at the same site as building 442.
  • the network portions in buildings 42 and 448 are interconnected by a bridge 450.
  • the network devices in building 452 are interconnected to the network in building 448 by interface devices 459 and 460, which may communicate by a packet switching system, a microwave link or a satellite link.
  • the network management system shown in Fig. 12 and described above is connected to the network of Fig. 13 at any convenient point, such as data bus 436.
  • the network management system shown in Fig. 12 performs two major operations during normal operation. It services user requests entered by the user at user interface 410 and provides network information such as alarms and events to user interface 410.
  • the virtual network machine 412 polls the network to obtain information for updating the network models as described hereinafter.
  • the network devices send status information to the network management system automatically without polling. In either case, the information received from the network is processed so that the topology, operational status, faults and other information pertaining to the network are presented to the user in a systematized and organized manner.
  • Each model includes a number a attributes and one or more inference handlers.
  • the attributes are data which define the characteristics and status of the network entity being modeled.
  • Basic attributes include a model name, a model type name, a model type handle, a polling interval, a next-time-to-poll, a retry count, a contact status, an activation status, a time-of-last-poll and statistics pertaining to the network entity which is being modeled. Polling of network devices will be described hereinafter.
  • attributes that are unique to a particular type of network device can be defined. For example, a network bridge will contain an address source table that defines the devices that are heard on each port of the bridge. A model of the network bridge can contain, as one of its attributes, a copy of the table.
  • the models used in the virtual network machine also include one or more inference handlers.
  • An inference handler is a C++ object which performs a specified computation, decision, action or inference.
  • the inference handlers collectively constitute the intelligence of the model.
  • An individual inference handler is defined by the type of processing performed, the source or sources of the stimulus and the destination of the result.
  • the result is an output of an inference handler and may include attribute changes, creation or destruction of models, alarms or any other valid output.
  • the operation of the inference handler is initiated by a trigger, which is an event occurring in the virtual network machine. Triggers include attribute changes in the same model, attribute changes in another model, relation changes, events, model creation or destruction, and the like.
  • each model includes inference handlers which perform specified functions upon the occurrence of predetermined events which trigger the inference handlers.
  • the inference handlers may process the source address lists according to the methods of this invention (Figs. 9-11).
  • a schematic diagram of a simple model configuration is shown in Fig. 14 to illustrate the concepts of this management system.
  • a device model 480 includes attributes 1 to x and inference handlers 1 to y.
  • a device model 482 includes attributes 1 to u and inference handlers 1 to v.
  • a connect relation 484 indicates that models 480 and 482 are connected in the physical network.
  • a room model 486 includes attributes 1 to m and inference handlers 1 to n.
  • a relation 488 indicates that model 480 is contained within room model 486, and a relation 490 indicates that model 482 is contained within room model 486.
  • Each of the models and the model relations shown in Fig. 14 is implemented as a C++ object. It will be understood that a representation of an actual network would be much more complex than the configuration shown in Fig. 14.
  • the collection of models and model relations in the virtual network machine form a representation of the physical network being managed.
  • the models represent not only the configuration of the network, but also represent its status on a dynamic basis.
  • the status of the network and other information and data relating to the network is obtained by the models in a number of different ways.
  • a primary technique for obtaining information from the network involves polling.
  • a model in the virtual network machine 412 requests the device communication manager 414 to poll the network device which corresponds to the model .
  • the device communication manager 414 converts the request to the necessary protocol for communicating with the network device.
  • the network device returns the requested information to the device communication manager 414, which extracts the device information and forwards it to the virtual network machine 412 for updating one or more attributes in the model of the network device.
  • the polling interval is specified individually for each model and corresponding network device, depending on the importance of the attribute, the frequency with which it is likely to change, and the like.
  • the polling interval in general, is a compromise between a desire that the models accurately reflect the present status of the network device and a desire to minimize network management traffic which could adversely impact normal network operation.
  • the network devices automatically transmit information to the network management system upon the occurrence of significant events without polling. This requires that the network devices be preprogrammed for such operation.
  • the network entity being modeled is not capable of communicating its status to the network management system.
  • models of buildings or rooms containing network devices and models of cables cannot communicate with the corresponding network entities.
  • the status of the network entity is inferred by the model from information contained in models of other network devices. Since successful polling of a network device connected to a cable may indicate that the cable is functioning properly, the status of the cable can be inferred from information contained in a model of the attached network device. Similarly, the operational status of a room can be inferred from the operational status contained in models of the network devices located within the room.
  • a model watch In order for a model to make such inferences, it is necessary for the model to obtain information from related models.
  • a function called a model watch an attribute in one model is monitored or watched by one or more other models. A change in the watched attribute may trigger inference handlers in the watching models.
  • One method is to make a manual data entry. Another is to conduct a "ping sweep" in which requests are sent blindly to a range of addresses, and the active devices respond with a reply. Another is to read system tables, such as the network name servers. Another is to read the address translation tables; and still another is to read the tables f om known network devices.
  • the above techniques are most useful for determining the presence of some network node at a given address; identification of the node as a data-relay device is then accomplished via the management protocol. In many cases, it will be necessary to translate the "management layer address" to a network address; the source address tables will typically utilize management level addresses, whereas the management protocol usually operates via network layer addresses.

Abstract

An apparatus and method for determining the topology of a computer network including data-relay devices (A, B, C) and node devices (d, ... k) based on a comparison of source addresses heard by the various data relay-devices. A source address table is compiled for each port of each data-relay device, and for each select pair of ports the addresses in the source tables are compared to determine whether there is an intersection of heard devices. In order to account for directed transmissions which are not heard at every port, further comparison is made of all of the other ports of the device, eliminating the ports for which the intersection is the empty set. From the determined connections, a topology of the network is graphed showing direct and transitive connections. In cases where there is both a direct and transitive connection, the redundant direct connection is eliminated.

Description

APPARATUS AND METHOD FOR DETERMINING NETWORK TOPOLOGY
Cross Reference To Related Application This is a continuation-in-part of copending and commonly assigned U.S. Serial No. 07/583,509 filed September 17, 1990 entitled "NETWORK MANAGEMENT SYSTEM USING MODEL-BASED INTELLIGENCE" by R.Dev et al.
Field of the Invention This invention relates to a system for determining the topology of a computer network which includes data-relay devices and node devices, the topology being determined based on a comparison of source addresses "heard" by the various data-relay devices.
Background of the Invention Computer networks are widely used to provide increased computing power, sharing of resources and communication between users. Computer systems and computer system components are interconnected to form a network. Networks may include a number of computer devices within a room, building or site that are interconnected by a high speed local data link such as local area network (LAN), token ring, Ethernet, or the like. Local networks in different locations may be interconnected by techniques such as packet switching, microwave links and satellite links to form a world-wide network. A network may include several hundred or more interconnected devices.
In computer networks, a number of issues arise, including traffic overload on parts of the network, optimum placement of network resources, security, isolation of network faults, and the like. These issues become more complex and difficult as networks become larger and more complex. For example, if a network device is not sending messages, it may be difficult to determine whether the fault is in the network device itself, the data communication link or an intermediate network device between the sending and receiving network devices.
Network management systems have been utilized in the past in attempts to address such issues. Prior art network management systems typically operated by remote access to and monitoring of information from network devices. The network management system collected large volumes of information which required evaluation by a network administrator. Prior art network management systems place a tremendous burden on the network administrator. He/she must be a networking expert in order to understand the implications of a change in a network device parameter. The administrator must also understand the topology of each section of the network in order to understand what may have caused the change. In addition, the administrator must sift. through reams of information and false alarms in order.to determine the cause of a problem.
An important aspect of any network management system is its ability to accurately represent interactions within the network and between network devices. Toward this end it is crucial that any network management system be provided with some means to determine the topology of the network.
It is a general object of the present invention to provide a method and apparatus for determining the topology of a computer network, which information may be used in managing the network.
It is another object of the present invention to provide an apparatuε and method which utilizes lists of source addresses "heard" by each port of a data-relay device (e.g., a bridge) and compares the same to addresses heard by other data-relay devices, in order to determine the topology of the network.
It is a further object of the present invention to provide an apparatus and method for determining connections which are not directly "heard" due to a directed transmission, but which connection can be determined by inference from the addresses heard by other ports of the data-relay device.
It is yet another object of the present invention to provide an apparatus and method which determines both transitive and direct connections between data-relay devices and for which redundant direct connections are eliminated to provide the final network topology.
Summary of the Invention According to the present invention, these and other objects and advantages are achieved in a method and apparatus for determining a network topology.
According to the method of the present invention, a list of network addresses "heard" by each port of a data-relay device, such as a bridge, is compiled for each data-relay device in a computer network. The network includes a plurality of data-relay devices and node devices interconnected by a data bus. The node devices (e.g., work stations or disk units) transmit data on the bus via data packets which include a source address. The data-relay devices also transmit their own source address with replies sent to the network management system. Each of the data-relay devices acquires and maintains a source address table which lists the addresses heard by each port of the data-relay device. These lists are then compared to determine whether there is a direct or transitive connection between select ports on different data-relay devices and the resulting connections are used to define a topology showing the interconnections between the devices in the network.
More specifically, if a port X. (on data-relay device X) hears another data-relay device Y, we know that there is at least a transitive (if not a direct) connection between some port Y. and X.. If there is a port Y. that hears
X, then we can conclude that X. is connected to Y. Due to directed transmissions, there will not always be a port Y. that hears X. In this case, another method is used to determine which port of Y is connected to X.. The correct port is found by excluding all ports that cannot be connected to port X..
More specifically, a node device z that is not in the set of addresses heard by X, must be in some set X where q does not equal i. If port Y. is connected to port X. , then z must be in set Y. and in no other set Y such that r does not equal j. Thus, to determine which port of Y is connected to port X. we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. If enough of the set X 'ε and Y 's are empty, there may be no ports remaining and the technique fails. Otherwise, one is left with one port Y which is the port that is connected to X^.
If the above methods are applied to each data- relay port in the network, one can then graph the interconnections between data-relay devices to define the network topology. As a final step, redundant connections are detected by checking each direct connection for the existence of a transitive connection between the same ports. If such a transitive connection exists, then the direct connection is redundant and is eliminated.
The apparatus of the present invention is a system for use with a digital computer network and preferably with a network management system. In a preferred embodiment, the network management system includes a virtual network having models representing network devices, each model containing network data relating to the corresponding network device and means for processing the network data to provide user information. The system includes means for transferring network data from the network devices to the corresponding models, and means for supplying user information from the virtual network to a user. Thus, a model of each data-relay device includes as network data a list of source addresses heard by each port of the data-relay device and the processing means are inference handlers which compare the sets of addresses of the various data-relay devices to determine the topology of the computer network. There are means for updating the network data,, e.g., by poling the network entities. The determined network topology is used by the management system for controlling traffic on the network, optimizing network resources, security, isolation of network faults, and the like. In addition, the determined topology may be provided to the user by a video display unit connected to the computer.
These and other advantages of the present invention will be further defined by the following drawings and detailed description.
Brief Description of the Drawings
Fig. 1 is a schematic diagram showing an example of a network;
Fig. 2 is a schematic diagram showing the partition imposed by data-relay device A in the network of Fig. 1;
Fig. 3 is a schematic diagram showing the partition imposed by data-relay device B in the network of Fig. 1;
Fig. 4 is a schematic diagram showing the partition imposed by data-relay device C in the network of Fig. 1;
Fig. 5 is a schematic diagram showing the combined partitions of the data-relay devices A, B and C in the network of Fig. 1;
Fig..6 is a schematic diagram showing trial connections between the data-relay devices of the network of Fig. 1, determined according to the present invention;
Fig. 7a is a schematic diagram showing the final connections between the data-relay devices in the network of Fig. 1, determined according to the present invention, and Fig. 7b is the same as Fig. 7a but further includes the node devices between the data-relay devices;
Fig. 8 shows an example of the source address tables for each port of the data-relay devices in the network of Fig. 1; Fig. 9 is a flow chart showing a method of determining a connection between data-relay devices according to the present invention;
Fig. 10 is a flow chart showing a further method of determining connections between data-relay devices according to the present invention;
Fig. 11a is a flow chart showing a method of making the trial connections and final connections (see Figs. 6-7) according to the present invention to determine the topology of the network of Fig. 1; Figs, lib and lie are schematic diagrams showing the elimination of redundant connections in two sample networks;
Fig. 12 is a block diagram of a network management system which may utilize the present invention;
Fig. 13 is a block diagram showing an example of a network connected to the network management system of Fig. 13; and
Fig. 14 is a schematic diagram showing the structure of models and the relations between models in the network management system of Fig. 13 which utilizes the present invention.
Detailed Description
The present invention is an apparatus and method for determining the topology of a computer network consisting of data-relay devices and node devices, which topology may be computed automatically by obtaining and processing information from the network devices.
A data-relay device is a network element that provides electrical and/or logical isolation of local area network segments. A data-relay device forwards network packets from one port of the device to other ports of the device, amplifying and retiming the electrical signals to (a) increase the maximum cable length, (b) increase the number of permissible nodes on the network, and/or (c) increase the number of network segments. Examples of data-relay devices include bridges, repeaters, hubs and the like. Bridges are defined in IEEE 802.1(d) specification, and repeaters are defined in IEEE 802.1(e) specification. In addition to the above functions, bridges also provide a storage and forward mechanism with packet filtering, to reduce the network load. Hubs are bridges but forward all traffic without filtering. For use in this invention, the hub must be an "intelligent" hub with a source address table for network management as described below.
In accordance with this invention, each data-relay device on the network has a source address table, which is a list of network addresses "heard" by each port of the device. When a network node transmits a packet, the source address of the transmitting device is included in the packet; when the packet is received by the data-relay device, the source address is stored in the source address table, along with the port that the packet was received on.
Entries in the source address table are usually "aged," so that an entry exists in the table for a finite period of time unless refreshed by another packet being received from the same source address on the same port. This maintains the currency of the table, since a device which is physically removed from the network will eventually disappear from the source address table.
If a node device on the network has not yet transmitted a packet, i.e., an inactive node, the data-relay device will not receive a packet from that node for entry on the source address table. However, this can be corrected for by periodically poling all of the node devices on the network to confirm the existence of all devices on the network.
In order to prevent loops in the network, and thus duplication of packets, data-relay devices must not be connected redundantly. Some devices, such as IEEE 802.1(d) compliant bridges, use a spanning tree algorithm to automatically detect and disable redundant devices. In performing their primary function of packet forwarding, data-relay devices are essentially transparent to the network; They do not transmit packets with their own addresses. However, data-relay devices do transmit responses to network management requests with their own address in the source address field of the packet. This feature is important to the computation of the network topology according to the present invention since, otherwise, the data-relay devices would not appear in each other's source address tables.
Some data-relay devices perform directed transmissions of management packets. Thus, instead of the management response being transmitted on all ports of the device, it is transmitted only on the port where the management request was received. This is usually evident in bridges where the management packet itself is subject to a filtering process. When a data-relay device uses directed transmission of management packets, the address of the device may or may not appear in the source address tables of other data-relay devices, depending on the relative placement of the data-relay devices and the management station. This fact is taken into account in accordance with the present invention when computing the network topology form the source address tables as described below.
In order to describe the operation of the present invention, a sample network is shown in Fig. 1. The network includes three data-relay devices, 10, 12 and 14, which have been identified as devices A, B and C. each with two numbered ports. The data-relay devices separate the network into three sections or partitions, with end-node devices d-k and management station m at various locations on the network. The partitions generated by devices A, B and C are shown in Figs. 2, 3 and 4, respectively. The combination of partitions is shown in Fig. 5. In summary:
• partition 9 includes node devices d, e and m connected to port 1 of data-relay device A; • partition 11 includes node devices f and g and between port 2 of data-relay device A and port 1 of data-relay device C;
• partition 13 includes node devices h and i between port 2 of data-relay device C and port 1 of data-relay device B; and
• partition 15 includes node devices j and k connected to port 2 of data-relay device B.
The source address table for each data-relay device reflects the partitioning. Each port "hears" all the devices contained in its partition. Thus, the sets of devices heard by the various ports are:
Al = (d, e, m)
A2 = (B, C, f, g, h, i, j, k)
Cl = (A, d, e, m, f, g)
C2 = (B, h, i, j, k)
Bl = (A, d, e, m, f, g, h, i)
B2 = (j, k) The above address tables for each port are shown schematically in Fig. 8.
The data-relay devices may not be heard by other data-relay devices due to directed transmission of management packets. For example, note that Bl does not hear device C, since in this example C is directing its management traffic toward the management station m. However, since in this example A does not direct its management traffic, Bl does hear A.
A transitive connection is defined as a connection between two nodes that crosses one or more data-relay devices. For example, in Fig. 1 port A2 is transitively connected to port Bl via device C.
In contrast, a direct connection has no data-relay device intervening. Again, referring to Fig. 1, port A2 is directly connected to port Cl.
A first method of the present invention for determining connections between the data-relay devices is illustrated by the flow chart of Fig. 9, wherein the address tables for a pair of data-relay devices X and Y, such as devices A and C in Fig. 8. According to the method, if a port X. hears another data-relay device Y (step 200), we know that there is at least a transitive (if not direct) connection between some port Y. and X.. If there is a port Y. that hears X (step 202), then we can conclude that X. is connected to Y. (step 205). For example, C is an element of A2, and A is an element of Cl, so A2 is connected to Cl (see boxes around relevant entries in the address tables of Fig. 8). If X. does not hear Y then there is no connection between X. and Y (step 201) and we advance to the next port on X (step 207). We repeat this process for every port on X (step 206), incrementing to the next port on X (step 207), and incrementing to the next data-relay device (step 208) when we reach the last port on X. Similarly, we repeat the process for every port on Y (step 203), incrementing to the next port on Y (step 204) .
Due ,to directed transmissions, there will not always be a port Y. that hears X. In this case, another method must be used to determine which port of Y is connected to X.. We find the correct port by excluding all ports that cannot be connected to X.. This method is illustrated in the flow chart of Fig. 10.
Thus a node device z that is not in the set X. must be in some set X such that q does not equal i. If port Y. is connected to port X., then z must be in the set Y. , and in no other set Y such that r does not equal j . In other words, for all q not equal to i, there is some intersection of the sets Xq and Yr if p cort Xl- is connected to port Y . If there is no intersection between X and Y (i.e., the null set), then port X. is not connected to port Y .
To determine which port of Y is connected to X. , we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. Thus, in step 300 of Fig. 10, we consider whether Y hears X . If the answer is, "yes" then there is a connection between X. and Y (step 301); if the answer is, "no" we check whether this is the last port on Y (step 302) and, if not, we increment to the next port on Y (step 303), After checking all of the ports of Y, if the intersection is the null set, then no connection can be determined between X. and any port on Y (the method fails)(step 304).
If we apply the above methods of Figs. 9 and 10 to each data-relay port in the network, one can graph the interconnections between the data-relay devices according to the trial connections shown in Fig. 6. Thus, the following direct connections are shown in Fig. 6:
• connection 20 between port A2 and port Cl;
• connection 22 between port C2 and port Bl; and
• connection 24 between port A2 and port Bl.
The connections, or edges between the two ports include both transitive connections as well as direct connections. If a transitive connection exists, then the direct connection is redundant and should be eliminated. For example, as shown in Fig. 7, because there is a transitive connection between A2 and Bl, the direct connection between A2 and Bl has been eliminated.
The method for eliminating redundant connections is illustrated in the flow chart of Fig. 11a. According to step 320, we first form trial connections between X. and Y. , as shown in Fig. 6. Next, we check whether X. is directly connected to Y. (step 321). If there is no direct connection, we check whether this is the last port on X (step 322) and either increment to the next port on X (step 323) or, if this is the last port on X, we increment to the next device (step 324). If we find that X. is directly connected to Y. , we then check whether X. is transitively connected to Y. (step 325). If the answer is, "yes" we remove the direct connection between X. and Y. (step 326) . In traversing from node to node to reach the desired port, the system can set a "flag" at each node visited since revisiting a node serves no purpose. The process of eliminating direct connections where transitive connections exist is illustrated in the simplest possible case in Figs. 6-7a - i.e., three data-relay devices. It should be pointed out that the same technique is applicable to an arbitrarily complex network of bridges, hubs and end-point devices (within the limitations imposed on bridges and hubs, i.e., no loops may exist). This is true because alternate paths involving more than three data-relay devices may consist of a combination of direct and transitive connections and may include branching as well. In Fig. lib, transitive connections (1 to 2 and 2 to 3) eliminate direct connection (1 to 4). Similarly, direction connection (2 to 4) is eliminated as well. In Fig. lie, a three port data-relay device allows for branching in the network; however, the elimination of redundant direct connections (5 to 7, 7 to 8 and 5 to 8 as shown) proceeds exactly as described for three data-relay devices. In real networks, data-relay devices having more than two ports, and network segments consisting of more than three devices are more the rule than the exception.
The topology of Fig. 7 can be further defined to include the node devices in each network segment. Each node device belongs to a segment if its address appears in the intersection of the port sets X. of each data-relay port which is directly connected to the segment. For example, as shown in Fig. 8, node devices h and i appear in the address tables for each ports C2 and Bl and, therefore, nodes h and i should be included as part of the connection 22 between devices A and C. Fig. 7b shows a topology further including the node devices.
In the preferred embodiment described herein, the method and apparatus for determining network topology form a part of a network management system, or provide network topology information to a network management system, such as that described in copending and commonly assigned U.S. Application Serial No. 07/583,509, filed September 17, 1990, entitled, NETWORK MANAGEMENT SYSTEM USING MODEL BASED INTELLIGENCE, by R. Dev et al; that application was also filed as PCT/US91/06725 on September 17, 1991 and published as Int'l Publ. No. WO92/05485 on April 2, 1992; that application is hereby incorporated by reference in its entirety. However, the present invention may also be used in combination with other network management systems, or in any application where it is desired to obtain the network topology.
In accordance with the network management system of the previously identified application, which is incorporated by reference, Fig. 12 shows a block diagram of the system.
The major components of the network management system are a user interface 410, a virtual network machine 412, and a device communication manager 414. The user interface 410, which may include a video display screen, keyboard, mouse and printer, provides all interaction with the user. The user interface controls the screen, keyboard, mouse and printer and provides the user with different views of the network that is being managed. The user interface receives network information from the virtual network machine 412. The virtual network machine 412 contains a software representation of the network being managed, including models that represent the devices and other entities associated with the network, and relations between the models. The virtual network machine 412 is associated with a database manager 416 which manages the storage and retrieval of disk-based data. Such data includes configuration data, an event log, statistics, history and current state information. The device communication manager 414 is connected to a network 418 and handles communication between the virtual network machine 412 and network devices. The data received from the network devices is provided by the device communication manager to the virtual network machine 412. The device communication manager 414 converts generic requests from the virtual network machine 412 to the required network management protocol for communicating with each network device. Existing network management protocols include Simple Network Management Protocol (SNMP), Internet Control Message Protocol (ICMP) and many proprietary network management protocols. Certain types of network devices are designed to communicate with a network management system using one of these protocols.
A view personality module 420 connected to the user interface 410 contains a collection of data modules which permit the user interface to provide different views of the network. A device personality module 422 connected to the virtual network machine 412 contains a collection of data modules which permit devices and other network entities to be configured and managed with the network management system. A protocol personality module 424 connected to the device communication manager contains a collection of data modules which permit communication with all devices that communicate using the network management protocols specified by the module 424. The personality modules 420, 422 and 424 provide a system that is highly flexible and user configurable. By altering the personality module 420, the user can specify customized views or displays. By changing the device personality module 422, the user can add new types of network devices to the system. Similarly, by changing the protocol personality module 424, the network management system can operate with new or different network management protocols. The personality modules permit the system to be reconfigured and customized without changing the basic control code of the system.
The hardware for supporting the system of Fig. 12 is typically a workstation such as a Sun Model 3 or 4, or a 386 PC compatible computer running Unix. A minimum of 8 megabytes of memory is required with a display device which supports a minimum of 640 x 680 pixels x 256 color resolution. The basic software includes a Unix release that supports sockets, X-windows and Open Software Foundation Motif 1.0. The network management system is implemented using the C++ programming language, but could be implemented in other object-oriented languages such as Eiffel, Smalltalk, ADA, or the like. The virtual network machine 412 and the device communication manager 414 may be run on a separate computer from the user interface 410 for increased operating speed.
An example of a network connected to this management system is shown in Fig. 13. The network includes workstations 430, 431, 432, 433 and disk units 434 and 435 interconnected by a data bus 436. Workstations 430 and 431 and disk unit 434 are located in a room 438, and workstations 432 and 433 and disk unit 435 are located in a room 440. The rooms 438 and 440 are located within a building 442. Network devices 444, 445 and 446 are interconnected by a data bus 447 and are located in a building 448 at the same site as building 442. The network portions in buildings 42 and 448 are interconnected by a bridge 450. A building 452 remotely located (in a different city, state or country) from buildings 442 and 448, contains network devices 453, 454, 455 and 456 interconnected by a data bus 457. The network devices in building 452 are interconnected to the network in building 448 by interface devices 459 and 460, which may communicate by a packet switching system, a microwave link or a satellite link. The network management system shown in Fig. 12 and described above is connected to the network of Fig. 13 at any convenient point, such as data bus 436.
In general, the network management system shown in Fig. 12 performs two major operations during normal operation. It services user requests entered by the user at user interface 410 and provides network information such as alarms and events to user interface 410. In addition, the virtual network machine 412 polls the network to obtain information for updating the network models as described hereinafter. In some cases, the network devices send status information to the network management system automatically without polling. In either case, the information received from the network is processed so that the topology, operational status, faults and other information pertaining to the network are presented to the user in a systematized and organized manner.
Each model includes a number a attributes and one or more inference handlers. The attributes are data which define the characteristics and status of the network entity being modeled. Basic attributes include a model name, a model type name, a model type handle, a polling interval, a next-time-to-poll, a retry count, a contact status, an activation status, a time-of-last-poll and statistics pertaining to the network entity which is being modeled. Polling of network devices will be described hereinafter. In addition, attributes that are unique to a particular type of network device can be defined. For example, a network bridge will contain an address source table that defines the devices that are heard on each port of the bridge. A model of the network bridge can contain, as one of its attributes, a copy of the table.
The models used in the virtual network machine also include one or more inference handlers. An inference handler is a C++ object which performs a specified computation, decision, action or inference. The inference handlers collectively constitute the intelligence of the model. An individual inference handler is defined by the type of processing performed, the source or sources of the stimulus and the destination of the result. The result is an output of an inference handler and may include attribute changes, creation or destruction of models, alarms or any other valid output. The operation of the inference handler is initiated by a trigger, which is an event occurring in the virtual network machine. Triggers include attribute changes in the same model, attribute changes in another model, relation changes, events, model creation or destruction, and the like. Thus, each model includes inference handlers which perform specified functions upon the occurrence of predetermined events which trigger the inference handlers. The inference handlers may process the source address lists according to the methods of this invention (Figs. 9-11).
A schematic diagram of a simple model configuration is shown in Fig. 14 to illustrate the concepts of this management system. A device model 480 includes attributes 1 to x and inference handlers 1 to y. A device model 482 includes attributes 1 to u and inference handlers 1 to v. A connect relation 484 indicates that models 480 and 482 are connected in the physical network. A room model 486 includes attributes 1 to m and inference handlers 1 to n. A relation 488 indicates that model 480 is contained within room model 486, and a relation 490 indicates that model 482 is contained within room model 486. Each of the models and the model relations shown in Fig. 14 is implemented as a C++ object. It will be understood that a representation of an actual network would be much more complex than the configuration shown in Fig. 14.
As discussed above, the collection of models and model relations in the virtual network machine form a representation of the physical network being managed. The models represent not only the configuration of the network, but also represent its status on a dynamic basis. The status of the network and other information and data relating to the network is obtained by the models in a number of different ways. A primary technique for obtaining information from the network involves polling. At specified intervals, a model in the virtual network machine 412 requests the device communication manager 414 to poll the network device which corresponds to the model . The device communication manager 414 converts the request to the necessary protocol for communicating with the network device. The network device returns the requested information to the device communication manager 414, which extracts the device information and forwards it to the virtual network machine 412 for updating one or more attributes in the model of the network device. The polling interval is specified individually for each model and corresponding network device, depending on the importance of the attribute, the frequency with which it is likely to change, and the like. The polling interval, in general, is a compromise between a desire that the models accurately reflect the present status of the network device and a desire to minimize network management traffic which could adversely impact normal network operation.
According to another technique for updating the information contained in the models, the network devices automatically transmit information to the network management system upon the occurrence of significant events without polling. This requires that the network devices be preprogrammed for such operation.
It will be understood that communication between a model and its corresponding network entity is possible only for certain types of devices such as bridges, card racks, hubs, etc. In other cases, the network entity being modeled is not capable of communicating its status to the network management system. For example, models of buildings or rooms containing network devices and models of cables cannot communicate with the corresponding network entities. In this case, the status of the network entity is inferred by the model from information contained in models of other network devices. Since successful polling of a network device connected to a cable may indicate that the cable is functioning properly, the status of the cable can be inferred from information contained in a model of the attached network device. Similarly, the operational status of a room can be inferred from the operational status contained in models of the network devices located within the room. In order for a model to make such inferences, it is necessary for the model to obtain information from related models. In a function called a model watch, an attribute in one model is monitored or watched by one or more other models. A change in the watched attribute may trigger inference handlers in the watching models.
There are several methods to acquire a list of the data-relay devices in the network. One method is to make a manual data entry. Another is to conduct a "ping sweep" in which requests are sent blindly to a range of addresses, and the active devices respond with a reply. Another is to read system tables, such as the network name servers. Another is to read the address translation tables; and still another is to read the tables f om known network devices. The above techniques are most useful for determining the presence of some network node at a given address; identification of the node as a data-relay device is then accomplished via the management protocol. In many cases, it will be necessary to translate the "management layer address" to a network address; the source address tables will typically utilize management level addresses, whereas the management protocol usually operates via network layer addresses.
While there have been shown and described select embodiments of the present invention, it will be obvious to those skilled in the art that various changes and modifications may be made therein without departing from the scope of the invention as defined by the appended claims.
What is claimed is:

Claims

1. A method for determining the topology of a computer network including node devices and data-relay devices, said method including the steps of: acquiring a list of data-relay devices in the computer network; acquiring a source address table for each port of each data-relay device, the source address table being a compilation of addresses of node devices and data-relay devices which communicate with the respective port; from the source address table, compiling sets of addresses heard by each port of each data-relay device for identifying connections between the data-relay devices and determining the topology of the computer network.
2. A method as defined in claim 1 wherein the step of determining the topology of the computer network includes: determining direct and transitive connections between the data-relay devices; generating a graph of connections between the data-relay devices; and removing transitive connections between the data-relay devices from the graph.
3. A method as defined in claim 1 wherein the step of determining the topology of the computer network includes: selecting a pair of ports X. and Y. on different data-relay devices X and Y; comparing the source address tables to determine whether port X. hears device Y and whether port Y. hears device X, and if the answer to both is affirmative establishing a connection between ports X. and Y. in the topology; wherein the above determining step is repeated for each different pair of ports of the data-relay devices in the network.
4. A method according to claim 1 wherein the step of determining the topology of the computer network includes: selecting a pair of ports X. and Y. on different data-relay devices X and Y; comparing the source address tables to determine whether port Y hears port X where q is not equal to i and r is not equal to j, and if the answer is affirmative, establishing a connection between X. and Y ; and wherein the above determining step is repeated for each different pair of ports of the data-relay devices in the network where Y. does not hear X.
5. A method according to claim 1 wherein the determined topology of the computer network is utilized by a network management system.
6. A method for determining the topology of a computer network including node devices and data-relay devices, the method comprising the steps of: acquiring a list of data-relay devices in the computer network; acquiring a source address table for each port of each of the data-relay devices, the source address table being a compilation of addresses of node devices and data-relay devices which communicate with the respective port; selecting a pair of ports X. and Y. on different data-relay devices X and Y; determining whether port X. hears device Y and whether port Y. hears device X, and if the answer to both is affirmative, establishing a connection between ports X. and Y. in the topology; repeating the above, selecting and determining steps for each different pair of ports of the data-relay devices in the network; for each pair of ports wherein Y. does not hear X, determining whether Y hears X where r is not equal to j, and q is not equal to i, and if the answer is affirmative, establishing a connection between X. and Y ; repeating the second determining step for each pair for which Y. does not hear X; generating a graph of connections between the data-relay devices to show the network topology; and removing direct connections between X. and Y. where there is a transitive connection between X. and Y..
7. A system for determining the topology of a computer network having node devices and data-relay devices, the apparatus comprising: means for storing a source address table of source addresses heard by each port of each data-relay device in the network; means for processing the addresses in the source address tables of different pairs of ports of different data-relay devices to determine whether there is a connection between the respective ports; and means for supplying information regarding the connections between the data-relay devices to a network management system.
8. A system as defined in claim 7 further including means for supplying information regarding the connections between the data-relay devices to a user.
9. A system for determining the topology of a computer network including node devices and data-relay devices, the system comprising: a virtual network including a plurality of models representing each of the data-relay devices, each model including a source address table which includes the addresses of the node devices which are heard by each port of the respective data-relay device, and each model including means for comparing the addresses in the source address tables for ports in different pairs of data-relay devices to determine whether there is a connection between the ports of the data-relay devices, and means for generating a graph of connections between the data-relay devices; means for transferring data from the network devices to the corresponding models in the virtual network in order to maintain the source address tables; and means for supplying the topology information from the virtual network to a user.
10. A system as defined in claim 9 wherein the models are software models.
11. A method as defined in claim 10 wherein the models contain network data relating to the corresponding network device and one or more inference handlers for processing the network data to provide user information.
EP94929738A 1993-09-01 1994-08-29 Apparatus and method for determining network topology Withdrawn EP0724795A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US115232 1987-10-30
US11523293A 1993-09-01 1993-09-01
PCT/US1994/009690 WO1995006989A1 (en) 1993-09-01 1994-08-29 Apparatus and method for determining network topology

Publications (1)

Publication Number Publication Date
EP0724795A1 true EP0724795A1 (en) 1996-08-07

Family

ID=22360078

Family Applications (1)

Application Number Title Priority Date Filing Date
EP94929738A Withdrawn EP0724795A1 (en) 1993-09-01 1994-08-29 Apparatus and method for determining network topology

Country Status (4)

Country Link
EP (1) EP0724795A1 (en)
JP (1) JPH09504913A (en)
AU (1) AU675362B2 (en)
WO (1) WO1995006989A1 (en)

Families Citing this family (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6883034B1 (en) 1995-06-23 2005-04-19 Cisco Technology, Inc. Method of resolving conflicts in access control lists in router by comparing elements in the lists based on subsumption relations
US6393486B1 (en) 1995-06-23 2002-05-21 Cisco Technology, Inc. System and method using level three protocol information for network centric problem analysis and topology construction of actual or planned routed network
EP0861546B1 (en) * 1995-11-16 2004-04-07 Loran Network Systems, L.L.C. Method of determining the topology of a network of objects
US5710885A (en) * 1995-11-28 1998-01-20 Ncr Corporation Network management system with improved node discovery and monitoring
WO1997049214A1 (en) * 1996-06-24 1997-12-24 Netsys Technologies, Inc. Method and apparatus for network centric problem analysis and topology construction
US6115362A (en) * 1997-03-28 2000-09-05 Cabletron Systems, Inc. Method and apparatus for determining frame relay connections
US6128729A (en) * 1997-12-16 2000-10-03 Hewlett-Packard Company Method and system for automatic configuration of network links to attached devices
US6108702A (en) * 1998-12-02 2000-08-22 Micromuse, Inc. Method and apparatus for determining accurate topology features of a network
CA2268495C (en) 1998-12-16 2008-11-18 Loran Network Management Ltd. Method for determining computer network topologies
US8040869B2 (en) * 2001-12-19 2011-10-18 Alcatel Lucent Method and apparatus for automatic discovery of logical links between network devices
US7856599B2 (en) 2001-12-19 2010-12-21 Alcatel-Lucent Canada Inc. Method and system for IP link management
DE10163606A1 (en) * 2001-12-21 2003-07-10 Daimler Chrysler Ag Operating network with several subscribers involves subscriber receiving network messages during initializing phase and forming topology of other types in accordance with received message
EP2587736A3 (en) 2007-09-26 2013-08-28 Nicira, Inc. Network operating system for managing and securing networks
EP2415221B1 (en) 2009-04-01 2014-05-07 Nicira, Inc. Method and apparatus for implementing and managing virtual switches
US10103939B2 (en) 2010-07-06 2018-10-16 Nicira, Inc. Network control apparatus and method for populating logical datapath sets
US8964528B2 (en) 2010-07-06 2015-02-24 Nicira, Inc. Method and apparatus for robust packet distribution among hierarchical managed switching elements
US8958292B2 (en) 2010-07-06 2015-02-17 Nicira, Inc. Network control apparatus and method with port security controls
US9525647B2 (en) 2010-07-06 2016-12-20 Nicira, Inc. Network control apparatus and method for creating and modifying logical switching elements
US9680750B2 (en) 2010-07-06 2017-06-13 Nicira, Inc. Use of tunnels to hide network addresses
US9043452B2 (en) 2011-05-04 2015-05-26 Nicira, Inc. Network control apparatus and method for port isolation
JP5941703B2 (en) * 2012-02-27 2016-06-29 株式会社日立製作所 Management server and management method
US9282019B2 (en) 2013-07-12 2016-03-08 Nicira, Inc. Tracing logical network packets through physical network
US9344349B2 (en) 2013-07-12 2016-05-17 Nicira, Inc. Tracing network packets by a cluster of network controllers
US9407580B2 (en) 2013-07-12 2016-08-02 Nicira, Inc. Maintaining data stored with a packet
US9264330B2 (en) 2013-10-13 2016-02-16 Nicira, Inc. Tracing host-originated logical network packets
US10193771B2 (en) 2013-12-09 2019-01-29 Nicira, Inc. Detecting and handling elephant flows
US9967199B2 (en) 2013-12-09 2018-05-08 Nicira, Inc. Inspecting operations of a machine to detect elephant flows
US9419889B2 (en) 2014-03-07 2016-08-16 Nicira, Inc. Method and system for discovering a path of network traffic
US9419874B2 (en) 2014-03-27 2016-08-16 Nicira, Inc. Packet tracing in a software-defined networking environment
US9553803B2 (en) 2014-06-30 2017-01-24 Nicira, Inc. Periodical generation of network measurement data
US9379956B2 (en) 2014-06-30 2016-06-28 Nicira, Inc. Identifying a network topology between two endpoints
US10469342B2 (en) 2014-10-10 2019-11-05 Nicira, Inc. Logical network traffic analysis
US10200306B2 (en) 2017-03-07 2019-02-05 Nicira, Inc. Visualization of packet tracing operation results
JPWO2019031258A1 (en) * 2017-08-08 2020-08-13 ソニー株式会社 Transmission terminal, transmission method, information processing terminal, and information processing method
US10608887B2 (en) 2017-10-06 2020-03-31 Nicira, Inc. Using packet tracing tool to automatically execute packet capture operations
US11570090B2 (en) 2020-07-29 2023-01-31 Vmware, Inc. Flow tracing operation in container cluster
US11196628B1 (en) 2020-07-29 2021-12-07 Vmware, Inc. Monitoring container clusters
US11558426B2 (en) 2020-07-29 2023-01-17 Vmware, Inc. Connection tracking for container cluster
US11736436B2 (en) 2020-12-31 2023-08-22 Vmware, Inc. Identifying routes with indirect addressing in a datacenter
US11336533B1 (en) 2021-01-08 2022-05-17 Vmware, Inc. Network visualization of correlations between logical elements and associated physical elements
US11687210B2 (en) 2021-07-05 2023-06-27 Vmware, Inc. Criteria-based expansion of group nodes in a network topology visualization
US11711278B2 (en) 2021-07-24 2023-07-25 Vmware, Inc. Visualization of flow trace operation across multiple sites
US11677645B2 (en) 2021-09-17 2023-06-13 Vmware, Inc. Traffic monitoring

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
ATE194238T1 (en) * 1990-09-17 2000-07-15 Cabletron Systems Inc SYSTEM AND METHOD FOR MODELING A COMPUTER NETWORK
US5179554A (en) * 1991-04-08 1993-01-12 Digital Equipment Corporation Automatic association of local area network station addresses with a repeater port
US5297138A (en) * 1991-04-30 1994-03-22 Hewlett-Packard Company Determining physical topology across repeaters and bridges in a computer network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO9506989A1 *

Also Published As

Publication number Publication date
AU7868994A (en) 1995-03-22
AU675362B2 (en) 1997-01-30
WO1995006989A1 (en) 1995-03-09
JPH09504913A (en) 1997-05-13

Similar Documents

Publication Publication Date Title
US5727157A (en) Apparatus and method for determining a computer network topology
AU675362B2 (en) Determination of network topology
EP1234407B1 (en) Network event correlation system using protocol models
US6205122B1 (en) Automatic network topology analysis
US6115743A (en) Interface system for integrated monitoring and management of network devices in a telecommunication network
AU700018B2 (en) Method and apparatus for testing the responsiveness of a network device
EP0737920B1 (en) Method for isolating a network fault
US5751933A (en) System for determining the status of an entity in a computer network
US6697338B1 (en) Determination of physical topology of a communication network
US5964837A (en) Computer network management using dynamic switching between event-driven and polling type of monitoring from manager station
US5774669A (en) Scalable hierarchical network management system for displaying network information in three dimensions
US5978845A (en) Network management relay mechanism
US20030225876A1 (en) Method and apparatus for graphically depicting network performance and connectivity
US20020165934A1 (en) Displaying a subset of network nodes based on discovered attributes
JP2005237018A (en) Data transmission to network management system
US7646729B2 (en) Method and apparatus for determination of network topology
CN110620693A (en) Railway station route remote restart control system and method based on Internet of things
GECKIL Apparatus and method for determining a computer network topology
JPH09247146A (en) Network management system
Wilbur Local area network management for distributed applications
Simmons et al. Knowledge sharing between distributed knowledge based systems
Markley et al. Management of space networks
Fried et al. Implementing Integrated Monitoring Systems for Heterogeneous Networks
Deodhar et al. Distributed analyzer architecture permits real-time monitoring of heterogeneous local area networks
Muller Managing the Enterprise Network with LightWatch/Open

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 19960320

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LI LU MC NL PT SE

17Q First examination report despatched

Effective date: 19971110

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 19980321