WO1995006989A1 - Apparatus and method for determining network topology - Google Patents

Apparatus and method for determining network topology

Info

Publication number
WO1995006989A1
WO1995006989A1 PCT/US1994/009690 US9409690W WO1995006989A1 WO 1995006989 A1 WO1995006989 A1 WO 1995006989A1 US 9409690 W US9409690 W US 9409690W WO 1995006989 A1 WO1995006989 A1 WO 1995006989A1
Authority
WO
Grant status
Application
Patent type
Prior art keywords
network
devices
data
relay
port
Prior art date
Application number
PCT/US1994/009690
Other languages
French (fr)
Inventor
Timothy L. Orr
Eric W. Gray
Original Assignee
Cabletron Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/02Topology update or discovery
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance or administration or management of packet switching networks
    • H04L41/12Arrangements for maintenance or administration or management of packet switching networks network topology discovery or management

Abstract

An apparatus and method for determining the topology of a computer network including data-relay devices (A, B, C) and node devices (d, ... k) based on a comparison of source addresses heard by the various data relay-devices. A source address table is compiled for each port of each data-relay device, and for each select pair of ports the addresses in the source tables are compared to determine whether there is an intersection of heard devices. In order to account for directed transmissions which are not heard at every port, further comparison is made of all of the other ports of the device, eliminating the ports for which the intersection is the empty set. From the determined connections, a topology of the network is graphed showing direct and transitive connections. In cases where there is both a direct and transitive connection, the redundant direct connection is eliminated.

Description

APPARATUS AND METHOD FOR DETERMINING NETWORK TOPOLOGY

Cross Reference To Related Application This is a continuation-in-part of copending and commonly assigned U.S. Serial No. 07/583,509 filed September 17, 1990 entitled "NETWORK MANAGEMENT SYSTEM USING MODEL-BASED INTELLIGENCE" by R.Dev et al.

Field of the Invention This invention relates to a system for determining the topology of a computer network which includes data-relay devices and node devices, the topology being determined based on a comparison of source addresses "heard" by the various data-relay devices.

Background of the Invention Computer networks are widely used to provide increased computing power, sharing of resources and communication between users. Computer systems and computer system components are interconnected to form a network. Networks may include a number of computer devices within a room, building or site that are interconnected by a high speed local data link such as local area network (LAN), token ring, Ethernet, or the like. Local networks in different locations may be interconnected by techniques such as packet switching, microwave links and satellite links to form a world-wide network. A network may include several hundred or more interconnected devices.

In computer networks, a number of issues arise, including traffic overload on parts of the network, optimum placement of network resources, security, isolation of network faults, and the like. These issues become more complex and difficult as networks become larger and more complex. For example, if a network device is not sending messages, it may be difficult to determine whether the fault is in the network device itself, the data communication link or an intermediate network device between the sending and receiving network devices.

Network management systems have been utilized in the past in attempts to address such issues. Prior art network management systems typically operated by remote access to and monitoring of information from network devices. The network management system collected large volumes of information which required evaluation by a network administrator. Prior art network management systems place a tremendous burden on the network administrator. He/she must be a networking expert in order to understand the implications of a change in a network device parameter. The administrator must also understand the topology of each section of the network in order to understand what may have caused the change. In addition, the administrator must sift. through reams of information and false alarms in order.to determine the cause of a problem.

An important aspect of any network management system is its ability to accurately represent interactions within the network and between network devices. Toward this end it is crucial that any network management system be provided with some means to determine the topology of the network.

It is a general object of the present invention to provide a method and apparatus for determining the topology of a computer network, which information may be used in managing the network.

It is another object of the present invention to provide an apparatuε and method which utilizes lists of source addresses "heard" by each port of a data-relay device (e.g., a bridge) and compares the same to addresses heard by other data-relay devices, in order to determine the topology of the network.

It is a further object of the present invention to provide an apparatus and method for determining connections which are not directly "heard" due to a directed transmission, but which connection can be determined by inference from the addresses heard by other ports of the data-relay device.

It is yet another object of the present invention to provide an apparatus and method which determines both transitive and direct connections between data-relay devices and for which redundant direct connections are eliminated to provide the final network topology.

Summary of the Invention According to the present invention, these and other objects and advantages are achieved in a method and apparatus for determining a network topology.

According to the method of the present invention, a list of network addresses "heard" by each port of a data-relay device, such as a bridge, is compiled for each data-relay device in a computer network. The network includes a plurality of data-relay devices and node devices interconnected by a data bus. The node devices (e.g., work stations or disk units) transmit data on the bus via data packets which include a source address. The data-relay devices also transmit their own source address with replies sent to the network management system. Each of the data-relay devices acquires and maintains a source address table which lists the addresses heard by each port of the data-relay device. These lists are then compared to determine whether there is a direct or transitive connection between select ports on different data-relay devices and the resulting connections are used to define a topology showing the interconnections between the devices in the network.

More specifically, if a port X. (on data-relay device X) hears another data-relay device Y, we know that there is at least a transitive (if not a direct) connection between some port Y. and X.. If there is a port Y. that hears

X, then we can conclude that X. is connected to Y. Due to directed transmissions, there will not always be a port Y. that hears X. In this case, another method is used to determine which port of Y is connected to X.. The correct port is found by excluding all ports that cannot be connected to port X..

More specifically, a node device z that is not in the set of addresses heard by X, must be in some set X where q does not equal i. If port Y. is connected to port X. , then z must be in set Y. and in no other set Y such that r does not equal j. Thus, to determine which port of Y is connected to port X. we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. If enough of the set X 'ε and Y 's are empty, there may be no ports remaining and the technique fails. Otherwise, one is left with one port Y which is the port that is connected to X^.

If the above methods are applied to each data- relay port in the network, one can then graph the interconnections between data-relay devices to define the network topology. As a final step, redundant connections are detected by checking each direct connection for the existence of a transitive connection between the same ports. If such a transitive connection exists, then the direct connection is redundant and is eliminated.

The apparatus of the present invention is a system for use with a digital computer network and preferably with a network management system. In a preferred embodiment, the network management system includes a virtual network having models representing network devices, each model containing network data relating to the corresponding network device and means for processing the network data to provide user information. The system includes means for transferring network data from the network devices to the corresponding models, and means for supplying user information from the virtual network to a user. Thus, a model of each data-relay device includes as network data a list of source addresses heard by each port of the data-relay device and the processing means are inference handlers which compare the sets of addresses of the various data-relay devices to determine the topology of the computer network. There are means for updating the network data,, e.g., by poling the network entities. The determined network topology is used by the management system for controlling traffic on the network, optimizing network resources, security, isolation of network faults, and the like. In addition, the determined topology may be provided to the user by a video display unit connected to the computer.

These and other advantages of the present invention will be further defined by the following drawings and detailed description.

Brief Description of the Drawings

Fig. 1 is a schematic diagram showing an example of a network;

Fig. 2 is a schematic diagram showing the partition imposed by data-relay device A in the network of Fig. 1;

Fig. 3 is a schematic diagram showing the partition imposed by data-relay device B in the network of Fig. 1;

Fig. 4 is a schematic diagram showing the partition imposed by data-relay device C in the network of Fig. 1;

Fig. 5 is a schematic diagram showing the combined partitions of the data-relay devices A, B and C in the network of Fig. 1;

Fig..6 is a schematic diagram showing trial connections between the data-relay devices of the network of Fig. 1, determined according to the present invention;

Fig. 7a is a schematic diagram showing the final connections between the data-relay devices in the network of Fig. 1, determined according to the present invention, and Fig. 7b is the same as Fig. 7a but further includes the node devices between the data-relay devices;

Fig. 8 shows an example of the source address tables for each port of the data-relay devices in the network of Fig. 1; Fig. 9 is a flow chart showing a method of determining a connection between data-relay devices according to the present invention;

Fig. 10 is a flow chart showing a further method of determining connections between data-relay devices according to the present invention;

Fig. 11a is a flow chart showing a method of making the trial connections and final connections (see Figs. 6-7) according to the present invention to determine the topology of the network of Fig. 1; Figs, lib and lie are schematic diagrams showing the elimination of redundant connections in two sample networks;

Fig. 12 is a block diagram of a network management system which may utilize the present invention;

Fig. 13 is a block diagram showing an example of a network connected to the network management system of Fig. 13; and

Fig. 14 is a schematic diagram showing the structure of models and the relations between models in the network management system of Fig. 13 which utilizes the present invention.

Detailed Description

The present invention is an apparatus and method for determining the topology of a computer network consisting of data-relay devices and node devices, which topology may be computed automatically by obtaining and processing information from the network devices.

A data-relay device is a network element that provides electrical and/or logical isolation of local area network segments. A data-relay device forwards network packets from one port of the device to other ports of the device, amplifying and retiming the electrical signals to (a) increase the maximum cable length, (b) increase the number of permissible nodes on the network, and/or (c) increase the number of network segments. Examples of data-relay devices include bridges, repeaters, hubs and the like. Bridges are defined in IEEE 802.1(d) specification, and repeaters are defined in IEEE 802.1(e) specification. In addition to the above functions, bridges also provide a storage and forward mechanism with packet filtering, to reduce the network load. Hubs are bridges but forward all traffic without filtering. For use in this invention, the hub must be an "intelligent" hub with a source address table for network management as described below.

In accordance with this invention, each data-relay device on the network has a source address table, which is a list of network addresses "heard" by each port of the device. When a network node transmits a packet, the source address of the transmitting device is included in the packet; when the packet is received by the data-relay device, the source address is stored in the source address table, along with the port that the packet was received on.

Entries in the source address table are usually "aged," so that an entry exists in the table for a finite period of time unless refreshed by another packet being received from the same source address on the same port. This maintains the currency of the table, since a device which is physically removed from the network will eventually disappear from the source address table.

If a node device on the network has not yet transmitted a packet, i.e., an inactive node, the data-relay device will not receive a packet from that node for entry on the source address table. However, this can be corrected for by periodically poling all of the node devices on the network to confirm the existence of all devices on the network.

In order to prevent loops in the network, and thus duplication of packets, data-relay devices must not be connected redundantly. Some devices, such as IEEE 802.1(d) compliant bridges, use a spanning tree algorithm to automatically detect and disable redundant devices. In performing their primary function of packet forwarding, data-relay devices are essentially transparent to the network; They do not transmit packets with their own addresses. However, data-relay devices do transmit responses to network management requests with their own address in the source address field of the packet. This feature is important to the computation of the network topology according to the present invention since, otherwise, the data-relay devices would not appear in each other's source address tables.

Some data-relay devices perform directed transmissions of management packets. Thus, instead of the management response being transmitted on all ports of the device, it is transmitted only on the port where the management request was received. This is usually evident in bridges where the management packet itself is subject to a filtering process. When a data-relay device uses directed transmission of management packets, the address of the device may or may not appear in the source address tables of other data-relay devices, depending on the relative placement of the data-relay devices and the management station. This fact is taken into account in accordance with the present invention when computing the network topology form the source address tables as described below.

In order to describe the operation of the present invention, a sample network is shown in Fig. 1. The network includes three data-relay devices, 10, 12 and 14, which have been identified as devices A, B and C. each with two numbered ports. The data-relay devices separate the network into three sections or partitions, with end-node devices d-k and management station m at various locations on the network. The partitions generated by devices A, B and C are shown in Figs. 2, 3 and 4, respectively. The combination of partitions is shown in Fig. 5. In summary:

• partition 9 includes node devices d, e and m connected to port 1 of data-relay device A; • partition 11 includes node devices f and g and between port 2 of data-relay device A and port 1 of data-relay device C;

• partition 13 includes node devices h and i between port 2 of data-relay device C and port 1 of data-relay device B; and

• partition 15 includes node devices j and k connected to port 2 of data-relay device B.

The source address table for each data-relay device reflects the partitioning. Each port "hears" all the devices contained in its partition. Thus, the sets of devices heard by the various ports are:

Al = (d, e, m)

A2 = (B, C, f, g, h, i, j, k)

Cl = (A, d, e, m, f, g)

C2 = (B, h, i, j, k)

Bl = (A, d, e, m, f, g, h, i)

B2 = (j, k) The above address tables for each port are shown schematically in Fig. 8.

The data-relay devices may not be heard by other data-relay devices due to directed transmission of management packets. For example, note that Bl does not hear device C, since in this example C is directing its management traffic toward the management station m. However, since in this example A does not direct its management traffic, Bl does hear A.

A transitive connection is defined as a connection between two nodes that crosses one or more data-relay devices. For example, in Fig. 1 port A2 is transitively connected to port Bl via device C.

In contrast, a direct connection has no data-relay device intervening. Again, referring to Fig. 1, port A2 is directly connected to port Cl.

A first method of the present invention for determining connections between the data-relay devices is illustrated by the flow chart of Fig. 9, wherein the address tables for a pair of data-relay devices X and Y, such as devices A and C in Fig. 8. According to the method, if a port X. hears another data-relay device Y (step 200), we know that there is at least a transitive (if not direct) connection between some port Y. and X.. If there is a port Y. that hears X (step 202), then we can conclude that X. is connected to Y. (step 205). For example, C is an element of A2, and A is an element of Cl, so A2 is connected to Cl (see boxes around relevant entries in the address tables of Fig. 8). If X. does not hear Y then there is no connection between X. and Y (step 201) and we advance to the next port on X (step 207). We repeat this process for every port on X (step 206), incrementing to the next port on X (step 207), and incrementing to the next data-relay device (step 208) when we reach the last port on X. Similarly, we repeat the process for every port on Y (step 203), incrementing to the next port on Y (step 204) .

Due ,to directed transmissions, there will not always be a port Y. that hears X. In this case, another method must be used to determine which port of Y is connected to X.. We find the correct port by excluding all ports that cannot be connected to X.. This method is illustrated in the flow chart of Fig. 10.

Thus a node device z that is not in the set X. must be in some set X such that q does not equal i. If port Y. is connected to port X., then z must be in the set Y. , and in no other set Y such that r does not equal j . In other words, for all q not equal to i, there is some intersection of the sets Xq and Yr if p cort Xl- is connected to port Y . If there is no intersection between X and Y (i.e., the null set), then port X. is not connected to port Y .

To determine which port of Y is connected to X. , we test each port of Y against the other ports of X, eliminating the ports of Y for which the intersection is the empty set. Thus, in step 300 of Fig. 10, we consider whether Y hears X . If the answer is, "yes" then there is a connection between X. and Y (step 301); if the answer is, "no" we check whether this is the last port on Y (step 302) and, if not, we increment to the next port on Y (step 303), After checking all of the ports of Y, if the intersection is the null set, then no connection can be determined between X. and any port on Y (the method fails)(step 304).

If we apply the above methods of Figs. 9 and 10 to each data-relay port in the network, one can graph the interconnections between the data-relay devices according to the trial connections shown in Fig. 6. Thus, the following direct connections are shown in Fig. 6:

• connection 20 between port A2 and port Cl;

• connection 22 between port C2 and port Bl; and

• connection 24 between port A2 and port Bl.

The connections, or edges between the two ports include both transitive connections as well as direct connections. If a transitive connection exists, then the direct connection is redundant and should be eliminated. For example, as shown in Fig. 7, because there is a transitive connection between A2 and Bl, the direct connection between A2 and Bl has been eliminated.

The method for eliminating redundant connections is illustrated in the flow chart of Fig. 11a. According to step 320, we first form trial connections between X. and Y. , as shown in Fig. 6. Next, we check whether X. is directly connected to Y. (step 321). If there is no direct connection, we check whether this is the last port on X (step 322) and either increment to the next port on X (step 323) or, if this is the last port on X, we increment to the next device (step 324). If we find that X. is directly connected to Y. , we then check whether X. is transitively connected to Y. (step 325). If the answer is, "yes" we remove the direct connection between X. and Y. (step 326) . In traversing from node to node to reach the desired port, the system can set a "flag" at each node visited since revisiting a node serves no purpose. The process of eliminating direct connections where transitive connections exist is illustrated in the simplest possible case in Figs. 6-7a - i.e., three data-relay devices. It should be pointed out that the same technique is applicable to an arbitrarily complex network of bridges, hubs and end-point devices (within the limitations imposed on bridges and hubs, i.e., no loops may exist). This is true because alternate paths involving more than three data-relay devices may consist of a combination of direct and transitive connections and may include branching as well. In Fig. lib, transitive connections (1 to 2 and 2 to 3) eliminate direct connection (1 to 4). Similarly, direction connection (2 to 4) is eliminated as well. In Fig. lie, a three port data-relay device allows for branching in the network; however, the elimination of redundant direct connections (5 to 7, 7 to 8 and 5 to 8 as shown) proceeds exactly as described for three data-relay devices. In real networks, data-relay devices having more than two ports, and network segments consisting of more than three devices are more the rule than the exception.

The topology of Fig. 7 can be further defined to include the node devices in each network segment. Each node device belongs to a segment if its address appears in the intersection of the port sets X. of each data-relay port which is directly connected to the segment. For example, as shown in Fig. 8, node devices h and i appear in the address tables for each ports C2 and Bl and, therefore, nodes h and i should be included as part of the connection 22 between devices A and C. Fig. 7b shows a topology further including the node devices.

In the preferred embodiment described herein, the method and apparatus for determining network topology form a part of a network management system, or provide network topology information to a network management system, such as that described in copending and commonly assigned U.S. Application Serial No. 07/583,509, filed September 17, 1990, entitled, NETWORK MANAGEMENT SYSTEM USING MODEL BASED INTELLIGENCE, by R. Dev et al; that application was also filed as PCT/US91/06725 on September 17, 1991 and published as Int'l Publ. No. WO92/05485 on April 2, 1992; that application is hereby incorporated by reference in its entirety. However, the present invention may also be used in combination with other network management systems, or in any application where it is desired to obtain the network topology.

In accordance with the network management system of the previously identified application, which is incorporated by reference, Fig. 12 shows a block diagram of the system.

The major components of the network management system are a user interface 410, a virtual network machine 412, and a device communication manager 414. The user interface 410, which may include a video display screen, keyboard, mouse and printer, provides all interaction with the user. The user interface controls the screen, keyboard, mouse and printer and provides the user with different views of the network that is being managed. The user interface receives network information from the virtual network machine 412. The virtual network machine 412 contains a software representation of the network being managed, including models that represent the devices and other entities associated with the network, and relations between the models. The virtual network machine 412 is associated with a database manager 416 which manages the storage and retrieval of disk-based data. Such data includes configuration data, an event log, statistics, history and current state information. The device communication manager 414 is connected to a network 418 and handles communication between the virtual network machine 412 and network devices. The data received from the network devices is provided by the device communication manager to the virtual network machine 412. The device communication manager 414 converts generic requests from the virtual network machine 412 to the required network management protocol for communicating with each network device. Existing network management protocols include Simple Network Management Protocol (SNMP), Internet Control Message Protocol (ICMP) and many proprietary network management protocols. Certain types of network devices are designed to communicate with a network management system using one of these protocols.

A view personality module 420 connected to the user interface 410 contains a collection of data modules which permit the user interface to provide different views of the network. A device personality module 422 connected to the virtual network machine 412 contains a collection of data modules which permit devices and other network entities to be configured and managed with the network management system. A protocol personality module 424 connected to the device communication manager contains a collection of data modules which permit communication with all devices that communicate using the network management protocols specified by the module 424. The personality modules 420, 422 and 424 provide a system that is highly flexible and user configurable. By altering the personality module 420, the user can specify customized views or displays. By changing the device personality module 422, the user can add new types of network devices to the system. Similarly, by changing the protocol personality module 424, the network management system can operate with new or different network management protocols. The personality modules permit the system to be reconfigured and customized without changing the basic control code of the system.

The hardware for supporting the system of Fig. 12 is typically a workstation such as a Sun Model 3 or 4, or a 386 PC compatible computer running Unix. A minimum of 8 megabytes of memory is required with a display device which supports a minimum of 640 x 680 pixels x 256 color resolution. The basic software includes a Unix release that supports sockets, X-windows and Open Software Foundation Motif 1.0. The network management system is implemented using the C++ programming language, but could be implemented in other object-oriented languages such as Eiffel, Smalltalk, ADA, or the like. The virtual network machine 412 and the device communication manager 414 may be run on a separate computer from the user interface 410 for increased operating speed.

An example of a network connected to this management system is shown in Fig. 13. The network includes workstations 430, 431, 432, 433 and disk units 434 and 435 interconnected by a data bus 436. Workstations 430 and 431 and disk unit 434 are located in a room 438, and workstations 432 and 433 and disk unit 435 are located in a room 440. The rooms 438 and 440 are located within a building 442. Network devices 444, 445 and 446 are interconnected by a data bus 447 and are located in a building 448 at the same site as building 442. The network portions in buildings 42 and 448 are interconnected by a bridge 450. A building 452 remotely located (in a different city, state or country) from buildings 442 and 448, contains network devices 453, 454, 455 and 456 interconnected by a data bus 457. The network devices in building 452 are interconnected to the network in building 448 by interface devices 459 and 460, which may communicate by a packet switching system, a microwave link or a satellite link. The network management system shown in Fig. 12 and described above is connected to the network of Fig. 13 at any convenient point, such as data bus 436.

In general, the network management system shown in Fig. 12 performs two major operations during normal operation. It services user requests entered by the user at user interface 410 and provides network information such as alarms and events to user interface 410. In addition, the virtual network machine 412 polls the network to obtain information for updating the network models as described hereinafter. In some cases, the network devices send status information to the network management system automatically without polling. In either case, the information received from the network is processed so that the topology, operational status, faults and other information pertaining to the network are presented to the user in a systematized and organized manner.

Each model includes a number a attributes and one or more inference handlers. The attributes are data which define the characteristics and status of the network entity being modeled. Basic attributes include a model name, a model type name, a model type handle, a polling interval, a next-time-to-poll, a retry count, a contact status, an activation status, a time-of-last-poll and statistics pertaining to the network entity which is being modeled. Polling of network devices will be described hereinafter. In addition, attributes that are unique to a particular type of network device can be defined. For example, a network bridge will contain an address source table that defines the devices that are heard on each port of the bridge. A model of the network bridge can contain, as one of its attributes, a copy of the table.

The models used in the virtual network machine also include one or more inference handlers. An inference handler is a C++ object which performs a specified computation, decision, action or inference. The inference handlers collectively constitute the intelligence of the model. An individual inference handler is defined by the type of processing performed, the source or sources of the stimulus and the destination of the result. The result is an output of an inference handler and may include attribute changes, creation or destruction of models, alarms or any other valid output. The operation of the inference handler is initiated by a trigger, which is an event occurring in the virtual network machine. Triggers include attribute changes in the same model, attribute changes in another model, relation changes, events, model creation or destruction, and the like. Thus, each model includes inference handlers which perform specified functions upon the occurrence of predetermined events which trigger the inference handlers. The inference handlers may process the source address lists according to the methods of this invention (Figs. 9-11).

A schematic diagram of a simple model configuration is shown in Fig. 14 to illustrate the concepts of this management system. A device model 480 includes attributes 1 to x and inference handlers 1 to y. A device model 482 includes attributes 1 to u and inference handlers 1 to v. A connect relation 484 indicates that models 480 and 482 are connected in the physical network. A room model 486 includes attributes 1 to m and inference handlers 1 to n. A relation 488 indicates that model 480 is contained within room model 486, and a relation 490 indicates that model 482 is contained within room model 486. Each of the models and the model relations shown in Fig. 14 is implemented as a C++ object. It will be understood that a representation of an actual network would be much more complex than the configuration shown in Fig. 14.

As discussed above, the collection of models and model relations in the virtual network machine form a representation of the physical network being managed. The models represent not only the configuration of the network, but also represent its status on a dynamic basis. The status of the network and other information and data relating to the network is obtained by the models in a number of different ways. A primary technique for obtaining information from the network involves polling. At specified intervals, a model in the virtual network machine 412 requests the device communication manager 414 to poll the network device which corresponds to the model . The device communication manager 414 converts the request to the necessary protocol for communicating with the network device. The network device returns the requested information to the device communication manager 414, which extracts the device information and forwards it to the virtual network machine 412 for updating one or more attributes in the model of the network device. The polling interval is specified individually for each model and corresponding network device, depending on the importance of the attribute, the frequency with which it is likely to change, and the like. The polling interval, in general, is a compromise between a desire that the models accurately reflect the present status of the network device and a desire to minimize network management traffic which could adversely impact normal network operation.

According to another technique for updating the information contained in the models, the network devices automatically transmit information to the network management system upon the occurrence of significant events without polling. This requires that the network devices be preprogrammed for such operation.

It will be understood that communication between a model and its corresponding network entity is possible only for certain types of devices such as bridges, card racks, hubs, etc. In other cases, the network entity being modeled is not capable of communicating its status to the network management system. For example, models of buildings or rooms containing network devices and models of cables cannot communicate with the corresponding network entities. In this case, the status of the network entity is inferred by the model from information contained in models of other network devices. Since successful polling of a network device connected to a cable may indicate that the cable is functioning properly, the status of the cable can be inferred from information contained in a model of the attached network device. Similarly, the operational status of a room can be inferred from the operational status contained in models of the network devices located within the room. In order for a model to make such inferences, it is necessary for the model to obtain information from related models. In a function called a model watch, an attribute in one model is monitored or watched by one or more other models. A change in the watched attribute may trigger inference handlers in the watching models.

There are several methods to acquire a list of the data-relay devices in the network. One method is to make a manual data entry. Another is to conduct a "ping sweep" in which requests are sent blindly to a range of addresses, and the active devices respond with a reply. Another is to read system tables, such as the network name servers. Another is to read the address translation tables; and still another is to read the tables f om known network devices. The above techniques are most useful for determining the presence of some network node at a given address; identification of the node as a data-relay device is then accomplished via the management protocol. In many cases, it will be necessary to translate the "management layer address" to a network address; the source address tables will typically utilize management level addresses, whereas the management protocol usually operates via network layer addresses.

While there have been shown and described select embodiments of the present invention, it will be obvious to those skilled in the art that various changes and modifications may be made therein without departing from the scope of the invention as defined by the appended claims.

What is claimed is:

Claims

1. A method for determining the topology of a computer network including node devices and data-relay devices, said method including the steps of: acquiring a list of data-relay devices in the computer network; acquiring a source address table for each port of each data-relay device, the source address table being a compilation of addresses of node devices and data-relay devices which communicate with the respective port; from the source address table, compiling sets of addresses heard by each port of each data-relay device for identifying connections between the data-relay devices and determining the topology of the computer network.
2. A method as defined in claim 1 wherein the step of determining the topology of the computer network includes: determining direct and transitive connections between the data-relay devices; generating a graph of connections between the data-relay devices; and removing transitive connections between the data-relay devices from the graph.
3. A method as defined in claim 1 wherein the step of determining the topology of the computer network includes: selecting a pair of ports X. and Y. on different data-relay devices X and Y; comparing the source address tables to determine whether port X. hears device Y and whether port Y. hears device X, and if the answer to both is affirmative establishing a connection between ports X. and Y. in the topology; wherein the above determining step is repeated for each different pair of ports of the data-relay devices in the network.
4. A method according to claim 1 wherein the step of determining the topology of the computer network includes: selecting a pair of ports X. and Y. on different data-relay devices X and Y; comparing the source address tables to determine whether port Y hears port X where q is not equal to i and r is not equal to j, and if the answer is affirmative, establishing a connection between X. and Y ; and wherein the above determining step is repeated for each different pair of ports of the data-relay devices in the network where Y. does not hear X.
5. A method according to claim 1 wherein the determined topology of the computer network is utilized by a network management system.
6. A method for determining the topology of a computer network including node devices and data-relay devices, the method comprising the steps of: acquiring a list of data-relay devices in the computer network; acquiring a source address table for each port of each of the data-relay devices, the source address table being a compilation of addresses of node devices and data-relay devices which communicate with the respective port; selecting a pair of ports X. and Y. on different data-relay devices X and Y; determining whether port X. hears device Y and whether port Y. hears device X, and if the answer to both is affirmative, establishing a connection between ports X. and Y. in the topology; repeating the above, selecting and determining steps for each different pair of ports of the data-relay devices in the network; for each pair of ports wherein Y. does not hear X, determining whether Y hears X where r is not equal to j, and q is not equal to i, and if the answer is affirmative, establishing a connection between X. and Y ; repeating the second determining step for each pair for which Y. does not hear X; generating a graph of connections between the data-relay devices to show the network topology; and removing direct connections between X. and Y. where there is a transitive connection between X. and Y..
7. A system for determining the topology of a computer network having node devices and data-relay devices, the apparatus comprising: means for storing a source address table of source addresses heard by each port of each data-relay device in the network; means for processing the addresses in the source address tables of different pairs of ports of different data-relay devices to determine whether there is a connection between the respective ports; and means for supplying information regarding the connections between the data-relay devices to a network management system.
8. A system as defined in claim 7 further including means for supplying information regarding the connections between the data-relay devices to a user.
9. A system for determining the topology of a computer network including node devices and data-relay devices, the system comprising: a virtual network including a plurality of models representing each of the data-relay devices, each model including a source address table which includes the addresses of the node devices which are heard by each port of the respective data-relay device, and each model including means for comparing the addresses in the source address tables for ports in different pairs of data-relay devices to determine whether there is a connection between the ports of the data-relay devices, and means for generating a graph of connections between the data-relay devices; means for transferring data from the network devices to the corresponding models in the virtual network in order to maintain the source address tables; and means for supplying the topology information from the virtual network to a user.
10. A system as defined in claim 9 wherein the models are software models.
11. A method as defined in claim 10 wherein the models contain network data relating to the corresponding network device and one or more inference handlers for processing the network data to provide user information.
PCT/US1994/009690 1993-09-01 1994-08-29 Apparatus and method for determining network topology WO1995006989A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US11523293 true 1993-09-01 1993-09-01
US08/115,232 1993-09-01

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP19940929738 EP0724795A1 (en) 1993-09-01 1994-08-29 Apparatus and method for determining network topology
JP50820094A JPH09504913A (en) 1993-09-01 1994-08-29 Network topology determination method and apparatus

Publications (1)

Publication Number Publication Date
WO1995006989A1 true true WO1995006989A1 (en) 1995-03-09

Family

ID=22360078

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1994/009690 WO1995006989A1 (en) 1993-09-01 1994-08-29 Apparatus and method for determining network topology

Country Status (3)

Country Link
EP (1) EP0724795A1 (en)
JP (1) JPH09504913A (en)
WO (1) WO1995006989A1 (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1997020419A2 (en) * 1995-11-16 1997-06-05 Loran Network Systems, L.L.C. Method of determining the topology of a network of objects
EP0777357A3 (en) * 1995-11-28 1997-10-22 Ncr Int Inc Network management system with improved node discovery and monitoring
WO1997049214A1 (en) * 1996-06-24 1997-12-24 Netsys Technologies, Inc. Method and apparatus for network centric problem analysis and topology construction
WO1998044682A1 (en) * 1997-03-28 1998-10-08 Cabletron Systems, Inc. Method and apparatus for determining frame relay connections
EP0926863A2 (en) * 1997-12-16 1999-06-30 Hewlett-Packard Company Automatic configuration of network links
WO2000036790A1 (en) * 1998-12-16 2000-06-22 Loran Network Management Ltd. Method for determining computer network topologies
US6108702A (en) * 1998-12-02 2000-08-22 Micromuse, Inc. Method and apparatus for determining accurate topology features of a network
US6393486B1 (en) 1995-06-23 2002-05-21 Cisco Technology, Inc. System and method using level three protocol information for network centric problem analysis and topology construction of actual or planned routed network
EP1322068A2 (en) * 2001-12-19 2003-06-25 Alcatel Canada Inc. Method and apparatus for automatic discovery of logical links between network devices
DE10163606A1 (en) * 2001-12-21 2003-07-10 Daimler Chrysler Ag Operating network with several subscribers involves subscriber receiving network messages during initializing phase and forming topology of other types in accordance with received message
US6883034B1 (en) 1995-06-23 2005-04-19 Cisco Technology, Inc. Method of resolving conflicts in access control lists in router by comparing elements in the lists based on subsumption relations
WO2009042919A3 (en) * 2007-09-26 2009-08-06 Nicira Networks Network operating system for managing and securing networks
US7856599B2 (en) 2001-12-19 2010-12-21 Alcatel-Lucent Canada Inc. Method and system for IP link management
US8837493B2 (en) 2010-07-06 2014-09-16 Nicira, Inc. Distributed network control apparatus and method
US8964528B2 (en) 2010-07-06 2015-02-24 Nicira, Inc. Method and apparatus for robust packet distribution among hierarchical managed switching elements
US8966035B2 (en) 2009-04-01 2015-02-24 Nicira, Inc. Method and apparatus for implementing and managing distributed virtual switches in several hosts and physical forwarding elements
US9043452B2 (en) 2011-05-04 2015-05-26 Nicira, Inc. Network control apparatus and method for port isolation
US9264330B2 (en) 2013-10-13 2016-02-16 Nicira, Inc. Tracing host-originated logical network packets
US9282019B2 (en) 2013-07-12 2016-03-08 Nicira, Inc. Tracing logical network packets through physical network
US9344349B2 (en) 2013-07-12 2016-05-17 Nicira, Inc. Tracing network packets by a cluster of network controllers
US9379956B2 (en) 2014-06-30 2016-06-28 Nicira, Inc. Identifying a network topology between two endpoints
US9407580B2 (en) 2013-07-12 2016-08-02 Nicira, Inc. Maintaining data stored with a packet
US9419874B2 (en) 2014-03-27 2016-08-16 Nicira, Inc. Packet tracing in a software-defined networking environment
US9419889B2 (en) 2014-03-07 2016-08-16 Nicira, Inc. Method and system for discovering a path of network traffic
US9525647B2 (en) 2010-07-06 2016-12-20 Nicira, Inc. Network control apparatus and method for creating and modifying logical switching elements
US9548924B2 (en) 2013-12-09 2017-01-17 Nicira, Inc. Detecting an elephant flow based on the size of a packet
US9553803B2 (en) 2014-06-30 2017-01-24 Nicira, Inc. Periodical generation of network measurement data
US9680750B2 (en) 2010-07-06 2017-06-13 Nicira, Inc. Use of tunnels to hide network addresses
US9967199B2 (en) 2013-12-09 2018-05-08 Nicira, Inc. Inspecting operations of a machine to detect elephant flows

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5941703B2 (en) * 2012-02-27 2016-06-29 株式会社日立製作所 Management server and management method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1992005485A2 (en) * 1990-09-17 1992-04-02 Cabletron Systems, Inc. Network management system using model-based intelligence
EP0511851A2 (en) * 1991-04-30 1992-11-04 Hewlett-Packard Company Determining physical topology across repeaters and bridges in a computer network
US5179554A (en) * 1991-04-08 1993-01-12 Digital Equipment Corporation Automatic association of local area network station addresses with a repeater port

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1992005485A2 (en) * 1990-09-17 1992-04-02 Cabletron Systems, Inc. Network management system using model-based intelligence
US5179554A (en) * 1991-04-08 1993-01-12 Digital Equipment Corporation Automatic association of local area network station addresses with a repeater port
EP0511851A2 (en) * 1991-04-30 1992-11-04 Hewlett-Packard Company Determining physical topology across repeaters and bridges in a computer network

Cited By (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7484004B2 (en) 1995-06-23 2009-01-27 Cisco Technology, Inc. Analyzing an access control list for a router to identify a subsumption relation between elements in the list
US6393486B1 (en) 1995-06-23 2002-05-21 Cisco Technology, Inc. System and method using level three protocol information for network centric problem analysis and topology construction of actual or planned routed network
US6883034B1 (en) 1995-06-23 2005-04-19 Cisco Technology, Inc. Method of resolving conflicts in access control lists in router by comparing elements in the lists based on subsumption relations
WO1997020419A2 (en) * 1995-11-16 1997-06-05 Loran Network Systems, L.L.C. Method of determining the topology of a network of objects
US6411997B1 (en) 1995-11-16 2002-06-25 Loran Network Systems Llc Method of determining the topology of a network of objects
WO1997020419A3 (en) * 1995-11-16 1997-11-13 Loran Network Systems Llc Method of determining the topology of a network of objects
EP0777357A3 (en) * 1995-11-28 1997-10-22 Ncr Int Inc Network management system with improved node discovery and monitoring
WO1997049214A1 (en) * 1996-06-24 1997-12-24 Netsys Technologies, Inc. Method and apparatus for network centric problem analysis and topology construction
US6115362A (en) * 1997-03-28 2000-09-05 Cabletron Systems, Inc. Method and apparatus for determining frame relay connections
WO1998044682A1 (en) * 1997-03-28 1998-10-08 Cabletron Systems, Inc. Method and apparatus for determining frame relay connections
EP0926863A3 (en) * 1997-12-16 1999-08-18 Hewlett-Packard Company Automatic configuration of network links
EP0926863A2 (en) * 1997-12-16 1999-06-30 Hewlett-Packard Company Automatic configuration of network links
US6108702A (en) * 1998-12-02 2000-08-22 Micromuse, Inc. Method and apparatus for determining accurate topology features of a network
US6405248B1 (en) 1998-12-02 2002-06-11 Micromuse, Inc. Method and apparatus for determining accurate topology features of a network
EP1171831A4 (en) * 1998-12-02 2004-11-03 Micromuse Inc Method and apparatus for determining accurate topology features of a network
EP1171831A1 (en) * 1998-12-02 2002-01-16 Micromuse, Inc. Method and apparatus for determining accurate topology features of a network
WO2000036790A1 (en) * 1998-12-16 2000-06-22 Loran Network Management Ltd. Method for determining computer network topologies
US6587440B1 (en) 1998-12-16 2003-07-01 Loran Networks Management Ltd. Method for determining computer network topologies
US8040869B2 (en) 2001-12-19 2011-10-18 Alcatel Lucent Method and apparatus for automatic discovery of logical links between network devices
EP1322068A3 (en) * 2001-12-19 2004-06-16 Alcatel Canada Inc. Method and apparatus for automatic discovery of logical links between network devices
EP1322068A2 (en) * 2001-12-19 2003-06-25 Alcatel Canada Inc. Method and apparatus for automatic discovery of logical links between network devices
US7856599B2 (en) 2001-12-19 2010-12-21 Alcatel-Lucent Canada Inc. Method and system for IP link management
DE10163606A1 (en) * 2001-12-21 2003-07-10 Daimler Chrysler Ag Operating network with several subscribers involves subscriber receiving network messages during initializing phase and forming topology of other types in accordance with received message
WO2009042919A3 (en) * 2007-09-26 2009-08-06 Nicira Networks Network operating system for managing and securing networks
US9083609B2 (en) 2007-09-26 2015-07-14 Nicira, Inc. Network operating system for managing and securing networks
US9876672B2 (en) 2007-09-26 2018-01-23 Nicira, Inc. Network operating system for managing and securing networks
US9590919B2 (en) 2009-04-01 2017-03-07 Nicira, Inc. Method and apparatus for implementing and managing virtual switches
US8966035B2 (en) 2009-04-01 2015-02-24 Nicira, Inc. Method and apparatus for implementing and managing distributed virtual switches in several hosts and physical forwarding elements
US8842679B2 (en) 2010-07-06 2014-09-23 Nicira, Inc. Control system that elects a master controller instance for switching elements
US8958292B2 (en) 2010-07-06 2015-02-17 Nicira, Inc. Network control apparatus and method with port security controls
US8966040B2 (en) 2010-07-06 2015-02-24 Nicira, Inc. Use of network information base structure to establish communication between applications
US8964598B2 (en) 2010-07-06 2015-02-24 Nicira, Inc. Mesh architectures for managed switching elements
US8964528B2 (en) 2010-07-06 2015-02-24 Nicira, Inc. Method and apparatus for robust packet distribution among hierarchical managed switching elements
US8959215B2 (en) 2010-07-06 2015-02-17 Nicira, Inc. Network virtualization
US8913483B2 (en) 2010-07-06 2014-12-16 Nicira, Inc. Fault tolerant managed switching element architecture
US9008087B2 (en) 2010-07-06 2015-04-14 Nicira, Inc. Processing requests in a network control system with multiple controller instances
US8880468B2 (en) 2010-07-06 2014-11-04 Nicira, Inc. Secondary storage architecture for a network control system that utilizes a primary network information base
US9049153B2 (en) 2010-07-06 2015-06-02 Nicira, Inc. Logical packet processing pipeline that retains state information to effectuate efficient processing of packets
US9077664B2 (en) 2010-07-06 2015-07-07 Nicira, Inc. One-hop packet processing in a network with managed switching elements
US8837493B2 (en) 2010-07-06 2014-09-16 Nicira, Inc. Distributed network control apparatus and method
US9106587B2 (en) 2010-07-06 2015-08-11 Nicira, Inc. Distributed network control system with one master controller per managed switching element
US9112811B2 (en) 2010-07-06 2015-08-18 Nicira, Inc. Managed switching elements used as extenders
US9172663B2 (en) 2010-07-06 2015-10-27 Nicira, Inc. Method and apparatus for replicating network information base in a distributed network control system with multiple controller instances
US9231891B2 (en) 2010-07-06 2016-01-05 Nicira, Inc. Deployment of hierarchical managed switching elements
US9692655B2 (en) 2010-07-06 2017-06-27 Nicira, Inc. Packet processing in a network with hierarchical managed switching elements
US9680750B2 (en) 2010-07-06 2017-06-13 Nicira, Inc. Use of tunnels to hide network addresses
US9300603B2 (en) 2010-07-06 2016-03-29 Nicira, Inc. Use of rich context tags in logical data processing
US9306875B2 (en) 2010-07-06 2016-04-05 Nicira, Inc. Managed switch architectures for implementing logical datapath sets
US9391928B2 (en) 2010-07-06 2016-07-12 Nicira, Inc. Method and apparatus for interacting with a network information base in a distributed network control system with multiple controller instances
US9363210B2 (en) 2010-07-06 2016-06-07 Nicira, Inc. Distributed network control system with one master controller per logical datapath set
US9525647B2 (en) 2010-07-06 2016-12-20 Nicira, Inc. Network control apparatus and method for creating and modifying logical switching elements
US9007903B2 (en) 2010-07-06 2015-04-14 Nicira, Inc. Managing a network by controlling edge and non-edge switching elements
US9043452B2 (en) 2011-05-04 2015-05-26 Nicira, Inc. Network control apparatus and method for port isolation
US9407580B2 (en) 2013-07-12 2016-08-02 Nicira, Inc. Maintaining data stored with a packet
US9282019B2 (en) 2013-07-12 2016-03-08 Nicira, Inc. Tracing logical network packets through physical network
US9344349B2 (en) 2013-07-12 2016-05-17 Nicira, Inc. Tracing network packets by a cluster of network controllers
US9860151B2 (en) 2013-07-12 2018-01-02 Nicira, Inc. Tracing network packets through logical and physical networks
US9602375B2 (en) 2013-10-13 2017-03-21 Nicira, Inc. Tracing host-originated logical network packets
US9264330B2 (en) 2013-10-13 2016-02-16 Nicira, Inc. Tracing host-originated logical network packets
US9548924B2 (en) 2013-12-09 2017-01-17 Nicira, Inc. Detecting an elephant flow based on the size of a packet
US9967199B2 (en) 2013-12-09 2018-05-08 Nicira, Inc. Inspecting operations of a machine to detect elephant flows
US9838276B2 (en) 2013-12-09 2017-12-05 Nicira, Inc. Detecting an elephant flow based on the size of a packet
US9419889B2 (en) 2014-03-07 2016-08-16 Nicira, Inc. Method and system for discovering a path of network traffic
US9419874B2 (en) 2014-03-27 2016-08-16 Nicira, Inc. Packet tracing in a software-defined networking environment
US9876704B2 (en) 2014-03-27 2018-01-23 Nicira, Inc. Packet tracing in a software-defined networking environment
US9397920B2 (en) 2014-06-30 2016-07-19 Nicira, Inc. Multi-path network bandwidth estimation
US9553803B2 (en) 2014-06-30 2017-01-24 Nicira, Inc. Periodical generation of network measurement data
US9379956B2 (en) 2014-06-30 2016-06-28 Nicira, Inc. Identifying a network topology between two endpoints
US9998369B2 (en) 2014-06-30 2018-06-12 Nicira, Inc. Periodical generation of network measurement data

Also Published As

Publication number Publication date Type
JPH09504913A (en) 1997-05-13 application
EP0724795A1 (en) 1996-08-07 application

Similar Documents

Publication Publication Date Title
US6137782A (en) Automatic network traffic analysis
Breitbart et al. Topology discovery in heterogeneous IP networks
US5922051A (en) System and method for traffic management in a network management system
US7328260B1 (en) Mapping discovered devices to SAN-manageable objects using configurable rules
US6430610B1 (en) TCP/IP address protection mechanism in a clustered server environment
US6058420A (en) Alarm server systems, apparatus, and processes
US6442144B1 (en) Method and apparatus for discovering network devices using internet protocol and producing a corresponding graphical network map
US6260062B1 (en) Element management system for heterogeneous telecommunications network
US5949759A (en) Fault correlation system and method in packet switching networks
US6526044B1 (en) Real-time analysis through capture buffer with real-time historical data correlation
US4937825A (en) Method and apparatus for diagnosing problems in data communication networks
US5321695A (en) Port arrival identification for computer network packets
US5588119A (en) Method for correlating logical device names with a hub port in a local area network
US6119159A (en) Distributed service subsystem protocol for distributed network management
US5293635A (en) Detection on a network by a mapping application of a relative location of a first device to a second device
US6651093B1 (en) Dynamic virtual local area network connection process
US5920257A (en) System and method for isolating an outage within a communications network
US6483812B1 (en) Token ring network topology discovery and display
US6006016A (en) Network fault correlation
US20060002311A1 (en) Network device with VLAN topology discovery functions
US20040155899A1 (en) Method and system for presenting an arrangement of management devices operable in a managed network
US6529954B1 (en) Knowledge based expert analysis system
US20080049627A1 (en) Method and Apparatus for Monitoring Physical Network Topology Information
US7027411B1 (en) Method and system for identifying and processing changes to a network topology
US20060282529A1 (en) Method and apparatus for monitoring physical network topology information

Legal Events

Date Code Title Description
AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

AK Designated states

Kind code of ref document: A1

Designated state(s): AU JP

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 1994929738

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1994929738

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1994929738

Country of ref document: EP