WO2016172060A1 - System and method of updating a network element - Google Patents
System and method of updating a network element Download PDFInfo
- Publication number
- WO2016172060A1 WO2016172060A1 PCT/US2016/028171 US2016028171W WO2016172060A1 WO 2016172060 A1 WO2016172060 A1 WO 2016172060A1 US 2016028171 W US2016028171 W US 2016028171W WO 2016172060 A1 WO2016172060 A1 WO 2016172060A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- network element
- network
- layer
- data processing
- update
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/65—Updates
- G06F8/654—Updates using techniques specially adapted for alterable solid state memories, e.g. for EEPROM or flash memories
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/61—Installation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/08—Configuration management of networks or network elements
- H04L41/0803—Configuration setting
- H04L41/0813—Configuration setting characterised by the conditions triggering a change of settings
- H04L41/082—Configuration setting characterised by the conditions triggering a change of settings the condition being updates or upgrades of network functionality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/65—Updates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/65—Updates
- G06F8/656—Updates while running
Definitions
- This invention relates generally to data networking, and more particularly, to updating network element boot image for a network element in a network without dropping data in the network.
- each network element runs a particular version of software along with a specific configuration for that software. Periodically, the software and/or the configuration of one or more of the network elements are updated.
- the software and/or configuration updates leads to poor performance or disruption because of problems with the software configuration, incompatibilities between different software configurations on different network elements, loss of connectivity due to incorrect Virtual Local Area Network (VLAN) membership change on a port, incorrect Access Control List (ACL) entry, incorrect route metric change, or other types of problems.
- VLAN Virtual Local Area Network
- ACL Access Control List
- the network element is upgraded to a new software image to add new features and/or fix problems with the current software image.
- the network image can be downgraded to a previous version of the software image.
- the network element loads the new software image, marks this new software image as the current software image, and reboots the network element. As the network element boots up, the new software image is loaded and programs the control and data planes of the network element.
- a problem with this way of upgrading or downgrading the software image is that, by rebooting the network element, the network element is inoperable for a relatively long time. For example, rebooting the network element can cause the network element to be off-line for upwards of five minutes. During this off-line time, the data processing functions of the network element are inoperable. This can lead to an interruption of data networking services that rely on the network element.
- a method and apparatus of a device that performs a hitless update a boot image of a network element.
- the device identifies the network element to update and determines if the network element has redundant paths. If this network element has redundant paths, the device configures the network element to drain data processing of the network element. In addition, the device updates the network element to a new boot image when the data processing of the network element is drained.
- Figure 1 is a block diagram of one embodiment of a network with a controller that updates the boot images of the network elements in the network.
- Figure 2 is a flow diagram of one embodiment of a process to update a network element by utilizing redundant network capacity.
- Figure 3 is a flow diagram of one embodiment of a process to drain data processing of a network element.
- Figure 4 is a block diagram of one embodiment of network layer hierarchy.
- Figure 5 is a flow diagram of one embodiment of a process to restore the data processing of a network element.
- Figure 6 is a block diagram of one embodiment of a network update module that updates a network element by utilizing redundant network capacity.
- FIG. 7 is a block diagram of one embodiment of a drain data module that drains the data processing of a network element.
- Figure 8 is a block diagram of one embodiment of a restore data module that restores the data processing of a network element.
- Figure 9 illustrates one example of a typical computer system, which may be used in conjunction with the embodiments described herein.
- Coupled is used to indicate that two or more elements, which may or may not be in direct physical or electrical contact with each other, co-operate or interact with each other.
- Connected is used to indicate the establishment of communication between two or more elements that are coupled with each other.
- processing logic that comprises hardware (e.g., circuitry, dedicated logic, etc.), software (such as is run on a general-purpose computer system or a dedicated machine), or a combination of both.
- processing logic comprises hardware (e.g., circuitry, dedicated logic, etc.), software (such as is run on a general-purpose computer system or a dedicated machine), or a combination of both.
- server The terms “server,” “client,” and “device” are intended to refer generally to data processing systems rather than specifically to a particular form factor for the server, client, and/or device.
- a method and apparatus of a device that performs a hitless update a boot image of a network element.
- a device updates the boot image by putting the network element into a maintenance mode before the network element is updated.
- the network element is part of redundant paths between one or more pairs of endpoints in the network.
- a maintenance mode for a network element is a configuration for the network element where one or more the paths that include the network element are disfavored paths such that other network elements will select other paths that do not include this network element that is in maintenance mode. Furthermore, the network element in maintenance mode is not taken out of the network.
- the device has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in for different forwarding decision making for the other network elements (e.g., routing, switching, and/or other forwarding decisions).
- the device puts the network element in maintenance mode by configuring each layer in a network stack for maintenance mode. For example and in one embodiment, the device configures the routing policy (layer 3) for maintenance mode. In this example, the device increases the cost of routes from that network element to other network elements and/or devices in the network.
- the device verifies the maintenance mode configuration for the routing policy. Furthermore, the device configures and verifies the virtual interface layer (layer 3) for maintenance mode.
- the device additionally configures and verifies the logical interface layer (layer 2) for maintenance mode by configuring the switch forwarding costs for paths that include this network element to be higher such that other network elements would not select these paths.
- the device increases spanning tree costs of paths that include this network element.
- the device further configures and verifies the physical interface layer (layer 1) for maintenance mode.
- the data processing of transported network data being performed by this network element is essentially zero, because this network data is being forwarded through other network elements.
- this network element can be updated with a new boot image with little or no impact on this transported data, because any interruption in the network element would not impact the transported data (e.g., a reboot of the network element during the update process would not interrupt the transported data processing).
- the device rolls back the maintenance mode layer by layer for the network elements, verifying the changes made at each layer, and in reverse order. For example and in one embodiment, the device restores and verifies the physical layer (layer 1) to the configuration this network element had before the software update. With the physical layer configuration restored, the device restores and verifies the virtual layer (layer 2) to the configuration this network element had before the software update.
- the device restores and verifies the logical layer (layer 3), followed by the routing policy, to the configuration this network element had before the software update.
- FIG. 1 is a block diagram of one embodiment of a network 100 with a controller 102 that updates the boot images of the network elements in the network.
- the network 100 includes spine network elements 104A-D that are coupled to each of the leaf network elements 106A-E.
- the leaf network elements 106A-E are further coupled to server racks 108A-E, respectively.
- a spine network element 104A-D is a network element that interconnects the leaf network elements 106A-E.
- each of the spine network elements 104A-D is coupled to each of the leaf network elements 106A-E.
- each of the spine network elements 104A-D is coupled with each other. While in one embodiment, the network elements 104A-D and 106A-E are illustrated in a spine and leaf topology, in alternate embodiments, the network elements 104A-D and 106A- E can be in a different topology. In addition, and in one embodiment, the network elements 104A-E are further coupled to a controller 102 and a database 114. In one embodiment, each of the network elements 104A-D and/or 106A-E can be a router, switch, bridge, gateway, load balancer, firewall, network security device, server, or any other type of device that can receive and process data from a network.
- the network elements 104A-D and 106A-E can be the same or different network elements in terms of manufacturer, type, configuration, or role.
- network elements 104A-D may be routers and network elements 106A-E may be switches.
- network elements 104A-D may be high capacity switches with relatively few 10 gigabit (Gb) or 40Gb ports and network elements 106A-E may be lower capacity switches with a large number of medium capacity port (e.g., 1Gb ports) and high capacity ports to couple to the spine network elements 104A-D.
- Gb gigabit
- 1Gb ports medium capacity port
- the controller 102 is a device that is used to support network- wide services for a network that includes multiple network elements.
- a network- wide service is a service that is used by two or more network elements.
- the network- wide service can be a service that gathers known Media Access Control (MAC) addresses and assembles these addresses into a global MAC address database.
- This global MAC address database can be used to support a VxLAN controller service.
- database 114 is an organized collection of data.
- the ability to rollback or roll forward network elements to different configuration points is a network-wide service.
- the configuration manager 112 can initiate software upgrades that update the boot images for some or all of the network elements 104A-D and 106A-E.
- an update to a boot image of a network element can be a downgrade to boot image previously used with a corresponding lower version software, an upgrade to boot image with a corresponding higher version software, or replacement boot image that has the same version software as currently running on that network element.
- This update to the boot image can be a hitless update or hit full update.
- a hitless update to a network element, or group of network elements is an update to the boot image that does not cause a loss in data processing.
- each of the network elements 104A-D and/or 106A-E includes a network update module 116 that performs the software update.
- a boot image update can be performed as a hitless or a hit full update.
- a hitless update can be performed by draining the data processing needs on a network element that has redundant links with other network elements.
- Figure 2 is a flow diagram of one embodiment of a process 200 to update a network element by utilizing redundant network capacity and putting that network element into maintenance mode.
- process 200 is performed by a network update module, such as the network update module 116 described in Figure 1 above.
- process 200 begins by receiving the network element information for the network element that is to be updated with a new boot image at block 202.
- the network element information received includes information whether the network element has redundant links such that data processing can be diverted to other network elements while the network element boot images being updated.
- process 200 determines if the network element has redundant paths.
- a network element has redundant paths if there are other network elements that can perform the data processing the network element is currently performing.
- the network elements may be part of an equal cost multipath group or a link aggregate group that allows for multiple network elements to appear as one network element. As illustrated in Figure 1 above, network elements 104A-D or 106A-E have redundant paths that allow others of these network elements to take over the data processing for one of the network elements.
- process 200 performs a hitless update on the network element at block 206.
- process 200 performs an accelerated software upgrade on the network element as described in in U.S. Patent Application No. 14/449,030, entitled “System and Method for Accelerated Software Upgrades", filed on July 31, 2014.
- process 200 configures the network element to be in maintenance mode.
- process 200 puts the network element in maintenance mode by draining the data processing from the network element.
- process 200 diverts the flow of data to the other network elements that have redundant paths for path(s) this network element participates in.
- process 200 drains the data processing from the network element by increasing the cost for forwarding the data in this network element.
- the network element will advertise its forwarding costs such that the flow of data will be diverted to the other network elements.
- network element 104A can increase the BGP cost to a large number (e.
- process 200 determines if the maintenance mode configuration was successful. In one embodiment, process 200 checks the maintenance mode configuration at each layer of configuration. If one of those layer configurations fails, process 200 reverts the previous layer configuration(s) to an original configuration for those layers and indicates a failure. If the maintenance mode configuration fails, execution proceeds to block 208 above.
- process 200 performs the boot image update at block 210.
- process 200 can perform the boot image update that involves a reboot of the network element. In this embodiment, even though the network element is rebooted during the boot image update process, this update is hitless because no data is lost during the rebooting.
- Process 200 restores the data processing configuration at block 214. In one embodiment, process 200 restores the data processing configuration by restoring the configuration for each layer of the network protocols for this network element. Restoring the data processing configuration is further described in Figure 5 below.
- process 200 drains the data processing for the network element by configuring the network element by putting the network element into maintenance mode.
- Figure 3 is a flow diagram of one embodiment of a process 300 to put the network element into maintenance mode.
- process 300 is performed by a drain data processing module, such as the drain data processing module 606 as described in Figure 6 below.
- process 300 begins by configuring the routing policy (layer 3) of network element, so as to drain the data processing of the network element at block 302.
- process 300 configures the routing policy (layer 3) of the network element so as to drain the layer 3 routing data processing from the network element.
- process 300 configures the layer 3 routing policy so as to discourage data being routed through this network element.
- the network element advertises the updated layer routing policy, such that other network elements would send the data to be routed to a different path that does not include this network element.
- process 300 can configure the layer routing costs to be higher, so that these routes can be advertised as having the high cost.
- other network element receiving the high cost advertised routes would select lower cost routes for the data.
- process 300 can configure the network element that routes for BGP, OSPF, or other types of routing protocol have a very high cost.
- the network element would advertise these high cost routes to other network elements via one or more of the routing protocols (e.g., BGP, OSPF, and/or another type of routing protocol) and the other network elements receiving these route advertisements would select other paths that do not include this network element.
- the routing protocols e.g., BGP, OSPF, and/or another type of routing protocol
- network element 104D includes routes to network elements 104A-C and 106A-E. For each of these routes, there are alternate layer 3 routing paths between network elements 104A-C and 106A-E that do not go through network element 104D.
- process 300 increases cost of the routes to the other network elements and network element 104D advertises these high cost routes to the other network elements (106A-C and 104A-E). The other network elements 106A-C and 104A-E use the high cost of the advertised routes from network element 104D to make routing decisions.
- the other network elements 104A-C and 106A-E would not route data to network element 104D if there are alternate paths available to route the data.
- the layer 3 routing data processing load can be effectively zero as the other network elements 104A-C and 106A-E will select alternate routes that do not include network element 104D.
- the network element configured by process 300 is not taken out of the network. Instead, process 300 has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in the layer 3 routing decision making for the other network elements.
- process 300 verifies that this configuration leaves the network element in a stable state. In one embodiment, process 300 verifies the routing configuration by that the routing metrics have been increased to a sufficient level for some or all of the routing protocols. As described above, this policy of increasing the routing metrics is to drain the layer 3 data, so that the routing metrics (or admin distance in the case of multiple protocol deployment) are increased to ensure that this is the least preferred path, which by definition will remove them from the routing table
- process 300 configures the logical layer 3 interface configuration.
- process 300 configures the logical layer 3 interface by changing the routing metrics for this interface and that another interface is preferred over the other. For example and in one embodiment, process 300 can use the interface configuration command "ip OSPF cost" to change the routing metrics of OSPF routes.
- process 300 verifies that this configuration leaves the network element in a stable state. In one embodiment, process 300 verifies the logical interface configuration by performing a routing table check that would verify that paths are or are not programmed as intended. If the routes are not properly programmed, process 300 re-programs these routes and re-checks that re-programmed routes.
- Process 300 configures the layer 2 virtual interface configuration to move the layer 2 data processing away from the network element at block 306.
- process 300 increases the cost of layer 2 paths that includes this network element using spanning tree protocol.
- process 300 increases the spanning tree path costs that involve the network element. For example and in one embodiment, if the network element 104D is to be updated, network element 104D includes layer 2 paths to network elements 104A-C and 106A-E. For each of these routes, there are alternate layer 2 paths between network elements 104A-C and 106A-E that do not go through network element 104D.
- process 300 increases the cost of the paths in the spanning tree such that network elements 104A-C and 106A-E would chose paths that do not include network element 104D.
- the other network elements 104A-C and 106A-E would not switch data to network element 104D if there are alternate paths available to switch this data.
- the layer 2 switching data processing load can be effectively zero as the other network elements 104A-C and 106A-E will select alternate paths that do not include network element 104D.
- process 300 As with the layer 3 routing configuration described above, the network element configured by process 300 is not taken out of the network. Instead, process 300 has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in the layer 2 switching decision making for the other network elements.
- process 300 configures a multi-layer link aggregation group (MLAG) policy for this network element, such that this network element is removed from the operation of the MLAG.
- process 300 signals to the network elements configured in the MLAG to move the MLAG data processing from the network element to be updated to the other network elements in the MLAG.
- process 300 shuts down the MLAG peer link(s) between the network element to be updated and the other network elements in the MLAG.
- process 300 shuts down the link between the network element to be updated and other network element(s) and/or devices that are downstream from this network element.
- process 300 By shutting down these link(s), process 300 removes an MLAG path through the network element to be updated and effectively diverting MLAG data processing to the other network elements involved in the MLAG. [0041] In one embodiment, after process 300 configures the layer 2 switching configuration (e.g., spanning tree and/or MLAG) as described above, process 300 verifies that this
- process 300 verifies the layer 2 switching configuration by determining if the MLAG, spanning tree, and/or any other type of layer 2 states have converged. For example and in one embodiment, process 300 can invoke commands to show the state of the MLAG and/or spanning tree (e.g., show mlag detail, show spanning-tree and/or any other type of L2 information command). If one or more of the these layer 2 states have converged, then the network element is in a stable state. If these states have not converged, process 300 can wait for a time period (seconds, minutes, or another time period) and determine if the layer 2 network states have converged. At block 308, process 300 configures the physical interfaces of the network element to be updated for maintenance mode. In one embodiment, process 300 shuts down the physical interface(s) of the network element.
- process 300 shuts down the physical interface(s) of the network element.
- FIG. 4 is a block diagram of one embodiment of network layer stack 400.
- the network layer stack 400 includes four layers: a physical layer (layer 1) 402, a virtual layer (layer 2), a logical layer (layer 3) 406, and routing policy (layer 3) 408.
- the physical layer 402 is the basic networking hardware transmission layer.
- the physical layer is an Ethernet physical layer.
- the virtual layer (layer 2) 404 is layer above the physical layer 402 and is used to support the layer 2 services (e.g., switching, virtual local area network (VLAN), MLAG, and/or other layer 2 services).
- the Media Access Control (MAC) addresses and the Ethernet protocol are defined in the virtual layer 402.
- the logical layer (layer 3) 406 is used to support layer 3 services (e.g., Internet Protocol (IP) addressing, User Datagram Protocol (UDP), Transmission Control Protocol (TCP), routing based on IP addresses, Equal-cost multi- path routing (ECMP), multicast services, firewall or other types of security services, and/or other layer 3 services).
- IP Internet Protocol
- UDP User Datagram Protocol
- TCP Transmission Control Protocol
- ECMP Equal-cost multi- path routing
- multicast services e.g., firewall or other types of security services, and/or other layer 3 services.
- routing policy (layer 3) 408 is built upon the logical layer 406.
- the routing policy 408 includes the one or more different routing protocols that are configured on the network element to be updated.
- the network element to be updated may be running the BGP and/or OSPF routing protocols, which are routing protocols that use the logical layer interfaces (e.g., IP addresses) configured on this network element.
- process 200 restores the configuration, so as to re-insert this network element gracefully into the data processing load of the network. For example and in one embodiment, if the network element is configured for BGP routing or as part of an MLAG, process 200 reconfigures the network element to be in the same configuration this network element had before the network element was updated.
- Figure 5 is a flow diagram of one embodiment of a process 500 to restore the data processing of a network element.
- process 500 is performed by a restore module, such as the restore module 612 as described in Figure 6 below.
- process 500 restores the network element configuration in the opposite order that was used in Figure 3. In Figure 5, process 500 begins by restoring and verifying the layer 1 configuration.
- process 500 restores the layer 1 configuration by applying the layer 1 configuration that this network element had before the network element was put into maintenance mode.
- process 500 verifies the restored layer 1 configuration is valid by looking at the physical layer of the interface and determining if this interface is up.
- process 500 can also check if there are any error conditions, such as link fault signaling errors. For example and in one embodiment, process 500 can use a command such as show interface ⁇ interface name> transceiver detail and parse the output for any errors.
- process 500 restores and verifies the layer 2 configuration.
- process 500 restores the layer 2 configuration by applying the layer 2 configuration that this network element had before the network element was put into maintenance mode. For example and in one embodiment, if the network element was part of an MLAG, process 500 applies the configuration that configures the network element for the MLAG. In addition, process 500 would bring up a MLAG peer link with the other network elements that are involved in the MLAG. In another example and embodiment, process 500 restores the spanning tree costs for paths that involve the network element being restored. In addition, further verifies that the restored layer 2 configuration.
- process 500 can verify the layer 3 configuration by comparing the MLAG and/or other layer 2 configurations before and after the network element was put into maintenance mode. If there are no (or no substantial) differences between the MLAG and/or other layer 2 services, process 500 indicates that the layer 2 configuration is verified. If not, process 500 indicates that the layer 2 configuration is not verified.
- Process 500 restores and verifies the layer 3 configuration at block 506.
- process 500 restores the layer 3 configuration by applying the layer 3 configuration that this network element had before the network element was put into maintenance mode.
- process 500 verifies the restored layer 3 configuration is valid by comparing the routing table before and after the network element was put into maintenance mode. If there are no (or no substantial) differences between the routing tables, process 500 indicates that the layer 2 configuration is verified. If not, process 500 indicates that the layer 2 configuration is not verified.
- process 500 restores and verifies the routing policy (layer 3) configuration of the network element. In one embodiment, process 500 restores the routing policy configuration by applying the routing policy configuration that this network element had before the network element was put into maintenance mode.
- process 500 restores the route costs back to previous values.
- the network element will advertised the restored route costs to other network elements using a routing policy configured on the network element (e.g., border gateway protocol (BGP), open shortest path first (OSPF), or some other routing policy).
- BGP border gateway protocol
- OSPF open shortest path first
- FIG. 6 is a block diagram of one embodiment of a network update module 116 that updates a network element by utilizing redundant network capacity.
- the network update module includes a receive information module 602, redundant paths module 604, maintenance mode module 606, maintenance mode check module 608, update module 610, and restore module 612.
- the receive information module 602 receives the information as described in Figure 3, block 302 above.
- the redundant paths module 604 determines if there are redundant paths for this network element as described in Figure 3, block 304 above.
- the maintenance mode module 606 puts the network element into maintenance mode as described in Figure 3, block 306 above.
- the maintenance mode check module 608 checks that the network element is in maintenance mode as described in Figure 3, block 308 above.
- the update module 610 performs the software update as described in Figure 3, block 310 above.
- the restore module 612 restores the network element from maintenance mode as described in Figure 3, block 312 above.
- FIG. 7 is a block diagram of one embodiment of a maintenance mode module 606 that puts a network element into maintenance mode.
- the maintenance mode module 606 includes configure routing module 702, configure layer 3 module 704, configure layer 2 module 706, and configure layer 1 module 708.
- the configure routing module 702 configures the routing policy as described in Figure 3, block 302 above.
- the configure layer 3 module 704 configures the layer 3 configuration as described in Figure 3, block 304 above.
- the configure layer 2 module 706 configures the layer 2 configuration as described in Figure 3, block 306 above.
- the configure layer 1 module 708 configures the layer 1
- Figure 8 is a block diagram of one embodiment of a restore module 612 that restores the data processing of a network element.
- the restore module 612 includes restore routing module 702, restore layer 3 module 704, restore layer 2 module 706, and restore layer 1 module 708.
- the restore routing module 702 restores the routing policy as described in Figure 5, block 502 above.
- the restore layer 3 module 704 restores the layer 3 configuration as described in Figure 5, block 504 above.
- the restore layer 2 module 706 restores the layer 2 configuration as described in Figure 5, block 506 above.
- the restore layer 1 module 708 restores the layer 1 configuration as described in Figure 5, block 508 above.
- Figure 9 shows one example of a data processing system 900, which may be used with one embodiment of the present invention.
- the system 900 may be implemented including a controller 102 as shown in Figure 1.
- Figure 9 illustrates various components of a computer system, it is not intended to represent any particular architecture or manner of interconnecting the components as such details are not germane to the present invention. It will also be appreciated that network computers and other data processing systems or other consumer electronic devices, which have fewer components or perhaps more components, may also be used with the present invention.
- the computer system 900 which is a form of a data processing system, includes a bus 903 which is coupled to a microprocessor(s) 905 and a ROM (Read Only Memory) 907 and volatile RAM 909 and a non- volatile memory 911.
- the microprocessor 905 may retrieve the instructions from the memories 907, 909, 911 and execute the instructions to perform operations described above.
- the bus 903 interconnects these various components together and also interconnects these components 905, 907, 909, and 911 to a display controller and display device 917 and to peripheral devices such as input/output (I/O) devices which may be mice, keyboards, modems, network interfaces, printers and other devices which are well known in the art.
- I/O input/output
- the system 900 includes a plurality of network interfaces of the same or different type (e.g., Ethernet copper interface, Ethernet fiber interfaces, wireless, and/or other types of network interfaces).
- the system 900 can include a forwarding engine to forward network date received on one interface out another interface.
- the input/output devices 915 are coupled to the system through input/output controllers 913.
- the volatile RAM (Random Access Memory) 909 is typically implemented as dynamic RAM (DRAM), which requires power continually in order to refresh or maintain the data in the memory.
- DRAM dynamic RAM
- the mass storage 911 is typically a magnetic hard drive or a magnetic optical drive or an optical drive or a DVD ROM/RAM or a flash memory or other types of memory systems, which maintains data (e.g. large amounts of data) even after power is removed from the system.
- the mass storage 911 will also be a random access memory although this is not required.
- FIG. 9 shows that the mass storage 911 is a local device coupled directly to the rest of the components in the data processing system
- the present invention may utilize a non-volatile memory which is remote from the system, such as a network storage device which is coupled to the data processing system through a network interface such as a modem, an Ethernet interface or a wireless network.
- the bus 903 may include one or more buses connected to each other through various bridges, controllers and/or adapters as is well known in the art.
- a “machine” may be a machine that converts intermediate form (or “abstract") instructions into processor specific instructions (e.g., an abstract execution environment such as a "process virtual machine” (e.g., a Java Virtual Machine), an interpreter, a Common Language Runtime, a high-level language virtual machine, etc.), and/or, electronic circuitry disposed on a semiconductor chip (e.g., "logic circuitry” implemented with transistors) designed to execute instructions such as a general- purpose processor and/or a special-purpose processor. Processes taught by the discussion above may also be performed by (in the alternative to a machine or in combination with a machine) electronic circuitry designed to perform the processes (or a portion thereof) without the execution of program code.
- processor specific instructions e.g., an abstract execution environment such as a "process virtual machine” (e.g., a Java Virtual Machine), an interpreter, a Common Language Runtime, a high-level language virtual machine, etc.)
- the present invention also relates to an apparatus for performing the operations described herein.
- This apparatus may be specially constructed for the required purpose, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer.
- a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), RAMs, EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
- a machine readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer).
- a machine readable medium includes read only memory ("ROM”); random access memory (“RAM”);
- magnetic disk storage media optical storage media; flash memory devices; etc.
- An article of manufacture may be used to store program code.
- An article of manufacture that stores program code may be embodied as, but is not limited to, one or more memories (e.g., one or more flash memories, random access memories (static, dynamic or other)), optical disks, CD-ROMs, DVD ROMs, EPROMs, EEPROMs, magnetic or optical cards or other type of machine-readable media suitable for storing electronic instructions.
- Program code may also be downloaded from a remote computer (e.g., a server) to a requesting computer (e.g., a client) by way of data signals embodied in a propagation medium (e.g., via a
Landscapes
- Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Security & Cryptography (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
A method and apparatus of a device that performs a hitless update a boot image of a network element. In this embodiment, the device identifies the network element to update and determines if the network element has redundant paths. If this network element has redundant paths, the device configures the network element to drain data processing of the network element. In addition, the device updates the network element to a new boot image when the data processing of the network element is drained.
Description
SYSTEM AND METHOD OF UPDATING A NETWORK ELEMENT
[0001] This claims the benefit of priority of prior, co-pending U.S. Patent Application No. 15/057,880, filed March 1, 2016, which claims the benefit of priority of prior, co-pending provisional applications U.S. Provisional Patent Application No. 62/300,710 filed February 26, 2016 and U.S. Provisional Patent Application No. 62,183,130 filed June 22, 2015 and U.S. Provisional Patent Application No. 62/150,787 filed April 21, 2015, of which are hereby incorporated by reference.
FIELD OF INVENTION
[0002] This invention relates generally to data networking, and more particularly, to updating network element boot image for a network element in a network without dropping data in the network.
BACKGROUND OF THE INVENTION
[0003] For a network with multiple network elements, each network element runs a particular version of software along with a specific configuration for that software. Periodically, the software and/or the configuration of one or more of the network elements are updated.
Sometimes, the software and/or configuration updates leads to poor performance or disruption because of problems with the software configuration, incompatibilities between different software configurations on different network elements, loss of connectivity due to incorrect Virtual Local Area Network (VLAN) membership change on a port, incorrect Access Control List (ACL) entry, incorrect route metric change, or other types of problems.
[0004] In one example, the network element is upgraded to a new software image to add new features and/or fix problems with the current software image. Alternatively, the network image can be downgraded to a previous version of the software image. To change the software image, the network element loads the new software image, marks this new software image as the current software image, and reboots the network element. As the network element boots up, the new software image is loaded and programs the control and data planes of the network element.
[0005] A problem with this way of upgrading or downgrading the software image is that, by rebooting the network element, the network element is inoperable for a relatively long time. For example, rebooting the network element can cause the network element to be off-line for upwards of five minutes. During this off-line time, the data processing functions of the network
element are inoperable. This can lead to an interruption of data networking services that rely on the network element.
SUMMARY OF THE DESCRIPTION
[0006] A method and apparatus of a device that performs a hitless update a boot image of a network element. In this embodiment, the device identifies the network element to update and determines if the network element has redundant paths. If this network element has redundant paths, the device configures the network element to drain data processing of the network element. In addition, the device updates the network element to a new boot image when the data processing of the network element is drained.
[0007] Other methods and apparatuses are also described.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] The present invention is illustrated by way of example and not limitation in the figures of the accompanying drawings in which like references indicate similar elements.
[0009] Figure 1 is a block diagram of one embodiment of a network with a controller that updates the boot images of the network elements in the network.
[0010] Figure 2 is a flow diagram of one embodiment of a process to update a network element by utilizing redundant network capacity.
[0011] Figure 3 is a flow diagram of one embodiment of a process to drain data processing of a network element.
[0012] Figure 4 is a block diagram of one embodiment of network layer hierarchy.
[0013] Figure 5 is a flow diagram of one embodiment of a process to restore the data processing of a network element.
[0014] Figure 6 is a block diagram of one embodiment of a network update module that updates a network element by utilizing redundant network capacity.
[0015] Figure 7 is a block diagram of one embodiment of a drain data module that drains the data processing of a network element.
[0016] Figure 8 is a block diagram of one embodiment of a restore data module that restores the data processing of a network element.
[0017] Figure 9 illustrates one example of a typical computer system, which may be used in conjunction with the embodiments described herein.
DETAILED DESCRIPTION
[0018] A method and apparatus of a device that performs a hitless update a boot image of a network element. In the following description, numerous specific details are set forth to provide thorough explanation of embodiments of the present invention. It will be apparent, however, to one skilled in the art, that embodiments of the present invention may be practiced without these
specific details. In other instances, well-known components, structures, and techniques have not been shown in detail in order not to obscure the understanding of this description.
[0019] Reference in the specification to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the invention. The appearances of the phrase "in one embodiment" in various places in the specification do not necessarily all refer to the same embodiment.
[0020] In the following description and claims, the terms "coupled" and "connected," along with their derivatives, may be used. It should be understood that these terms are not intended as synonyms for each other. "Coupled" is used to indicate that two or more elements, which may or may not be in direct physical or electrical contact with each other, co-operate or interact with each other. "Connected" is used to indicate the establishment of communication between two or more elements that are coupled with each other.
[0021] The processes depicted in the figures that follow, are performed by processing logic that comprises hardware (e.g., circuitry, dedicated logic, etc.), software (such as is run on a general-purpose computer system or a dedicated machine), or a combination of both. Although the processes are described below in terms of some sequential operations, it should be appreciated that some of the operations described may be performed in different order.
Moreover, some operations may be performed in parallel rather than sequentially.
[0022] The terms "server," "client," and "device" are intended to refer generally to data processing systems rather than specifically to a particular form factor for the server, client, and/or device.
[0023] A method and apparatus of a device that performs a hitless update a boot image of a network element. In one embodiment, a device updates the boot image by putting the network element into a maintenance mode before the network element is updated. In this embodiment, the network element is part of redundant paths between one or more pairs of endpoints in the network. A maintenance mode for a network element is a configuration for the network element where one or more the paths that include the network element are disfavored paths such that other network elements will select other paths that do not include this network element that is in maintenance mode. Furthermore, the network element in maintenance mode is not taken out of the network. Instead, the device has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in for different forwarding decision making for the other network elements (e.g., routing, switching, and/or other forwarding decisions).
[0024] In one embodiment, the device puts the network element in maintenance mode by configuring each layer in a network stack for maintenance mode. For example and in one embodiment, the device configures the routing policy (layer 3) for maintenance mode. In this example, the device increases the cost of routes from that network element to other network elements and/or devices in the network. In addition, the device verifies the maintenance mode configuration for the routing policy. Furthermore, the device configures and verifies the virtual interface layer (layer 3) for maintenance mode. The device additionally configures and verifies the logical interface layer (layer 2) for maintenance mode by configuring the switch forwarding costs for paths that include this network element to be higher such that other network elements would not select these paths. In one example, the device increases spanning tree costs of paths that include this network element. The device further configures and verifies the physical interface layer (layer 1) for maintenance mode.
[0025] With the network element in maintenance mode, the data processing of transported network data being performed by this network element is essentially zero, because this network data is being forwarded through other network elements. With no or very little data being transported through the network element to be updated, this network element can be updated with a new boot image with little or no impact on this transported data, because any interruption in the network element would not impact the transported data (e.g., a reboot of the network element during the update process would not interrupt the transported data processing).
[0026] In one embodiment, once the network element is updated, the device rolls back the maintenance mode layer by layer for the network elements, verifying the changes made at each layer, and in reverse order. For example and in one embodiment, the device restores and verifies the physical layer (layer 1) to the configuration this network element had before the software update. With the physical layer configuration restored, the device restores and verifies the virtual layer (layer 2) to the configuration this network element had before the software update.
Similarly, the device restores and verifies the logical layer (layer 3), followed by the routing policy, to the configuration this network element had before the software update.
[0027] Figure 1 is a block diagram of one embodiment of a network 100 with a controller 102 that updates the boot images of the network elements in the network. In Figure 1, the network 100 includes spine network elements 104A-D that are coupled to each of the leaf network elements 106A-E. The leaf network elements 106A-E are further coupled to server racks 108A-E, respectively. In one embodiment, a spine network element 104A-D is a network element that interconnects the leaf network elements 106A-E. In this embodiment, each of the spine network elements 104A-D is coupled to each of the leaf network elements 106A-E.
Furthermore, in this embodiment, each of the spine network elements 104A-D is coupled with
each other. While in one embodiment, the network elements 104A-D and 106A-E are illustrated in a spine and leaf topology, in alternate embodiments, the network elements 104A-D and 106A- E can be in a different topology. In addition, and in one embodiment, the network elements 104A-E are further coupled to a controller 102 and a database 114. In one embodiment, each of the network elements 104A-D and/or 106A-E can be a router, switch, bridge, gateway, load balancer, firewall, network security device, server, or any other type of device that can receive and process data from a network. In one embodiment, the network elements 104A-D and 106A-E can be the same or different network elements in terms of manufacturer, type, configuration, or role. For example and in one embodiment, network elements 104A-D may be routers and network elements 106A-E may be switches. As another example and embodiment, network elements 104A-D may be high capacity switches with relatively few 10 gigabit (Gb) or 40Gb ports and network elements 106A-E may be lower capacity switches with a large number of medium capacity port (e.g., 1Gb ports) and high capacity ports to couple to the spine network elements 104A-D.
[0028] In one embodiment, the controller 102 is a device that is used to support network- wide services for a network that includes multiple network elements. A network- wide service is a service that is used by two or more network elements. For example, the network- wide service can be a service that gathers known Media Access Control (MAC) addresses and assembles these addresses into a global MAC address database. This global MAC address database can be used to support a VxLAN controller service. In one embodiment, database 114 is an organized collection of data. As another example and embodiment, the ability to rollback or roll forward network elements to different configuration points is a network-wide service.
[0029] In one embodiment, the configuration manager 112 can initiate software upgrades that update the boot images for some or all of the network elements 104A-D and 106A-E. In this embodiment, an update to a boot image of a network element can be a downgrade to boot image previously used with a corresponding lower version software, an upgrade to boot image with a corresponding higher version software, or replacement boot image that has the same version software as currently running on that network element. This update to the boot image can be a hitless update or hit full update. In one embodiment, a hitless update to a network element, or group of network elements, is an update to the boot image that does not cause a loss in data processing. In this embodiment, a hitless update is performed and there is no loss or dropping of data packets by the network elements that are being updated. In contrast, a hit full update is an update of a boot image to a network element, or group of network elements, where there can be a loss or dropping of data packets by the network elements that are being updated. In one
embodiment, each of the network elements 104A-D and/or 106A-E includes a network update module 116 that performs the software update.
[0030] As above, a boot image update can be performed as a hitless or a hit full update. In one embodiment, a hitless update can be performed by draining the data processing needs on a network element that has redundant links with other network elements. Figure 2 is a flow diagram of one embodiment of a process 200 to update a network element by utilizing redundant network capacity and putting that network element into maintenance mode. In one embodiment, process 200 is performed by a network update module, such as the network update module 116 described in Figure 1 above. In Figure 2, process 200 begins by receiving the network element information for the network element that is to be updated with a new boot image at block 202. In one embodiment, the network element information received includes information whether the network element has redundant links such that data processing can be diverted to other network elements while the network element boot images being updated. At block 204, process 200 determines if the network element has redundant paths. In one embodiment, a network element has redundant paths if there are other network elements that can perform the data processing the network element is currently performing. For example and in one embodiment, the network elements may be part of an equal cost multipath group or a link aggregate group that allows for multiple network elements to appear as one network element. As illustrated in Figure 1 above, network elements 104A-D or 106A-E have redundant paths that allow others of these network elements to take over the data processing for one of the network elements. For example and in one embodiment, if network element 104A went off-line (e.g., to update a boot image), the other network elements 104B-D can take over the data processing for network element 104A. If the network element does not have redundant paths, process 200 performs a hitless update on the network element at block 206. In one embodiment, process 200 performs an accelerated software upgrade on the network element as described in in U.S. Patent Application No. 14/449,030, entitled "System and Method for Accelerated Software Upgrades", filed on July 31, 2014.
[0031] If the network element does have redundant paths, at block 208, process 200 configures the network element to be in maintenance mode. In one embodiment, process 200 puts the network element in maintenance mode by draining the data processing from the network element. In this embodiment, process 200 diverts the flow of data to the other network elements that have redundant paths for path(s) this network element participates in. In one embodiment, process 200 drains the data processing from the network element by increasing the cost for forwarding the data in this network element. In this embodiment, the network element will advertise its forwarding costs such that the flow of data will be diverted to the other network elements. For example and in one embodiment, network element 104A can increase the BGP
cost to a large number (e. g., a large value or infinity) and advertise this cost, such that data will be diverted to network elements 104B-D. By increasing this forwarding cost, where the network element will advertise this cost, no data will eventually be forwarded to this network element. When no data is being forwarded to this network element, process 200 can update the boot image of this network element without any data loss. Putting the network element in maintenance mode is further described in Figure 3 below. At block 210, process 200 determines if the maintenance mode configuration was successful. In one embodiment, process 200 checks the maintenance mode configuration at each layer of configuration. If one of those layer configurations fails, process 200 reverts the previous layer configuration(s) to an original configuration for those layers and indicates a failure. If the maintenance mode configuration fails, execution proceeds to block 208 above. If the maintenance mode configuration is successful, process 200 performs the boot image update at block 210. In one embodiment, process 200 can perform the boot image update that involves a reboot of the network element. In this embodiment, even though the network element is rebooted during the boot image update process, this update is hitless because no data is lost during the rebooting. Process 200 restores the data processing configuration at block 214. In one embodiment, process 200 restores the data processing configuration by restoring the configuration for each layer of the network protocols for this network element. Restoring the data processing configuration is further described in Figure 5 below.
[0032] As described above, process 200 drains the data processing for the network element by configuring the network element by putting the network element into maintenance mode. Figure 3 is a flow diagram of one embodiment of a process 300 to put the network element into maintenance mode. In one embodiment, process 300 is performed by a drain data processing module, such as the drain data processing module 606 as described in Figure 6 below. In Figure 3, process 300 begins by configuring the routing policy (layer 3) of network element, so as to drain the data processing of the network element at block 302. In one embodiment, process 300 configures the routing policy (layer 3) of the network element so as to drain the layer 3 routing data processing from the network element. In one embodiment, process 300 configures the layer 3 routing policy so as to discourage data being routed through this network element. In this embodiment, the network element advertises the updated layer routing policy, such that other network elements would send the data to be routed to a different path that does not include this network element. In one embodiment, process 300 can configure the layer routing costs to be higher, so that these routes can be advertised as having the high cost. In this embodiment, other network element receiving the high cost advertised routes would select lower cost routes for the data. For example and in one embodiment, process 300 can configure the network element that routes for BGP, OSPF, or other types of routing protocol have a very high cost. By increasing the
route costs, the network element would advertise these high cost routes to other network elements via one or more of the routing protocols (e.g., BGP, OSPF, and/or another type of routing protocol) and the other network elements receiving these route advertisements would select other paths that do not include this network element.
[0033] As another example and embodiment and referring to Figure 1 above, if network element 104D is to be updated, network element 104D includes routes to network elements 104A-C and 106A-E. For each of these routes, there are alternate layer 3 routing paths between network elements 104A-C and 106A-E that do not go through network element 104D. In this example, process 300 increases cost of the routes to the other network elements and network element 104D advertises these high cost routes to the other network elements (106A-C and 104A-E). The other network elements 106A-C and 104A-E use the high cost of the advertised routes from network element 104D to make routing decisions. In this example, the other network elements 104A-C and 106A-E would not route data to network element 104D if there are alternate paths available to route the data. Thus, in this example, the layer 3 routing data processing load can be effectively zero as the other network elements 104A-C and 106A-E will select alternate routes that do not include network element 104D.
[0034] In this embodiment, the network element configured by process 300 is not taken out of the network. Instead, process 300 has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in the layer 3 routing decision making for the other network elements.
[0035] In one embodiment, after process 300 configures the layer 3 routing configuration as described above, process 300 verifies that this configuration leaves the network element in a stable state. In one embodiment, process 300 verifies the routing configuration by that the routing metrics have been increased to a sufficient level for some or all of the routing protocols. As described above, this policy of increasing the routing metrics is to drain the layer 3 data, so that the routing metrics (or admin distance in the case of multiple protocol deployment) are increased to ensure that this is the least preferred path, which by definition will remove them from the routing table
[0036] At block 304, process 300 configures the logical layer 3 interface configuration. In one embodiment, process 300 configures the logical layer 3 interface by changing the routing metrics for this interface and that another interface is preferred over the other. For example and in one embodiment, process 300 can use the interface configuration command "ip OSPF cost" to change the routing metrics of OSPF routes.
[0037] In one embodiment, after process 300 configures the layer 3 routing configuration as described above, process 300 verifies that this configuration leaves the network element in a
stable state. In one embodiment, process 300 verifies the logical interface configuration by performing a routing table check that would verify that paths are or are not programmed as intended. If the routes are not properly programmed, process 300 re-programs these routes and re-checks that re-programmed routes.
[0038] Process 300 configures the layer 2 virtual interface configuration to move the layer 2 data processing away from the network element at block 306. In one embodiment, process 300 increases the cost of layer 2 paths that includes this network element using spanning tree protocol. In this embodiment, process 300 increases the spanning tree path costs that involve the network element. For example and in one embodiment, if the network element 104D is to be updated, network element 104D includes layer 2 paths to network elements 104A-C and 106A-E. For each of these routes, there are alternate layer 2 paths between network elements 104A-C and 106A-E that do not go through network element 104D. In this example, process 300 increases the cost of the paths in the spanning tree such that network elements 104A-C and 106A-E would chose paths that do not include network element 104D. In this example, the other network elements 104A-C and 106A-E would not switch data to network element 104D if there are alternate paths available to switch this data. Thus, in this example, the layer 2 switching data processing load can be effectively zero as the other network elements 104A-C and 106A-E will select alternate paths that do not include network element 104D.
[0039] As with the layer 3 routing configuration described above, the network element configured by process 300 is not taken out of the network. Instead, process 300 has configured the network element such that this network element is still an active device in the network, but this network element is a disfavored destination in the layer 2 switching decision making for the other network elements.
[0040] In another embodiment, process 300 configures a multi-layer link aggregation group (MLAG) policy for this network element, such that this network element is removed from the operation of the MLAG. In this embodiment, process 300 signals to the network elements configured in the MLAG to move the MLAG data processing from the network element to be updated to the other network elements in the MLAG. In addition, process 300 shuts down the MLAG peer link(s) between the network element to be updated and the other network elements in the MLAG. Furthermore, process 300 shuts down the link between the network element to be updated and other network element(s) and/or devices that are downstream from this network element. By shutting down these link(s), process 300 removes an MLAG path through the network element to be updated and effectively diverting MLAG data processing to the other network elements involved in the MLAG.
[0041] In one embodiment, after process 300 configures the layer 2 switching configuration (e.g., spanning tree and/or MLAG) as described above, process 300 verifies that this
configuration leaves the network element in a stable state. In one embodiment, process 300 verifies the layer 2 switching configuration by determining if the MLAG, spanning tree, and/or any other type of layer 2 states have converged. For example and in one embodiment, process 300 can invoke commands to show the state of the MLAG and/or spanning tree (e.g., show mlag detail, show spanning-tree and/or any other type of L2 information command). If one or more of the these layer 2 states have converged, then the network element is in a stable state. If these states have not converged, process 300 can wait for a time period (seconds, minutes, or another time period) and determine if the layer 2 network states have converged. At block 308, process 300 configures the physical interfaces of the network element to be updated for maintenance mode. In one embodiment, process 300 shuts down the physical interface(s) of the network element.
[0042] As described above, process 300 successively configures and verifies each layer of network protocols down to the physical layer. Furthermore, as will be described further below, a process (e.g., process 500 of Figure 5) restores the configuration of the network element after that network element has been updated by successively configuring and verifying each layer up the network protocol stack. Figure 4 is a block diagram of one embodiment of network layer stack 400. In Figure 4, the network layer stack 400 includes four layers: a physical layer (layer 1) 402, a virtual layer (layer 2), a logical layer (layer 3) 406, and routing policy (layer 3) 408. In one embodiment, the physical layer 402 is the basic networking hardware transmission layer. In one embodiment, the physical layer is an Ethernet physical layer.
[0043] In one embodiment, the virtual layer (layer 2) 404 is layer above the physical layer 402 and is used to support the layer 2 services (e.g., switching, virtual local area network (VLAN), MLAG, and/or other layer 2 services). In one embodiment, the Media Access Control (MAC) addresses and the Ethernet protocol are defined in the virtual layer 402. Above the virtual layer 404 is the logical layer (layer 3) 406. In one embodiment, the logical layer (layer 3) 406 is used to support layer 3 services (e.g., Internet Protocol (IP) addressing, User Datagram Protocol (UDP), Transmission Control Protocol (TCP), routing based on IP addresses, Equal-cost multi- path routing (ECMP), multicast services, firewall or other types of security services, and/or other layer 3 services). In particular, and in one embodiment, routing policy (layer 3) 408 is built upon the logical layer 406. In this embodiment, the routing policy 408 includes the one or more different routing protocols that are configured on the network element to be updated. For example and in one embodiment, the network element to be updated may be running the BGP
and/or OSPF routing protocols, which are routing protocols that use the logical layer interfaces (e.g., IP addresses) configured on this network element.
[0044] In one embodiment, after the network element is updated, process 200 restores the configuration, so as to re-insert this network element gracefully into the data processing load of the network. For example and in one embodiment, if the network element is configured for BGP routing or as part of an MLAG, process 200 reconfigures the network element to be in the same configuration this network element had before the network element was updated. Figure 5 is a flow diagram of one embodiment of a process 500 to restore the data processing of a network element. In one embodiment, process 500 is performed by a restore module, such as the restore module 612 as described in Figure 6 below. In one embodiment, process 500 restores the network element configuration in the opposite order that was used in Figure 3. In Figure 5, process 500 begins by restoring and verifying the layer 1 configuration. In one embodiment, process 500 restores the layer 1 configuration by applying the layer 1 configuration that this network element had before the network element was put into maintenance mode. In addition and in this embodiment, process 500 verifies the restored layer 1 configuration is valid by looking at the physical layer of the interface and determining if this interface is up. In addition, process 500 can also check if there are any error conditions, such as link fault signaling errors. For example and in one embodiment, process 500 can use a command such as show interface <interface name> transceiver detail and parse the output for any errors.
[0045] At block 504, process 500 restores and verifies the layer 2 configuration. In one embodiment, process 500 restores the layer 2 configuration by applying the layer 2 configuration that this network element had before the network element was put into maintenance mode. For example and in one embodiment, if the network element was part of an MLAG, process 500 applies the configuration that configures the network element for the MLAG. In addition, process 500 would bring up a MLAG peer link with the other network elements that are involved in the MLAG. In another example and embodiment, process 500 restores the spanning tree costs for paths that involve the network element being restored. In addition, further verifies that the restored layer 2 configuration. In one embodiment, process 500 can verify the layer 3 configuration by comparing the MLAG and/or other layer 2 configurations before and after the network element was put into maintenance mode. If there are no (or no substantial) differences between the MLAG and/or other layer 2 services, process 500 indicates that the layer 2 configuration is verified. If not, process 500 indicates that the layer 2 configuration is not verified.
[0046] Process 500 restores and verifies the layer 3 configuration at block 506. In one embodiment, process 500 restores the layer 3 configuration by applying the layer 3 configuration
that this network element had before the network element was put into maintenance mode. In addition and in this embodiment, process 500 verifies the restored layer 3 configuration is valid by comparing the routing table before and after the network element was put into maintenance mode. If there are no (or no substantial) differences between the routing tables, process 500 indicates that the layer 2 configuration is verified. If not, process 500 indicates that the layer 2 configuration is not verified. At block 508, process 500 restores and verifies the routing policy (layer 3) configuration of the network element. In one embodiment, process 500 restores the routing policy configuration by applying the routing policy configuration that this network element had before the network element was put into maintenance mode. For example and in one embodiment, if the route costs were increased as the network element was put into maintenance mode, process 500 restores the route costs back to previous values. In this example, the network element will advertised the restored route costs to other network elements using a routing policy configured on the network element (e.g., border gateway protocol (BGP), open shortest path first (OSPF), or some other routing policy).
[0047] Figure 6 is a block diagram of one embodiment of a network update module 116 that updates a network element by utilizing redundant network capacity. In one embodiment, the network update module includes a receive information module 602, redundant paths module 604, maintenance mode module 606, maintenance mode check module 608, update module 610, and restore module 612. In one embodiment, the receive information module 602 receives the information as described in Figure 3, block 302 above. The redundant paths module 604 determines if there are redundant paths for this network element as described in Figure 3, block 304 above. The maintenance mode module 606 puts the network element into maintenance mode as described in Figure 3, block 306 above. The maintenance mode check module 608 checks that the network element is in maintenance mode as described in Figure 3, block 308 above. The update module 610 performs the software update as described in Figure 3, block 310 above. The restore module 612 restores the network element from maintenance mode as described in Figure 3, block 312 above.
[0048] Figure 7 is a block diagram of one embodiment of a maintenance mode module 606 that puts a network element into maintenance mode. In one embodiment, the maintenance mode module 606 includes configure routing module 702, configure layer 3 module 704, configure layer 2 module 706, and configure layer 1 module 708. In one embodiment, the configure routing module 702 configures the routing policy as described in Figure 3, block 302 above. The configure layer 3 module 704 configures the layer 3 configuration as described in Figure 3, block 304 above. The configure layer 2 module 706 configures the layer 2 configuration as described in
Figure 3, block 306 above. The configure layer 1 module 708 configures the layer 1
configuration as described in Figure 3, block 308 above.
[0049] Figure 8 is a block diagram of one embodiment of a restore module 612 that restores the data processing of a network element. In one embodiment, the restore module 612 includes restore routing module 702, restore layer 3 module 704, restore layer 2 module 706, and restore layer 1 module 708. In one embodiment, the restore routing module 702 restores the routing policy as described in Figure 5, block 502 above. The restore layer 3 module 704 restores the layer 3 configuration as described in Figure 5, block 504 above. The restore layer 2 module 706 restores the layer 2 configuration as described in Figure 5, block 506 above. The restore layer 1 module 708 restores the layer 1 configuration as described in Figure 5, block 508 above.
[0050] Figure 9 shows one example of a data processing system 900, which may be used with one embodiment of the present invention. For example, the system 900 may be implemented including a controller 102 as shown in Figure 1. Note that while Figure 9 illustrates various components of a computer system, it is not intended to represent any particular architecture or manner of interconnecting the components as such details are not germane to the present invention. It will also be appreciated that network computers and other data processing systems or other consumer electronic devices, which have fewer components or perhaps more components, may also be used with the present invention.
[0051] As shown in Figure 9, the computer system 900, which is a form of a data processing system, includes a bus 903 which is coupled to a microprocessor(s) 905 and a ROM (Read Only Memory) 907 and volatile RAM 909 and a non- volatile memory 911. The microprocessor 905 may retrieve the instructions from the memories 907, 909, 911 and execute the instructions to perform operations described above. The bus 903 interconnects these various components together and also interconnects these components 905, 907, 909, and 911 to a display controller and display device 917 and to peripheral devices such as input/output (I/O) devices which may be mice, keyboards, modems, network interfaces, printers and other devices which are well known in the art. In one embodiment, the system 900 includes a plurality of network interfaces of the same or different type (e.g., Ethernet copper interface, Ethernet fiber interfaces, wireless, and/or other types of network interfaces). In this embodiment, the system 900 can include a forwarding engine to forward network date received on one interface out another interface.
[0052] Typically, the input/output devices 915 are coupled to the system through input/output controllers 913. The volatile RAM (Random Access Memory) 909 is typically implemented as dynamic RAM (DRAM), which requires power continually in order to refresh or maintain the data in the memory.
[0053] The mass storage 911 is typically a magnetic hard drive or a magnetic optical drive or an optical drive or a DVD ROM/RAM or a flash memory or other types of memory systems, which maintains data (e.g. large amounts of data) even after power is removed from the system. Typically, the mass storage 911 will also be a random access memory although this is not required. While Figure 9 shows that the mass storage 911 is a local device coupled directly to the rest of the components in the data processing system, it will be appreciated that the present invention may utilize a non-volatile memory which is remote from the system, such as a network storage device which is coupled to the data processing system through a network interface such as a modem, an Ethernet interface or a wireless network. The bus 903 may include one or more buses connected to each other through various bridges, controllers and/or adapters as is well known in the art.
[0054] Portions of what was described above may be implemented with logic circuitry such as a dedicated logic circuit or with a microcontroller or other form of processing core that executes program code instructions. Thus processes taught by the discussion above may be performed with program code such as machine-executable instructions that cause a machine that executes these instructions to perform certain functions. In this context, a "machine" may be a machine that converts intermediate form (or "abstract") instructions into processor specific instructions (e.g., an abstract execution environment such as a "process virtual machine" (e.g., a Java Virtual Machine), an interpreter, a Common Language Runtime, a high-level language virtual machine, etc.), and/or, electronic circuitry disposed on a semiconductor chip (e.g., "logic circuitry" implemented with transistors) designed to execute instructions such as a general- purpose processor and/or a special-purpose processor. Processes taught by the discussion above may also be performed by (in the alternative to a machine or in combination with a machine) electronic circuitry designed to perform the processes (or a portion thereof) without the execution of program code.
[0055] The present invention also relates to an apparatus for performing the operations described herein. This apparatus may be specially constructed for the required purpose, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), RAMs, EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, and each coupled to a computer system bus.
[0056] A machine readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine
readable medium includes read only memory ("ROM"); random access memory ("RAM");
magnetic disk storage media; optical storage media; flash memory devices; etc.
[0057] An article of manufacture may be used to store program code. An article of manufacture that stores program code may be embodied as, but is not limited to, one or more memories (e.g., one or more flash memories, random access memories (static, dynamic or other)), optical disks, CD-ROMs, DVD ROMs, EPROMs, EEPROMs, magnetic or optical cards or other type of machine-readable media suitable for storing electronic instructions. Program code may also be downloaded from a remote computer (e.g., a server) to a requesting computer (e.g., a client) by way of data signals embodied in a propagation medium (e.g., via a
communication link (e.g., a network connection)).
[0058] The preceding detailed descriptions are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the tools used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of operations leading to a desired result. The operations are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
[0059] It should be kept in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the above discussion, it is appreciated that throughout the description, discussions utilizing terms such as "receiving," "identifying," "determining," "updating," "failing," "signaling," "configuring," "increasing," or the like, refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
[0060] The processes and displays presented herein are not inherently related to any particular computer or other apparatus. Various general-purpose systems may be used with programs in accordance with the teachings herein, or it may prove convenient to construct a more specialized apparatus to perform the operations described. The required structure for a variety of these systems will be evident from the description below. In addition, the present invention is
not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the invention as described herein.
[0061] The foregoing discussion merely describes some exemplary embodiments of the present invention. One skilled in the art will readily recognize from such discussion, the accompanying drawings and the claims that various modifications can be made without departing from the spirit and scope of the invention.
Claims
1. A non-transitory machine-readable medium having executable instructions to cause one or more processing units to perform a method to perform a hitless update a boot image of a network element, the method comprising:
identifying the network element to update;
determining if the network element is involved in redundant paths; and
if that network element is involved in redundant paths,
configuring the network element to drain data processing of the network element, and
updating the network element to a new boot image when the data processing of the network element is drained.
2. The machine-readable medium of claim 1 , wherein the configuring the network element comprises:
increasing the forwarding cost of the data processing.
3. The machine-readable medium of claim 2, wherein the forwarding cost is a routing metric of a routing protocol.
4. The machine-readable medium of claim 3, wherein the routing protocol is selected from the group consisting of Border Gateway Protocol and Open Shortest Path First.
5. The machine-readable medium of claim 3, wherein the routing metric is selected from the group consisting of hop count, path length, bandwidth, load, path cost, delay, reliability, and communication cost.
6. The machine-readable medium of claim 2, wherein the forwarding cost is a switching metric of a switching protocol.
7. The machine-readable medium of claim 1, wherein a network element is drained on data processing if data that could be processed by the network element is processed by one or more other network elements coupled to this network element by the redundant links.
8. The machine-readable medium of claim 1, wherein the configuring the network element comprises:
successively configuring lower levels of networking protocol layers of forwarding costs at each layer, wherein each of the networking protocol layers is used to make forwarding decisions by the network element.
9. The machine-readable medium of claim 1, further comprising:
rebooting the network element.
10. The machine-readable medium of claim 1, further comprising:
restoring a configuration of the network element.
11. A method to perform a hitless update a boot image of a network element, the method comprising:
identifying the network element to update;
determining if the network element is involved in redundant paths; and
if that network element is involved in redundant paths,
configuring the network element to drain data processing of the network element, and
updating the network element to a new boot image when the data processing of the network element is drained.
12. The method of claim 11, wherein the configuring the network element comprises:
increasing the forwarding cost of the data processing.
13. The method of claim 12, wherein the forwarding cost is a routing metric of a routing protocol.
14. The method of claim 13, wherein the routing protocol is selected from the group consisting of Border Gateway Protocol and Open Shortest Path First.
15. The method of claim 11, wherein the routing metric is selected from the group consisting of hop count, path length, bandwidth, load, path cost, delay, reliability, and communication cost.
16. The method of claim 11, wherein the forwarding cost is a switching metric of a switching protocol.
17. The method of claim 11, wherein a network element is drained on data processing if data that could be processed by the network element is processed by one or more other network elements coupled to this network element by the redundant links.
18. The method of claim 11, wherein the configuring the network element comprises:
successively configuring lower levels of networking protocol layers of forwarding costs at each layer, wherein each of the networking protocol layers is used to make forwarding decisions by the network element.
19. A network element that performs a hitless update a boot image of the network element, the network method comprising:
a processor;
a memory coupled to the processor though a bus; and
a process executed from the memory by the processor causes the processor to identify the network element to update, determine if the network element is involved in redundant paths, and if that network element is involved in redundant paths, configure the network element to drain data processing of the network element, and update the network element to a new boot image when the data processing of the network element is drained.
20. The network element of claim 19, wherein the process further causes the processor to increase the forwarding cost of the data processing.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP16783664.2A EP3278215B1 (en) | 2015-04-21 | 2016-04-18 | System and method of updating a network element |
Applications Claiming Priority (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201562150787P | 2015-04-21 | 2015-04-21 | |
US62/150,787 | 2015-04-21 | ||
US201562183130P | 2015-06-22 | 2015-06-22 | |
US62/183,130 | 2015-06-22 | ||
US201662300710P | 2016-02-26 | 2016-02-26 | |
US62/300,710 | 2016-02-26 | ||
US15/057,880 US10545753B2 (en) | 2015-04-21 | 2016-03-01 | System and method of updating a network element |
US15/057,880 | 2016-03-01 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016172060A1 true WO2016172060A1 (en) | 2016-10-27 |
Family
ID=57144199
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2016/028171 WO2016172060A1 (en) | 2015-04-21 | 2016-04-18 | System and method of updating a network element |
Country Status (3)
Country | Link |
---|---|
US (3) | US10545753B2 (en) |
EP (1) | EP3278215B1 (en) |
WO (1) | WO2016172060A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106953798A (en) * | 2017-05-25 | 2017-07-14 | 深圳市伊特利网络科技有限公司 | Link selecting method and system based on access point |
US11303533B2 (en) | 2019-07-09 | 2022-04-12 | Cisco Technology, Inc. | Self-healing fabrics |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10545753B2 (en) * | 2015-04-21 | 2020-01-28 | Arista Networks, Inc. | System and method of updating a network element |
US10938653B2 (en) | 2015-04-21 | 2021-03-02 | Arista Networks, Inc. | System and method of updating a network |
WO2017019108A1 (en) | 2015-07-30 | 2017-02-02 | Hewlett Packard Enterprise Development Lp | Firmware update of an interconnect device |
US11182191B2 (en) | 2018-12-19 | 2021-11-23 | Vmware, Inc. | Nested host manager in a hyper-converged infrastructure |
US11231951B2 (en) * | 2018-12-19 | 2022-01-25 | Vmware, Inc. | Fault tolerant hyper-converged infrastructure upgrades in an environment with no additional physical infrastructure |
CN109901867B (en) * | 2019-02-14 | 2022-07-26 | 浙江可胜技术股份有限公司 | BOOT area code online upgrading method, heliostat controller and heliostat |
US11681519B2 (en) | 2019-10-14 | 2023-06-20 | Hewlett Packard Enterprise Development Lp | Software update on a network device forming a multi-chassis link aggregation group (MC-LAG) |
US11658899B2 (en) * | 2021-01-22 | 2023-05-23 | Vmware, Inc. | Routing configuration for data center fabric maintenance |
WO2023117593A1 (en) * | 2021-12-22 | 2023-06-29 | Hirschmann Automation And Control Gmbh | Software update sequence using link layer discovery protocol (lldp) |
US20230327981A1 (en) * | 2022-04-12 | 2023-10-12 | Hewlett Packard Enterprise Development Lp | Efficient traffic redirection for an mclag for controlled unavailability events |
US20240223448A1 (en) * | 2022-12-28 | 2024-07-04 | Dish Wireless L.L.C. | In service upgrade of radio access network radio unit of a cellular network |
US20240220231A1 (en) * | 2022-12-28 | 2024-07-04 | Dish Wireless L.L.C. | In service upgrade of radio access network distributed unit of a cellular network |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030202473A1 (en) * | 2002-04-25 | 2003-10-30 | General Instrument Corporation | Traffic network flow control using dynamically modified metrics for redundancy connections |
US20110158113A1 (en) * | 2009-12-31 | 2011-06-30 | Nanda Avoy K | Driven multicast traffic distribution on link-aggregate-group |
US20140019661A1 (en) * | 2012-05-18 | 2014-01-16 | Dell Products, Lp | System and Method for Providing Network Access for a Processing Node |
US20140032889A1 (en) * | 2012-07-30 | 2014-01-30 | David J. Koenen | Network booting a machine coupled to the network by a link aggregation group |
US8943490B1 (en) * | 2012-09-28 | 2015-01-27 | Juniper Networks, Inc. | Intelligent non-stop software upgrade |
Family Cites Families (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6490244B1 (en) * | 2000-03-09 | 2002-12-03 | Nortel Networks Limited | Layer 3 routing in self-healing networks |
US6594744B1 (en) | 2000-12-11 | 2003-07-15 | Lsi Logic Corporation | Managing a snapshot volume or one or more checkpoint volumes with multiple point-in-time images in a single repository |
US6944732B2 (en) | 2002-05-08 | 2005-09-13 | Hewlett-Packard Development Company, L.P. | Method and apparatus for supporting snapshots with direct I/O in a storage area network |
CA2428517A1 (en) * | 2002-05-13 | 2003-11-13 | Tropic Networks Inc. | System and method for distributed resource reservation protocol - traffic engineering (rsvp-te) hitless restart in multi-protocol label switching (mpls) network |
US7383541B1 (en) * | 2003-08-07 | 2008-06-03 | Cisco Technology, Inc. | Method and apparatus providing interoperation of execution images of different versions |
US7580360B2 (en) * | 2003-10-14 | 2009-08-25 | Cisco Technology, Inc. | Method and apparatus for generating routing information in a data communications network |
US7506335B1 (en) | 2003-11-29 | 2009-03-17 | Cisco Technology, Inc. | Method and apparatus for software loading and initialization in a distributed network |
US7483370B1 (en) * | 2003-12-22 | 2009-01-27 | Extreme Networks, Inc. | Methods and systems for hitless switch management module failover and upgrade |
US7360208B2 (en) | 2004-05-17 | 2008-04-15 | Oracle International Corp. | Rolling upgrade of distributed software with automatic completion |
US7779404B2 (en) | 2004-06-10 | 2010-08-17 | Cisco Technology, Inc. | Managing network device configuration using versioning and partitioning |
US7398382B2 (en) * | 2004-12-29 | 2008-07-08 | Intel Corporation | Method and apparatus to enhance platform boot efficiency |
US7484131B2 (en) | 2005-09-13 | 2009-01-27 | International Business Machines Corporation | System and method for recovering from a hang condition in a data processing system |
US8448162B2 (en) * | 2005-12-28 | 2013-05-21 | Foundry Networks, Llc | Hitless software upgrades |
US8274989B1 (en) * | 2006-03-31 | 2012-09-25 | Rockstar Bidco, LP | Point-to-multipoint (P2MP) resilience for GMPLS control of ethernet |
US7916657B2 (en) * | 2008-01-22 | 2011-03-29 | At&T Intellectual Property Ii, L.P. | Network performance and reliability evaluation taking into account abstract components |
US7830785B2 (en) * | 2008-01-25 | 2010-11-09 | At&T Labs, Inc. | System and method for restoration in a multimedia IP network |
US8315157B2 (en) * | 2008-07-02 | 2012-11-20 | Cisco Technology, Inc. | Graceful removal and/or insertion of a network element |
US7995481B2 (en) * | 2008-07-31 | 2011-08-09 | Cisco Technology, Inc. | Hop cost as secondary metric for equal cost multi-paths |
US8627137B1 (en) * | 2010-09-16 | 2014-01-07 | Cisco Technology, Inc. | Graceful handling of critical traffic blackholing faults |
EP2469762A1 (en) * | 2010-12-24 | 2012-06-27 | British Telecommunications Public Limited Company | Communications network management |
CN102231902B (en) | 2011-01-18 | 2014-07-30 | 华为技术有限公司 | Terminal selection network method and dual-function terminal |
JP5677173B2 (en) | 2011-04-11 | 2015-02-25 | キヤノン株式会社 | Image forming apparatus, network system, image forming apparatus control method, and program |
US8930318B1 (en) | 2011-09-29 | 2015-01-06 | Google Inc. | Systems and methods for handling interruptions while updating of an electronic device |
US8966467B2 (en) * | 2012-04-30 | 2015-02-24 | Dell Products, L.P. | System and method for performing an in-service software upgrade in non-redundant systems |
US9131014B2 (en) * | 2012-08-20 | 2015-09-08 | Cisco Technology, Inc. | Hitless pruning protocol upgrade on single supervisor network devices |
US9038053B2 (en) * | 2012-08-27 | 2015-05-19 | Lenovo Enterprise Solutions (Singapore) Pte. Ltd | Non-disruptive software updates for servers processing network traffic |
US9755900B2 (en) * | 2013-03-11 | 2017-09-05 | Amazon Technologies, Inc. | Managing configuration updates |
US9338055B2 (en) * | 2013-03-15 | 2016-05-10 | Cisco Technology, Inc. | Virtual router upgrade via graceful restart |
US9141373B2 (en) * | 2013-07-31 | 2015-09-22 | Arista Networks, Inc. | System and method for accelerated software upgrades |
US9537757B2 (en) * | 2013-08-12 | 2017-01-03 | Nec Corporation | Adding link aggregation group to hitless protection in packet switched networks |
US10009261B2 (en) | 2014-08-21 | 2018-06-26 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and system of checkpoint and rollback recovery for forwarding states of a software-defined networking (SDN) system |
US20160055579A1 (en) * | 2014-08-22 | 2016-02-25 | Vmware, Inc. | Decreasing time to market of a pre-configured hyper-converged computing device |
US10938653B2 (en) | 2015-04-21 | 2021-03-02 | Arista Networks, Inc. | System and method of updating a network |
US10545753B2 (en) * | 2015-04-21 | 2020-01-28 | Arista Networks, Inc. | System and method of updating a network element |
US9880829B2 (en) * | 2015-12-14 | 2018-01-30 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and apparatus for performing hitless update of line cards of a network device |
US10732961B1 (en) * | 2017-12-02 | 2020-08-04 | Juniper Networks, Inc | Systems and methods for performing in-service software upgrades on active network devices |
US10938623B2 (en) * | 2018-10-23 | 2021-03-02 | Hewlett Packard Enterprise Development Lp | Computing element failure identification mechanism |
US10846179B2 (en) * | 2018-11-05 | 2020-11-24 | Arista Networks, Inc. | Hitless repair for network device components |
-
2016
- 2016-03-01 US US15/057,880 patent/US10545753B2/en active Active
- 2016-04-18 WO PCT/US2016/028171 patent/WO2016172060A1/en unknown
- 2016-04-18 EP EP16783664.2A patent/EP3278215B1/en active Active
-
2019
- 2019-12-18 US US16/719,579 patent/US11249747B2/en active Active
-
2021
- 2021-12-30 US US17/566,227 patent/US20220121439A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030202473A1 (en) * | 2002-04-25 | 2003-10-30 | General Instrument Corporation | Traffic network flow control using dynamically modified metrics for redundancy connections |
US20110158113A1 (en) * | 2009-12-31 | 2011-06-30 | Nanda Avoy K | Driven multicast traffic distribution on link-aggregate-group |
US20140019661A1 (en) * | 2012-05-18 | 2014-01-16 | Dell Products, Lp | System and Method for Providing Network Access for a Processing Node |
US20140032889A1 (en) * | 2012-07-30 | 2014-01-30 | David J. Koenen | Network booting a machine coupled to the network by a link aggregation group |
US8943490B1 (en) * | 2012-09-28 | 2015-01-27 | Juniper Networks, Inc. | Intelligent non-stop software upgrade |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106953798A (en) * | 2017-05-25 | 2017-07-14 | 深圳市伊特利网络科技有限公司 | Link selecting method and system based on access point |
US11303533B2 (en) | 2019-07-09 | 2022-04-12 | Cisco Technology, Inc. | Self-healing fabrics |
Also Published As
Publication number | Publication date |
---|---|
EP3278215B1 (en) | 2020-01-29 |
US20200117441A1 (en) | 2020-04-16 |
EP3278215A4 (en) | 2018-10-24 |
EP3278215A1 (en) | 2018-02-07 |
US11249747B2 (en) | 2022-02-15 |
US20220121439A1 (en) | 2022-04-21 |
US20160313985A1 (en) | 2016-10-27 |
US10545753B2 (en) | 2020-01-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11249747B2 (en) | System and method of updating a network element | |
US9141373B2 (en) | System and method for accelerated software upgrades | |
US20180248769A1 (en) | Simulating a topology of network elements | |
US7769886B2 (en) | Application based active-active data center network using route health injection and IGP | |
EP2774329B1 (en) | Data center network architecture | |
US7463579B2 (en) | Routed split multilink trunking | |
US9300528B2 (en) | Trill network with multipath redundancy | |
US9774524B2 (en) | Method and apparatus for fast reroute, control plane and forwarding plane synchronization | |
US11290417B2 (en) | System and method of persistent address resolution synchronization | |
US20110231833A1 (en) | Upgrading system software in a chassis without traffic loss | |
US11018977B2 (en) | Pre-built match-action tables | |
US20160019044A1 (en) | Systems and methods for performing uninterrupted network upgrades with controllers | |
CN110061912B (en) | Arbitrating mastership between redundant control planes of virtual nodes | |
US20110149743A1 (en) | Network distribution prevention when virtual chassis system undergoes splits and merges | |
WO2016174598A1 (en) | Sdn network element affinity based data partition and flexible migration schemes | |
US9806989B2 (en) | Layer 3 (L3) best route selection rule for shortest path bridging multicast (SPBM) networks | |
US10616046B2 (en) | System and method of handling a fault detection mechanism during a control plane failover | |
US11349758B2 (en) | Multihoming optimizations for fast failover in single-active networks | |
US10110474B2 (en) | Updating interface list of multi-link group | |
WO2017144944A1 (en) | Method and apparatus for improving convergence in a spring network | |
US20230283542A1 (en) | Control plane based enhanced TI-LFA node protection scheme for SR-TE paths |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16783664 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |