CN109617716B - Data center exception handling method and device - Google Patents

Data center exception handling method and device Download PDF

Info

Publication number
CN109617716B
CN109617716B CN201811452209.6A CN201811452209A CN109617716B CN 109617716 B CN109617716 B CN 109617716B CN 201811452209 A CN201811452209 A CN 201811452209A CN 109617716 B CN109617716 B CN 109617716B
Authority
CN
China
Prior art keywords
data center
resource pool
vxlan tunnel
nfvo
gateway
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811452209.6A
Other languages
Chinese (zh)
Other versions
CN109617716A (en
Inventor
袁甲鸣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
New H3C Technologies Co Ltd Hefei Branch
Original Assignee
New H3C Technologies Co Ltd Hefei Branch
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by New H3C Technologies Co Ltd Hefei Branch filed Critical New H3C Technologies Co Ltd Hefei Branch
Priority to CN201811452209.6A priority Critical patent/CN109617716B/en
Publication of CN109617716A publication Critical patent/CN109617716A/en
Application granted granted Critical
Publication of CN109617716B publication Critical patent/CN109617716B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/06Management of faults, events, alarms or notifications
    • H04L41/0654Management of faults, events, alarms or notifications using network fault recovery
    • H04L41/0668Management of faults, events, alarms or notifications using network fault recovery by dynamic selection of recovery network elements, e.g. replacement by the most appropriate element after failure
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2854Wide area networks, e.g. public data networks
    • H04L12/2856Access arrangements, e.g. Internet access
    • H04L12/2869Operational details of access network equipments
    • H04L12/287Remote access server, e.g. BRAS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/46Interconnection of networks
    • H04L12/4633Interconnection of networks using encapsulation techniques, e.g. tunneling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/46Interconnection of networks
    • H04L12/4641Virtual LANs, VLANs, e.g. virtual private networks [VPN]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0823Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability
    • H04L41/0836Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability to enhance reliability, e.g. reduce downtime
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • H04L67/025Protocols based on web technology, e.g. hypertext transfer protocol [HTTP] for remote control or remote monitoring of applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1004Server selection for load balancing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1095Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]

Abstract

The disclosure relates to a data center exception handling method and device, which comprises the steps of monitoring the state of each data center; when the second data center is monitored to be in an abnormal state, an interface opening instruction is sent to a first resource pool gateway deployed in the first data center so as to open a first VXLAN tunnel established by the first resource pool gateway and a convergence switch borne by the second data center and open a second VXLAN tunnel established by the first resource pool gateway and a backup vBRAS in the first data center, so that the first data center bears the service of a user accessed through the convergence switch. By switching the VXLAN tunnel of the resource pool gateway of the abnormal data center to the VXLAN tunnel of the resource pool gateway of the backup data center, the data center abnormality processing method and device according to the embodiment of the disclosure can improve the reliability of the data center.

Description

Data center exception handling method and device
Technical Field
The present disclosure relates to the field of communications technologies, and in particular, to a method and an apparatus for processing an exception in a data center.
Background
NFVO (Network Function Virtualization editor) is a system for NFV (Network Function Virtualization) product lifecycle management and virtual, physical Network element service orchestration. The NFVO can manage the life cycle of the NFV and flexibly coordinate the management and scene deployment of various resources according to the business requirements of users.
In a resource Pool deployment mode of a metropolitan Area Network (DC) (Data Center), a vBRAS (Virtualized Broadband Remote Access Server) resource Pool is connected to a Pool-Gateway (Pool Gateway), a Virtual eXtensible Local Area Network (VXLAN) tunnel is established between the vBRAS and the Pool-Gateway through NFVO, a VXLAN tunnel is established between the HJSW and the Pool-Gateway, and the VXLAN tunnel is established between the HJSW and the Pool-Gateway, and then two sections of VXLAN tunnels are communicated through a VXLAN tunnel bonding technology, so that the vBRAS resource Pool in the DC carries services of the HJW Access user.
In the related technology, in a dual-DC-vBRAS resource pool deployment mode, NFVO is deployed in DC in a cluster mode, and vBRAS resource pools in different DCs bear services of different HJSW access users, so that load balancing is realized. If a certain DC fails, the DC cannot continuously bear the service of the HJSF access user, and service interruption is caused.
Disclosure of Invention
In view of this, the present disclosure provides a method and an apparatus for processing an exception in a data center, which can improve reliability of the data center.
According to an aspect of the present disclosure, a data center exception handling method is provided, where an NFVO in an NFVO cluster is deployed in a data center of a dual data center networking, where the data center of the dual data center networking includes a first data center and a second data center, and the method is applied to the NFVO cluster, and the method includes: monitoring the state of each data center; when the second data center is monitored to be in an abnormal state, an interface opening instruction is sent to a first resource pool gateway deployed in the first data center so as to open a first VXLAN tunnel established by the first resource pool gateway and a convergence switch borne by the second data center and open a second VXLAN tunnel established by the first resource pool gateway and a backup virtual broadband remote access server vBRAS in the first data center, so that the first data center bears the service of a user accessed through the convergence switch.
According to another aspect of the present disclosure, there is provided a data center exception handling apparatus, an NFVO in an NFVO cluster is deployed in a data center of a dual data center networking, where the data center of the dual data center networking includes a first data center and a second data center, and the apparatus is applied to the NFVO cluster, and the apparatus includes: the monitoring module is used for monitoring the state of each data center; and the sending module is used for sending an interface opening instruction to a first resource pool gateway deployed in the first data center when the second data center is monitored to be in an abnormal state so as to open a first VXLAN tunnel established by the first resource pool gateway and a convergence switch borne by the second data center and open a second VXLAN tunnel established by the first resource pool gateway and a backup virtual broadband remote access server vBRAS in the first data center, so that the first data center bears the service of a user accessed through the convergence switch.
In the embodiment of the present disclosure, the state of each data center is monitored by the NFVO cluster, and when the data center is abnormal, the VXLAN tunnel of the resource pool gateway of the abnormal data center is switched to the VXLAN tunnel of the resource pool gateway of the backup data center, so that the backup data center can bear the service borne by the abnormal data center, thereby avoiding interruption of the service borne by the abnormal data center, and improving the reliability of the data center.
Other features and aspects of the present disclosure will become apparent from the following detailed description of exemplary embodiments, which proceeds with reference to the accompanying drawings.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate exemplary embodiments, features, and aspects of the disclosure and, together with the description, serve to explain the principles of the disclosure.
Fig. 1a shows a schematic diagram of a dual DC networking in the related art.
Fig. 1b shows a schematic diagram of a configuration of the VSI on the POOL-GW1 based on fig. 1 a.
Fig. 1c shows a schematic diagram of the configuration of VSI on HJSW1 based on fig. 1 a.
Fig. 2a shows a schematic diagram of a dual DC networking according to an embodiment of the present disclosure.
Fig. 2b shows a schematic configuration diagram of a VSI on the POOL-GW1 based on fig. 2a in the embodiment of the present disclosure.
Fig. 2c shows a schematic configuration diagram of a VSI on HJSW1 based on fig. 2a in an embodiment of the present disclosure.
FIG. 3 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure.
FIG. 4 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure.
FIG. 5 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure.
Fig. 6 illustrates a block diagram of a data center exception handling apparatus according to an embodiment of the present disclosure.
FIG. 7 is a block diagram illustrating an apparatus 900 for data center exception handling in accordance with an exemplary embodiment.
Detailed Description
Various exemplary embodiments, features and aspects of the present disclosure will be described in detail below with reference to the accompanying drawings. In the drawings, like reference numbers can indicate functionally identical or similar elements. While the various aspects of the embodiments are presented in drawings, the drawings are not necessarily drawn to scale unless specifically indicated.
The word "exemplary" is used exclusively herein to mean "serving as an example, embodiment, or illustration. Any embodiment described herein as "exemplary" is not necessarily to be construed as preferred or advantageous over other embodiments.
Furthermore, in the following detailed description, numerous specific details are set forth in order to provide a better understanding of the present disclosure. It will be understood by those skilled in the art that the present disclosure may be practiced without some of these specific details. In some instances, methods, means, elements and circuits that are well known to those skilled in the art have not been described in detail so as not to obscure the present disclosure.
In the related technology, in a dual DC-vbars resource pool deployment manner, NFVO is deployed in a data center in a cluster manner, and different data centers carry different services of users accessed by HJSW, thereby realizing load balancing.
Fig. 1a shows a schematic diagram of a dual DC networking in the related art. As shown in fig. 1a, NFVO1 and NFVO2 represent NFVO within NFVO cluster, DC1 and DC2 represent data centers, POOL-GW1 and POOL-GW2 represent resource POOL gateways, and HJSW1 and HJSW2 represent aggregation switches.
As shown in fig. 1a, NFVO1 and NFVO2 are disposed within DC1 and DC2, respectively. A VXLAN tunnel from POOL-GW1 to HJSW1 and a VXLAN tunnel from POOL-GW1 to service vbars in DC1 are respectively configured on POOL-GW1 deployed inside DC 1. The service vbars of the DC1 may be used to carry the service of the user accessed by the HJSW 1.
Fig. 1b shows a schematic configuration diagram of VSI on the POOL-GW1 based on fig. 1a, where DC1_1 is the identifier of the VSI on the POOL-GW1, VXLAN1 is the VXLAN identifier corresponding to the VSI, tunnel11 may represent VXLAN tunnels from the POOL-GW1 to HJSW1, and tunnel21 may represent VXLAN tunnels from the POOL-GW1 to the service vbars inside DC 1.
Fig. 1c shows a schematic diagram of the configuration of VSI on HJSW1 based on fig. 1 a. Wherein, DC1 is the identifier of VSI on HJSW1, VXLAN1 is the VXLAN identifier corresponding to the VSI, and tunnel1 can represent the VXLAN tunnel from HJSW1 to POOL-GW 1.
Through a VXLAN tunnel between POOL-GW1 and HJSW1 and a VXLAN tunnel between POOL-GW1 and a service vBRAS inside DC1, the DC1 can bear the service of a user accessed by HJSW 1. DC2 in fig. 1a may refer to DC1, the configuration of VSI on POOL-GW2 may refer to POOL-GW1, and the configuration of VSI on HJSW2 may refer to HJSW1, which is not described herein.
Thus, in the related art, DC1 carries traffic of users accessed by HJSW1, and DC2 carries traffic of users accessed by HJSW 2. If the DC1 is abnormal, the service interruption of the user accessed by the HJSW1 is caused; if the DC2 is abnormal, the service of the user accessed by the HJSW2 is interrupted. Therefore, in the related art, the reliability of the data center is poor.
In order to make those skilled in the art better understand the technical solutions in the embodiments of the present disclosure, some technical terms related to the embodiments of the present disclosure are briefly described below.
A resource pool gateway represents a gateway device of a vbars resource pool consisting of a plurality of vbars.
The aggregation switch carried by the first data center represents an aggregation switch for carrying traffic of a user accessing the first data center. In this way, the first data center can carry the traffic of the user accessed by the aggregation switch.
The aggregation switch carried by the second data center represents an aggregation switch for carrying traffic of users accessing the second data center. In this way, the second data center can carry the traffic of the users accessed by the aggregation switch.
Fig. 2a shows a schematic diagram of a dual DC networking according to an embodiment of the present disclosure. As shown in fig. 2a, the dual DC networking of the embodiment of the present disclosure adds backup vbars in DC1 and DC2, and adds VXLAN tunnel between the POOL-GW1 and the backup vbars of DC1, and VXLAN tunnel between the POOL-GW1 and the HJSW2, respectively, on the basis of the dual DC networking shown in fig. 1 a. Accordingly, a VXLAN tunnel between POOL-GW2 and HJSW1 is added, as well as a VXLAN tunnel between POOL-GW2 and the backup vbars of DC 2. The backup vBRAS can be created through the NFVO cluster, and the backup vBRAS does not carry services under normal conditions.
It should be noted that the DC2 shown in fig. 2a can carry the services of users accessed by multiple aggregation switches, that is, the DC2 can carry not only the services of users accessed by the HJSW2, but also the DC2 can carry the services of users accessed by other aggregation switches. Wherein each aggregation switch may correspond to a first VXLAN tunnel and a second VXLAN tunnel. In the embodiment of the present disclosure, the second data center DC2 is used to carry the service of the user accessed by one aggregation switch (for example, HJSW2 shown in fig. 2 a), and the embodiment of the present disclosure is also applicable to the case where the second data center DC2 carries the service of the user accessed by multiple aggregation switches.
For ease of description, the method of the present invention is described based on a dual DC networking as shown in fig. 2a, although the present disclosure is equally applicable to a networking where there are multiple DCs. In the embodiment of the present disclosure, DC1 shown in fig. 2a is taken as a first data center, DC2 is taken as a second data center, POOL-GW1 is taken as a first resource POOL gateway, HJSW2 is a convergence switch carried by the second data center, a VXLAN tunnel between POOL-GW1 and HJSW2 is taken as a first VXLAN tunnel, and a VXLAN tunnel between POOL-GW1 and a backup vBRAS of DC1 is taken as a second VXLAN tunnel.
FIG. 3 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure. The data center exception handling method may be applied to an NFVO cluster, where NFVOs within the NFVO cluster are deployed in data centers of a dual-DC networking shown in fig. 2a, where the data centers of the dual-DC networking include a first data center (DC 1 shown in fig. 2 a) and a second data center (DC 2 shown in fig. 2 a). As shown in fig. 3, the data center exception handling method may include:
step S11, the status of each data center is monitored.
Step S12, when it is monitored that the second data center is in an abnormal state, sending an interface start command to a first resource pool gateway deployed inside the first data center, so as to start a first VXLAN tunnel established between the first resource pool gateway and a convergence switch carried by the second data center, and start a second VXLAN tunnel established between the first resource pool gateway and a backup vBRAS inside the first data center, so that the first data center carries a service of a user accessed through the convergence switch.
In the embodiment of the present disclosure, the state of the data center is monitored by the NFVO cluster, and when the data center is abnormal, the VXLAN tunnel of the resource pool gateway of the abnormal data center is switched to the VXLAN tunnel of the resource pool gateway of the backup data center, so that the backup data center can bear the service borne by the abnormal data center, thereby avoiding interruption of the service borne by the abnormal data center, and improving reliability of the data center.
The NFVOs are deployed in the data centers in a cluster mode, and the NFVO cluster can monitor the state of each data center through the NFVOs deployed in each data center. The state of the data center includes a normal state and an abnormal state. The normal state may indicate that the data center is working normally, and the data center may normally carry the service of the user accessed by the corresponding aggregation switch. The abnormal state may indicate that the data center fails to normally carry the service of the user accessed by the corresponding aggregation switch. For example, when a first data center and a second data center are included in a dual DC networking, the NFVO cluster may monitor the status of the first data center through the NFVO (NFVO 1 shown in fig. 2 a) deployed in the first data center, and monitor the status of the second data center through the NFVO (NFVO 2 shown in fig. 2 a) deployed in the second data center.
In this embodiment of the present disclosure, the second data center may be any one data center, and the first data center may be any one data center, where the first data center may also serve as a backup data center of the second data center while bearing a service of a user corresponding to the access of the aggregation switch, so as to bear a service originally borne by the second data center when the second data center is abnormal. The NFVO cluster may create a backup vBRAS inside a first data center, establish a first VXLAN tunnel between the first resource pool gateway and a convergence switch carried by a second data center, and establish a second VXLAN tunnel between the first resource pool gateway and the backup vBRAS inside the first data center.
In a possible implementation manner, the second data center includes a second resource pool gateway and a service vbars, a third VXLAN tunnel is established between the second resource pool gateway and a convergence switch carried by the second data center, and a fourth VXLAN tunnel is established between the second resource pool gateway and the service vbars in the second data center. A connection between the aggregation switch carried by the second data center and the service vbars in the second data center may be established through the third VXLAN tunnel and the fourth VXLAN tunnel. In this way, the user equipment accessed by the aggregation switch carried by the second data center may send the service packet to the service vbars in the second data center.
The first VXLAN tunnel may represent a VXLAN tunnel between the first resource pool gateway and a convergence switch carried by the second data center, and the second VXLAN tunnel may represent a VXLAN tunnel between the first resource pool gateway and a backup vBRAS inside the first data center. A backup vbars in a first data center refers to a backup of a service vbars in a second data center.
A connection between the aggregation switch carried by the second data center and the backup vbars inside the first data center can be established through the first VXLAN tunnel and the second VXLAN tunnel.
In this way, the aggregation switch carried by the second data center is connected to both the backup vbars in the first data center and the service vbars in the second data center.
The NFVO cluster may close the first VXLAN tunnel and the second VXLAN tunnel while the second data center is in a normal state. Therefore, when the second data center is in a normal state, the user equipment accessed by the aggregation switch carried by the second data center can send the service message to the service vBRAS inside the second data center, but cannot send the service message to the backup vBRAS inside the first data center.
In a possible implementation manner, the NFVO cluster may set a first VSI corresponding to a first VXLAN tunnel and a second VXLAN tunnel in a first resource pool gateway deployed inside a second data center to a closed state, that is, a shutdown state. When the first VSI is in a closed state, although the user equipment accessed by the aggregation switch carried by the second data center may send the service packet to the first resource pool gateway, the service packet cannot be sent to the backup vbars inside the first data center.
In one example, fig. 2b shows a schematic configuration diagram based on VSI on POOL-GW1 of fig. 2a in an embodiment of the present disclosure. As shown in fig. 2b, DC2_1 is the identity of the VSI on the POOL-GW 1; VXLAN3 is the VXLAN identifier corresponding to the VSI; tunnel13 may represent the VXLAN tunnel from POOL-GW1 to HJSW2, i.e., the first VXLAN tunnel; tunnel22 may represent a tunnel from POOL-GW1 to a backup vbars deployed inside DC1, i.e., a second VXLAN tunnel. When DC1 is in a normal state, DC2_1 corresponding to tunnel13 and tunnel22 is in a shutdown state, that is, the first VSI corresponding to the first VXLAN tunnel and the second VXLAN tunnel is in a closed state. Fig. 2c shows a schematic configuration diagram of a VSI on HJSW1 based on fig. 2a in an embodiment of the present disclosure. Wherein, DC1 is the identifier of VSI on HJSW1, VXLAN1 is the VXLAN identifier corresponding to the VSI, and tunnel2 can represent the VXLAN tunnel from HJSW1 to POOL-GW 2. Reference may be made to HJSW1 for HJSW2, which is not described here.
When the second data center is in an abnormal state, it indicates that the second data center may have a fault, and the second data center may not be able to continue to carry the service of the user originally accessed by the aggregation switch carried in the second data, thereby causing the service interruption. The NFVO cluster may start a backup data center of the second data center when it is monitored that the second data center is in an abnormal state, so that the backup data center carries a service of a user, which is originally carried by the aggregation switch carried by the second data center and accessed by the aggregation switch.
In the embodiment of the present disclosure, a first data center is taken as a backup data center of a second data center as an example. The NFVO cluster may send an interface start command to a first resource pool gateway deployed in the first data center when it is monitored that the second data center is in an abnormal state. The interface open command may indicate opening of the first VXLAN tunnel and opening of the second VXLAN tunnel. After the first VXLAN tunnel and the second VXLAN tunnel are opened, the user equipment accessed by the aggregation switch carried by the second data center can send the service message to the first resource pool gateway in the first data center, and then send the service message to the backup vbars in the first data center, so that the service of the user accessed by the aggregation switch originally carried by the second data center can be normally processed.
In a possible implementation manner, the interface start command may be an undo shutdown command, and the interface start command may instruct the first resource pool gateway to start the first VSI corresponding to the first VXLAN tunnel and the second VXLAN tunnel. And after receiving the interface opening command, the first resource pool gateway sets the first VSI to be in an opening state. When the first VSI is in the on state, the user equipment accessed by the aggregation switch carried by the second data center may send the service packet to the first resource pool gateway, and then to the backup vbars in the first data center.
Application example
As shown in fig. 2a, the NFVO cluster may monitor the status of DC1 and DC 2.
When the DC2 is in a normal state, the HJSW2 can access the service vbars deployed inside the DC2 through the third VXLAN tunnel and the fourth VXLAN tunnel, so that the DC2 can bear the service of the user accessed by the HJSW 2.
When the second data center DC2 is monitored to be in an abnormal state, the NFVO cluster may open the first VXLAN tunnel and the second VXLAN tunnel. At this time, the HJSW2 accesses the backup vbars of the first data center DC1 through the first VXLAN tunnel and the second VXLAN tunnel, so that the first data center DC2 carries the services of the user accessed by the HJSW 2.
In one possible implementation manner, when it is monitored that the second data center recovers to a normal state, the first VXLAN tunnel and the second VXLAN tunnel are closed. For example, the NFVO cluster may send a shutdowm instruction to the first resource pool gateway to close the first VSI.
The first data center is used as a backup data center of the second data center, and after the second data center is restored, the second data center can normally bear the originally-borne service of the user accessed by the aggregation switch. Therefore, the NFVO cluster can close the first VXLAN tunnel and the second VXLAN tunnel, so that the first data center does not bear the service of the user accessed by the aggregation switch originally borne by the second data center any more, and resources of the vbars resource pool are saved.
FIG. 4 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure. As shown in fig. 4, the data center exception handling method may further include:
and step S13, establishing a session with a second resource pool gateway deployed in the second data center.
And step S14, acquiring the state of the session at intervals of a first time length.
Step S15, when the number of times that the session is in the unavailable state is greater than a number threshold, determining that the second data center is in an abnormal state.
When a resource pool gateway in the data center fails, the data center cannot bear the service of the aggregation switch access user. In this case, the NFVO cluster may determine that the data center is failed, i.e., the data center is in an abnormal state.
The NFVO cluster may establish a session with a second resource pool gateway. In one example, the NFVO2 deployed in the second data center in the NFVO cluster may establish a session with the second resource pool gateway through NETCONF (Network Configuration Protocol). The NFVO2 may obtain the state of the session once every first duration. When the number of times that the session is in the unavailable state is greater than the number threshold, indicating that the connection between the NFVO2 and the second resource pool gateway is abnormal, the NFVO2 may determine that the second resource pool gateway fails, and may further determine that the second data center is in an abnormal state.
The first duration and the number threshold may be set as needed, and the disclosure is not limited thereto.
In a possible implementation manner, the NFVO cluster may establish a session with the second resource pool gateway through a NETCONF protocol.
FIG. 5 shows a flow diagram of a data center exception handling method according to an embodiment of the present disclosure. As shown in fig. 5, the data center exception handling method may further include:
step S16, monitoring heartbeat messages of each NFVO in the cluster.
Step S17, if the heartbeat message sent by the NFVO corresponding to the second data center is not received in the cluster within the second duration, it is determined that the second data center is in an abnormal state.
In the embodiment of the present disclosure, the NFVO may monitor the status of other NFVOs in the cluster through inter-cluster heartbeat messages (e.g., ping). If the heartbeat message sent by the NFVO corresponding to the second data center is not received in the cluster within the second duration, the NFVO cluster may determine that the NFVO fails. At this time, the NFVO cluster cannot monitor the state of the resource pool gateway in the data center through the NFVO deployed in the data center, or the NFVO cluster cannot know the state information monitored by the NFVO deployed in the data center. Under the condition, the NFVO cluster determines that the second data center is in an abnormal state, so that the backup data center can be started, and the reliability of the data center is improved.
The second duration may be set as needed, and the disclosure is not limited thereto.
It should be noted that, in the dual-DC networking, the NFVO cluster may deploy the second data center as the same as the first data center, so that the second data center may also serve as a backup data center of the first data center while bearing the service of the user corresponding to the access of the aggregation switch, so as to bear the service originally borne by the first data center when the first data center is abnormal. The processing mode of the first data center for the occurrence of the abnormality may refer to the processing mode of the second data center for the occurrence of the abnormality in the embodiment of the present disclosure, and details are not described here.
Fig. 6 illustrates a block diagram of a data center exception handling apparatus according to an embodiment of the present disclosure. The NFVO in the NFVO cluster is deployed in a data center of a dual-data-center networking, the data center of the dual-data-center networking comprises a first data center and a second data center, and the device is applied to the NFVO cluster. As shown in fig. 6, the apparatus 60 may include:
the monitoring module 61 is used for monitoring the state of each data center;
a sending module 62, configured to send an interface start instruction to a first resource pool gateway deployed in the first data center when it is monitored that the second data center is in an abnormal state, so as to start a first VXLAN tunnel established between the first resource pool gateway and a convergence switch carried by the second data center, and start a second VXLAN tunnel established between the first resource pool gateway and a backup virtual broadband remote access server vbrs inside the first data center, so that the first data center carries a service of a user accessed through the convergence switch.
In a possible implementation manner, the monitoring module is specifically configured to:
establishing a session with a second resource pool gateway deployed within the second data center;
acquiring the state of the session at intervals of a first duration;
when the number of times that the session is in the unavailable state is greater than a number threshold, determining that the second data center is in an abnormal state.
In a possible implementation manner, the monitoring module is specifically configured to:
monitoring heartbeat messages of each NFVO in the cluster;
and if the heartbeat message sent by the NFVO corresponding to the second data center is not received in the cluster within the second duration, determining that the second data center is in an abnormal state.
In one possible implementation, the apparatus 60 may further include:
and the closing module is used for closing the first VXLAN tunnel and the second VXLAN tunnel when the second data center is monitored to be recovered to the normal state.
In one possible implementation, establishing a session with the second resource pool gateway includes:
and establishing a session with the second resource pool gateway through a NETCONF protocol.
In the embodiment of the present disclosure, the state of each data center is monitored by the NFVO cluster, and when the data center is abnormal, the VXLAN tunnel of the resource pool gateway of the abnormal data center is switched to the VXLAN tunnel of the resource pool gateway of the backup data center, so that the backup data center can bear the service borne by the abnormal data center, thereby avoiding interruption of the service borne by the abnormal data center, and improving the reliability of the data center.
FIG. 7 is a block diagram illustrating an apparatus 900 for data center exception handling in accordance with an exemplary embodiment. Referring to fig. 7, the apparatus 900 may include a processor 901, a machine-readable storage medium 902 having stored thereon machine-executable instructions. The processor 901 and the machine-readable storage medium 902 may communicate via a system bus 903. Also, the processor 901 performs the data center exception handling method described above by reading machine executable instructions in the machine readable storage medium 902 corresponding to the data center exception handling logic.
The machine-readable storage medium 902 referred to herein may be any electronic, magnetic, optical, or other physical storage device that can contain or store information such as executable instructions, data, and the like. For example, the machine-readable storage medium may be: a RAM (random Access Memory), a volatile Memory, a non-volatile Memory, a flash Memory, a storage drive (e.g., a hard drive), a solid state drive, any type of storage disk (e.g., an optical disk, a dvd, etc.), or similar storage medium, or a combination thereof.
Having described embodiments of the present disclosure, the foregoing description is intended to be exemplary, not exhaustive, and not limited to the disclosed embodiments. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terms used herein were chosen in order to best explain the principles of the embodiments, the practical application, or technical improvements to the techniques in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.

Claims (10)

1. A data center exception handling method is characterized in that an NFVO in a Network Function Virtualization Orchestrator (NFVO) cluster is deployed in a data center of a dual-data-center networking, wherein the data center of the dual-data-center networking comprises a first data center and a second data center, and the method is applied to the NFVO cluster, and comprises the following steps:
monitoring the state of each data center;
when the second data center is monitored to be in an abnormal state, sending an interface opening instruction to a first resource pool gateway deployed in the first data center so as to open a first VXLAN tunnel established by the first resource pool gateway and a convergence switch borne by the second data center and open a second VXLAN tunnel established by the first resource pool gateway and a backup virtual broadband remote access server vBRAS in the first data center, so that the first data center bears the service of a user accessed through the convergence switch;
wherein, opening a first VXLAN tunnel established by the first resource pool gateway and the aggregation switch carried by the second data center, and opening a second VXLAN tunnel established by the first resource pool gateway and the backup virtual broadband remote access server vBRAS inside the first data center includes:
and opening a first virtual switch instance VSI which is configured on the first resource pool gateway and corresponds to the first VXLAN tunnel and the second VXLAN tunnel.
2. The method of claim 1, wherein the monitoring the status of each data center comprises:
establishing a session with a second resource pool gateway deployed within the second data center;
acquiring the state of the session at intervals of a first duration;
when the number of times that the session is in the unavailable state is greater than a number threshold, determining that the second data center is in an abnormal state.
3. The method of claim 1 or 2, wherein the monitoring the status of each data center comprises:
monitoring heartbeat messages of each NFVO in the cluster;
and if the heartbeat message sent by the NFVO corresponding to the second data center is not received in the cluster within the second duration, determining that the second data center is in an abnormal state.
4. The method of claim 1, further comprising:
and when the second data center is monitored to be recovered to a normal state, closing the first VXLAN tunnel and the second VXLAN tunnel.
5. The method of claim 2, wherein establishing a session with the second resource pool gateway comprises:
and establishing a session with the second resource pool gateway through a NETCONF protocol.
6. A data center exception handling apparatus, wherein an NFVO in a Network Function Virtualization Orchestrator (NFVO) cluster is deployed in a data center of a dual data center networking, the data center of the dual data center networking comprises a first data center and a second data center, and the apparatus is applied to the NFVO cluster, and the apparatus comprises:
the monitoring module is used for monitoring the state of each data center;
a sending module, configured to send an interface start instruction to a first resource pool gateway deployed in the first data center when it is monitored that the second data center is in an abnormal state, so as to start a first VXLAN tunnel established between the first resource pool gateway and a convergence switch carried by the second data center, and start a second VXLAN tunnel established between the first resource pool gateway and a backup virtual broadband remote access server vbars in the first data center, so that the first data center carries a service of a user accessed through the convergence switch;
wherein, opening a first VXLAN tunnel established by the first resource pool gateway and the aggregation switch carried by the second data center, and opening a second VXLAN tunnel established by the first resource pool gateway and the backup virtual broadband remote access server vBRAS inside the first data center includes:
and opening a first virtual switch instance VSI which is configured on the first resource pool gateway and corresponds to the first VXLAN tunnel and the second VXLAN tunnel.
7. The apparatus according to claim 6, wherein the monitoring module is specifically configured to:
establishing a session with a second resource pool gateway deployed within the second data center;
acquiring the state of the session at intervals of a first duration;
when the number of times that the session is in the unavailable state is greater than a number threshold, determining that the second data center is in an abnormal state.
8. The apparatus according to claim 6 or 7, wherein the monitoring module is specifically configured to:
monitoring heartbeat messages of each NFVO in the cluster;
and if the heartbeat message sent by the NFVO corresponding to the second data center is not received in the cluster within the second duration, determining that the second data center is in an abnormal state.
9. The apparatus of claim 6, further comprising:
and the closing module is used for closing the first VXLAN tunnel and the second VXLAN tunnel when the second data center is monitored to be recovered to the normal state.
10. The apparatus of claim 7, wherein establishing a session with the second resource pool gateway comprises:
and establishing a session with the second resource pool gateway through a NETCONF protocol.
CN201811452209.6A 2018-11-30 2018-11-30 Data center exception handling method and device Active CN109617716B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811452209.6A CN109617716B (en) 2018-11-30 2018-11-30 Data center exception handling method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811452209.6A CN109617716B (en) 2018-11-30 2018-11-30 Data center exception handling method and device

Publications (2)

Publication Number Publication Date
CN109617716A CN109617716A (en) 2019-04-12
CN109617716B true CN109617716B (en) 2022-02-25

Family

ID=66005529

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811452209.6A Active CN109617716B (en) 2018-11-30 2018-11-30 Data center exception handling method and device

Country Status (1)

Country Link
CN (1) CN109617716B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114760224A (en) * 2021-12-24 2022-07-15 中国银联股份有限公司 System, method, apparatus, and storage medium for monitoring status of network channels
CN115499299A (en) * 2022-09-13 2022-12-20 航天信息股份有限公司 Cluster equipment monitoring method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103516554A (en) * 2013-10-22 2014-01-15 杭州华三通信技术有限公司 Method and equipment for VM (Virtual Machine) migration
CN103631652A (en) * 2012-08-28 2014-03-12 中兴通讯股份有限公司 Method and system for achieving virtual machine migration
CN104798342A (en) * 2014-11-17 2015-07-22 华为技术有限公司 Service migration method of data center, device and system thereof
US9274907B1 (en) * 2013-12-19 2016-03-01 Emc Corporation Decommissioning of virtual backup appliances
CN105490937A (en) * 2014-09-17 2016-04-13 华为技术有限公司 Ethernet virtual network gateway switching method and service provider edge node equipment
CN105574590A (en) * 2015-12-28 2016-05-11 中国民航信息网络股份有限公司 Adaptive general control disaster recovery switching device and system, and signal generation method
CN108512702A (en) * 2018-03-27 2018-09-07 中国联合网络通信有限公司广东省分公司 A kind of system and method for realizing end-to-end quality guarantee between DC based on SR and ESI

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102025646B (en) * 2011-01-05 2013-08-28 杭州华三通信技术有限公司 Link switching method and device thereof
CN103259727B (en) * 2012-08-31 2016-12-28 杭州华三通信技术有限公司 A kind of message forwarding method and equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103631652A (en) * 2012-08-28 2014-03-12 中兴通讯股份有限公司 Method and system for achieving virtual machine migration
CN103516554A (en) * 2013-10-22 2014-01-15 杭州华三通信技术有限公司 Method and equipment for VM (Virtual Machine) migration
US9274907B1 (en) * 2013-12-19 2016-03-01 Emc Corporation Decommissioning of virtual backup appliances
CN105490937A (en) * 2014-09-17 2016-04-13 华为技术有限公司 Ethernet virtual network gateway switching method and service provider edge node equipment
CN104798342A (en) * 2014-11-17 2015-07-22 华为技术有限公司 Service migration method of data center, device and system thereof
CN105574590A (en) * 2015-12-28 2016-05-11 中国民航信息网络股份有限公司 Adaptive general control disaster recovery switching device and system, and signal generation method
CN108512702A (en) * 2018-03-27 2018-09-07 中国联合网络通信有限公司广东省分公司 A kind of system and method for realizing end-to-end quality guarantee between DC based on SR and ESI

Also Published As

Publication number Publication date
CN109617716A (en) 2019-04-12

Similar Documents

Publication Publication Date Title
US11223514B2 (en) Method and system of a dynamic high-availability mode based on current wide area network connectivity
CN112003961B (en) Resource exposure method, system, equipment and medium in kubernets
EP3895388B1 (en) Server redundant network paths
TWI612415B (en) System and management method of server and computer-readable storage medium
JP6466003B2 (en) Method and apparatus for VNF failover
CN109861867B (en) MEC service processing method and device
CN109756382B (en) Fault positioning method and device
CN109617716B (en) Data center exception handling method and device
CN113489149B (en) Power grid monitoring system service master node selection method based on real-time state sensing
CN110661599B (en) HA implementation method, device and storage medium between main node and standby node
CN107547255B (en) Cluster system upgrading method, main control board and cluster system
CN114124803B (en) Device management method and device, electronic device and storage medium
US10516625B2 (en) Network entities on ring networks
CN107360015B (en) Method and device for switching shared storage
US20180139113A1 (en) Efficiently Calculating Per Service Impact Of Ethernet Ring Status Changes
CN112131201B (en) Method, system, equipment and medium for high availability of network additional storage
CN116248484B (en) Management method and device of cloud primary integrated machine, electronic equipment and storage medium
US10122588B2 (en) Ring network uplink designation
CN108337147B (en) Message forwarding method and device
CN116846740A (en) Node fault processing method and device
CN115643161A (en) Out-of-band management method and system
CN111866041A (en) Method and device for selecting service equipment and updating cloud storage cluster and storage medium
CN117215722A (en) Virtual machine management method and device, electronic equipment and storage medium
CN113505022A (en) Hot backup processing method, device, equipment and machine readable storage medium
CN114726779A (en) Main/standby switching method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant