WO2017184807A1 - Parallel multipath routing architecture - Google Patents

Parallel multipath routing architecture Download PDF

Info

Publication number
WO2017184807A1
WO2017184807A1 PCT/US2017/028504 US2017028504W WO2017184807A1 WO 2017184807 A1 WO2017184807 A1 WO 2017184807A1 US 2017028504 W US2017028504 W US 2017028504W WO 2017184807 A1 WO2017184807 A1 WO 2017184807A1
Authority
WO
WIPO (PCT)
Prior art keywords
packets
overlay
computing device
paths
baseline
Prior art date
Application number
PCT/US2017/028504
Other languages
French (fr)
Inventor
Taric Mirza
Original Assignee
Taric Mirza
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Taric Mirza filed Critical Taric Mirza
Publication of WO2017184807A1 publication Critical patent/WO2017184807A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/24Multipath

Definitions

  • FIG. 1A is a drawing of a networked environment according to various embodiments of the present disclosure.
  • FIG. IB is a drawing of a networked environment according to various embodiments of the present disclosure.
  • FIG. 2 is a flowchart illustrating one example of functionality implemented as portions of an edge node 114b in the networked environment of FIGS. 1A or FIG. IB according to various embodiments of the present disclosure.
  • FIG. 3 is a flowchart illustrating one example of functionality implemented as portions of an edge node 114a in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
  • FIG. 4 is a flowchart illustrating one example of functionality implemented as portions of a controller application 137 executed in a controller computing environment in the networked environment of FIG. IB according to various embodiments of the present disclosure.
  • FIG. 5A is a schematic block diagram that provides one example illustration of a server computing environment employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
  • FIG. 5B is a schematic block diagram that provides one example illustration of a client employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure
  • FIG. 5C is a schematic block diagram that provides one example illustration of a controller computing environment employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
  • Internet access consumers may experience performance degradation due to packet loss.
  • a player in a networked video game may experience packet loss between the player and a server. Additional resources must be used in order to resend a lost packet between the player and server, which may introduce lag, speed degradation, and other issues that may affect many players.
  • stock traders communicating with a trading system may experience delays in consummating transactions due to additional resource expenditure in resending lost packets.
  • a parallel multipath architecture implements an overlay network on existing Internet Protocol network frameworks.
  • the overlay network includes edge nodes communicatively coupled to communication endpoints.
  • Relay nodes are communicatively coupled to the edge nodes and other relay nodes.
  • Packets from a source communication endpoint are modified at an edge node to include metadata including sequence numbers, destination network address information, or other data.
  • the edge node forwards the packets along multiple parallel paths of relay nodes within the overlay network.
  • An edge node communicatively coupled to the destination communication endpoint removes the added metadata, discards duplicate packets, and forwards the sequenced payloads to the destination.
  • the networked environment 100a includes a server computing environment 101, a client 104, which are in data communication with each other via an overlay network 107.
  • the overlay network 107 includes, for example, the Internet, wired networks, wireless networks, or other suitable networks, etc., or any combination of two or more such networks.
  • such networks may comprise satellite networks, cable networks, Ethernet networks, and other types of networks.
  • the overlay network 107 comprises one or more relay nodes 11 la-n and edge nodes 114a/b, which can include a subset of network components or nodes of a network upon which the overlay network 107 lies.
  • Relay nodes 11 la-n are communicatively coupled to other relay nodes 11 la-n and/or to edge nodes 114a/b.
  • Edge nodes 114a/b are communicatively coupled to relay nodes 11 la-n or network source or destination endpoints, such as the server computing environment 101 and client 104.
  • Relay nodes 11 la-n may correspond to data centers, network locations, routers, communications nexus, or other network components communicatively coupled to other relay nodes 11 la-n and edge nodes 114a/b via an overlay network approach.
  • relay nodes 11 la-n may be communicatively coupled to other relay nodes 11 la-n or edge nodes 114a/b using tunneling, including Transmission Control Protocol (TCP) over Internet Protocol (IP) /Universal Datagram Protocol (UDP) tunneling, UDP over IP/UDP tunneling, Secure Shell (SSH) tunneling, Virtual Private Networks (VPNs), or other approaches as can be appreciated.
  • TCP Transmission Control Protocol
  • IP Internet Protocol
  • UDP Universal Datagram Protocol
  • SSH Secure Shell
  • VPNs Virtual Private Networks
  • the edge nodes 114a/b may include dedicated networking devices, such as routers, switches, or other devices configured to perform the operations of edge nodes 114a/b as will be described below.
  • the edge nodes 114a/b may also include software, applications, services, or other functionality configured to perform the operations of edge nodes 114a/b and executed in one or more computing devices.
  • the edge nodes 114a/b are shown as distinct from the server computing environment 101 and client 104, it is understood that the edge nodes 114a/b may also include components or functionality executed within the server computing environment 101 or client 104.
  • a client 104 may be configured to execute an application facilitating the operations of an edge node 114b as can be appreciated.
  • the server computing environment 101 may comprise, for example, a server computer or any other system providing computing capability.
  • the server computing environment 101 may employ a plurality of computing devices that may be arranged, for example, in one or more server banks or computer banks or other arrangements. Such computing devices may be located in a single installation or may be distributed among many different geographical locations.
  • the server computing environment 101 may include a plurality of computing devices that together may comprise a hosted computing resource, a grid computing resource and/or any other distributed computing arrangement.
  • the server computing environment 101 may correspond to an elastic computing resource where the allotted capacity of processing, network, storage, or other computing- related resources may vary over time.
  • Various applications and/or other functionality may be executed in the server computing environment 101 according to various embodiments.
  • the components executed on the server computing environment 101 include a server application 117, and other applications, services, processes, systems, engines, or functionality not discussed in detail herein.
  • the server application 117 is executed to communicate baseline packets 121 of data to and receive baseline packets 121 of data from a client 104 via the overlay network 107.
  • the client 104 is representative of a plurality of client devices that may be coupled to the overlay network 107.
  • the client 104 may comprise, for example, a processor-based system such as a computer system.
  • a computer system may be embodied in the form of a desktop computer, a laptop computer, personal digital assistants, cellular telephones, smartphones, set-top boxes, music players, web pads, tablet computer systems, game consoles, electronic book readers, or other devices with like capability.
  • the client 104 may include a display.
  • the display may comprise, for example, one or more devices such as liquid crystal display (LCD) displays, gas plasma-based flat panel displays, organic light emitting diode (OLED) displays, electrophoretic ink (E ink) displays, LCD projectors, or other types of display devices, etc.
  • LCD liquid crystal display
  • OLED organic light emitting diode
  • E ink electrophoretic ink
  • the client 104 may be configured to execute various applications such as a client application 124 and/or other applications.
  • the client application 124 may be executed in a client 104, for example, to access network content served up by the server computing environment 101 and/or other servers, thereby rendering a user interface on the display.
  • the client application 124 may comprise, for example, a browser, a dedicated application, etc.
  • the user interface may comprise a network page, an application screen, etc.
  • the client 104 may be configured to execute applications beyond the client application 124 such as, for example, email applications, social networking applications, word processors, spreadsheets, and/or other applications.
  • the client application 124 may be configured, for example, to generate baseline packets 121 for communication to the server computing environment 101.
  • the client application 124 may also be configured to access data of baseline packets 121 received from the server computing environment 101 to perform its functionality.
  • a client application 124 generates one or more baseline packets 121 for communication to a server application 117 executed in the server computing environment 101.
  • the baseline packets 121 may be encoded, for example, for communication according to Transmission Control Protocol (TCP), Universal Datagram Protocol (UDP), Internet Protocol (IP), or another protocol.
  • TCP Transmission Control Protocol
  • UDP Universal Datagram Protocol
  • IP Internet Protocol
  • the baseline packets 121 are then obtained by an edge node 1 14b communicatively coupled to the client 104.
  • the edge node 1 14b may comprise a router, switch, server, or other dedicated network component distinct from the client 104.
  • obtaining the baseline packets 121 may include intercepting or otherwise capturing the baseline packets 121 as forwarded by the client 104.
  • the client 104 may be considered the edge node 1 14b.
  • the client 104 may be configured to execute an application or service to intercept or access the baseline packets 121 prior to forwarding to the overlay network 107.
  • the edge node 114b After obtaining the baseline packets 121, the edge node 114b generates overlay packets 127 from the baseline packets 121.
  • the baseline packets 121 are included in a data payload of the overlay packets 127.
  • generating the overlay packets 127 may include splitting the baseline packets 121 into smaller portions of data.
  • baseline packets 121 may be split such that the resulting overlay packets 127 conform to a Maximum Transmission Unit (MTU) size of a link in the overlay network 107, or by another approach.
  • MTU Maximum Transmission Unit
  • the split portions of baseline packets 121 may correspond to respective data payloads of overlay packets 127.
  • Generating the overlay packets 127 may also include generating metadata 131 for the respective overlay packets 127.
  • Metadata 131 may include, for example, sequence numbers corresponding to a sequence of the encoded baseline packets 121. Metadata 131 may also include routing data such as, for example, a network address of the client 104, an originating port of the client 104, a network address or port of the server application 1 17 or server computing environment 101 to which the baseline packets 121 will be communicated, a network address or port of an edge node 114a to which the overlay packets 127 will be communicated, and potentially other data.
  • Metadata 131 may also include, for example, authentication data, session identifiers, unique identifiers, or other data facilitating an authentication or identification of a client 104.
  • Generating overlay packets 127 may also include applying an encryption algorithm to all or a portion of the baseline packet 121 , overlay packet 127, or performing other operations.
  • the edge node 1 14b then forwards the overlay packets 127 to one or more relay nodes l l l a-n in the overlay network 107. In some embodiments, this may include performing a path finding operation applied to relay nodes l l la-n.
  • the edge nodes 1 14a/b and relay nodes l l la-n may be encoded or considered as a weighted graph.
  • the weights of the graph may be based on network performance data such as latency between linked nodes, capacity of relay nodes l l la-n, or other data.
  • the latency data, capacity data, or other data upon which a path is determined may be obtained by querying respective relay nodes 1 1 l a-n, obtained from a central data repository, monitoring service, data feed, or obtained by another approach.
  • the edge node 114b may then determine a plurality of paths of the graph according to a lowest combined latency of the path, a lowest combined burden of the path, or other criteria.
  • the edge node 1 14b After determining a plurality of paths, the edge node 1 14b then communicates the overlay packets 127 to a first relay node l l l a-n in each of the respective paths.
  • the relay nodes l l la-n then forward the overlay packets 127 to the edge node 1 14a via other relay nodes l l l a-n in the overlay network 107.
  • the edge node 1 14b may have been configured to encode an indication of a routing path in the overlay packets 127. In such an embodiment, the relay nodes l l l a-n would forward the overlay packets 127 according to the indicated routing path.
  • the relay nodes 11 la-n may be configured to determine an optimal path to the edge node 1 14a similar to the approaches described above. In such an embodiment, the relay nodes l l l a-n would then forward the overlay packets 127 to the next relay node 11 la-n in a determined path.
  • the edge node 1 14a Upon arrival at the edge node 114a, the edge node 1 14a generates the baseline packets 121 from the overlay packets 127. As the overlay packets 127 are sent along multiple paths in the overlay network 107, the edge node 1 14a may receive duplicate copies of one or more of the overlay packets 127. Accordingly, in some embodiments, this may include discarding duplicate overlay packets 127 as determined by a sequence number encoded in metadata 131 , by applying a hashing algorithm to the overlay packets 127, or by another approach.
  • Generating the baseline packets 121 from the overlay packets 127 may also include removing metadata 131, packet headers, or other data added to the baseline packets 121 by the edge node 114b to generate the overlay packets 127. In embodiments in which the baseline packets 121 were split into components of reduced size, this may also include combining or reassembling the baseline packets 121. The baseline packets 121 may also be regenerated from the overlay packets 127 by another approach. The edge node 114a then forwards the baseline packets 121 to the server application 117.
  • baseline packets 121 generated by a client application 124 for communication to a server application 1 17
  • baseline packets 121 may be similarly generated by a server application 1 17 for communication to the client application 124.
  • the edge node 114a would generate overlay packets 127 from the baseline packets 121 for forwarding to an edge node 114b via relay nodes 1 1 l a-n.
  • the networked environment 100b includes components similar to the networked environment 100a of FIG. 1A, including a server computing environment 101, a client 104, which are in data communication with each other via an overlay network 107. Such components may perform operations similar to those set forth with respect to the discussion of FIG. 1A. Additionally, the networked environment 100b includes a controller computing environment 134 communicatively coupled to the overlay network 107.
  • the controller computing environment 134 may comprise, for example, a server computer or any other system providing computing capability.
  • the controller computing environment 134 may employ a plurality of computing devices that may be arranged, for example, in one or more server banks or computer banks or other arrangements. Such computing devices may be located in a single installation or may be distributed among many different geographical locations.
  • the controller computing environment 134 may include a plurality of computing devices that together may comprise a hosted computing resource, a grid computing resource and/or any other distributed computing arrangement.
  • the controller computing environment 134 may correspond to an elastic computing resource where the allotted capacity of processing, network, storage, or other computing-related resources may vary over time.
  • Various applications and/or other functionality may be executed in the controller computing environment 134 according to various embodiments.
  • the components executed on the controller computing environment 134 include a controller application 137, and other applications, services, processes, systems, engines, or functionality not discussed in detail herein.
  • the controller application 137 is executed to query relay nodes l l l a-n for relay data 141 indicating an operational status of a respective relay node l l l a-n.
  • the relay data 141 may indicate, for example, a latency between the respective relay node l l l a-n and another relay node l l l a-n, edge node 114a/b, or other component of the overlay network 107.
  • the relay data 141 may also indicate a current pending workload or capacity of the respective relay node l l la-n, or other data. Using the relay data 141, the controller application 137 may then determine an optimal route or portion of a route between edge nodes 1 14a/b.
  • a client 104 communicates a request to the controller application 137 to establish a route between a corresponding edge node 1 14b and an edge node 114a of a server computing environment 101.
  • the controller application 137 queries one or more relay nodes 114a-n for their relay data 141.
  • the controller application 137 uses a path finding or graph search algorithm to generate a plurality of routes between the edge nodes 114a/b according to the relay data 141.
  • the routes may be generated to minimize a latency between the edge node 1 14a and 114b.
  • the routes may be generated to avoid or preferably avoid the use of relay nodes l l la-n having a pending workload meeting or exceeding a threshold, or a capacity meeting or falling below another threshold.
  • the controller application 137 communicates an indication of the routes to the edge node 1 14b, such that overlay packets 127 may be encoded with an indication of a respective route. This allows the overlay packets 127 to be communicated to relay nodes l l l a-n with an indication of a communications path through the overlay network 107.
  • the controller application 137 communicates, to the edge node 1 14b, an indication of one or more first relay nodes l l l a-n in the path.
  • the controller application 137 may also communicate to relay nodes l l la-n included in the routes an indication of a respective subsequent relay node l l la-n to which overlay packets 127 should be forwarded.
  • the edge node 1 14b and relay nodes l l l a-n are only instructed a next node to which overlay packets 127 should be forwarded.
  • the forwarding instructions to the relay nodes l l l a-n may be specific to overlay packets 127 to or from a particular client 104, client application 124, or destination edge node 1 14a.
  • relay nodes l l l a-n would select a next relay node 1 1 la-n according to the received instructions and a source or destination of the overlay packets 127.
  • the controller application 137 may repeatedly query the relay nodes l l l a-n for relay data 141 at a predefined interval, in response to a request, or according to other criteria. For example, an edge node 1 14a/b, relay node l l l a-n, or other component of the networked environment 100b may detect a network status such as a network component outage, a latency meeting or exceeding a threshold, or another event. The detecting component may then communicate a request to the controller application 137 to requery the relay nodes l l la-n for relay data 141. In such embodiments, the controller application 137 may then update the edge node 114b or relay nodes l l la-n with updated routes generated according to the updated relay data 141.
  • controller application 137 may also generate routes between an edge node 114b and 114a under a similar approach, thereby facilitating communications between a server application 117 and client application 127.
  • FIG. 2 shown is a flowchart that provides one example of the operation of a portion of the edge node 114b according to various embodiments. It is understood that the flowchart of FIG. 2 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of the edge node 114b as described herein. As an alternative, the flowchart of FIG. 2 may be viewed as depicting an example of elements of a method implemented in the edge node 114b according to one or more embodiments.
  • the edge node 114b receives baseline packets 121 (FIG. 1A) for communication to a network destination address via the overlay network 107 (FIG. 1A).
  • the baseline packets 121 may be received from a client application 124 (FIG. 1A) or other application executed in a client 104 (FIG. 1A) communicatively coupled to the edge node 114b.
  • the edge node 114b may serve as a router, switch, access point, modem, or other networking device in communication with the client 104.
  • the client 104 may serve as the edge node 114b, and the baseline packets 121 are received by a process or application distinct from a client application 124 or other application executed within the client 104.
  • the edge node 114b generates overlay packets 127 (FIG. 2) from the received baseline packets 121. In some embodiments, this includes generating the overlay packets 127 as having a data payload including a corresponding baseline packet 121. In other embodiments, the edge node 114b may split the baseline packets 121 into multiple data portions, with each data portion serving as a data payload for an overlay packet 127. In such an embodiment, the baseline packets 121 may be split such that the overlay packet 127 having the payload of the resulting data portions satisfy a Maximum Transmission Unit (MTU) size or other threshold. The MTU size may be defined according to a relay node l l la-n (FIG. 1A) or other component of the overlay network 107.
  • MTU Maximum Transmission Unit
  • Generating the overlay packets 127 may also include generating metadata 131 (FIG. 1A) for the overlay packets 127.
  • metadata 131 can include sequence numbers or other identifying information, integrity or verification data such as a hash value, cyclical redundancy check value, or other data.
  • the metadata 131 can also include login credentials or other identifying information corresponding to a client 104 or a user account.
  • the metadata 131 may also include indications of relay nodes l l la-n or paths of the overlay network 107 through which the packets should be communicated, as will be described in more detail below.
  • the metadata 131 can also include other data.
  • Generating the overlay packets may further include applying an encryption algorithm to the baseline packets 121 or other portions of the overlay packets 127.
  • the edge node 114b selects one or more relay nodes l l la-n to which the overlay packets 127 will be communicated. In some embodiments, this may include performing a path finding operation on relay nodes l l la-n to select a plurality of paths of the overlay network 107 leading to an edge node 114a (FIG. 1A).
  • the path finding operation may be performed based on a weighted graph of relay nodes l l la-n, with weights based on capacity or latency of relay nodes l l la-n, computational overhead, or other criteria. In such an embodiment, the edge node 114b may then select a plurality of paths based on the path finding operation.
  • the edge node 114b would then select the relay nodes l l la-n as the first relay node l l la-n for each of the selected paths.
  • the edge node 114b may encode an indication of a selected path for a corresponding duplicate set of overlay packets 127, indicating through which path the corresponding overlay packets 127 should be communicated.
  • the edge node 114b may select one or more relay nodes l l la-n based on the capacity, latency, nearness, or other criteria of communicatively adjacent relay nodes l l la-n. After selecting the relay nodes l l la-n, in box 211, the edge node 114b communicates a copy of the overlay packets 127 to each of the selected relay nodes 11 la-n, after which the process ends.
  • FIG. 3 shown is a flowchart that provides one example of the operation of a portion of the edge node 114a according to various embodiments. It is understood that the flowchart of FIG. 3 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of the edge node 114a as described herein. As an alternative, the flowchart of FIG. 3 may be viewed as depicting an example of elements of a method implemented in the edge node 114a according to one or more embodiments.
  • the edge node 114a receives overlay packets 127 (FIG. 1A) for communication to a network destination address via one or more relay nodes l l la-n (FIG. 1A) of the overlay network 107 (FIG. 1A).
  • the edge node 114a discards duplicate instances of the received overlay packets 127 in box 304. This may include, for example, tracking sequence numbers, hash values or other identifiers encoded in overlay packet 127 metadata 131.
  • a duplicate overlay packet 127 may be identified by determining if the identifier of an overlay packet 127 corresponds to a previously received overlay packet 127. Additionally, the edge node 114a may discard those packets failing a data verification, integrity or validation operation. The duplicate overlay packets 127 may then be deleted or otherwise discarded such that they are not forwarded to a destination address, as will be described below.
  • the edge node 114a After discarding duplicate overlay packets 127, the edge node 114a generates baseline packets 121 (FIG. 1A) from the retained overlay packets 127 in box 307. This may include applying a decryption algorithm to the overlay packets 127 or the pay loads of the overlay packets 127. In embodiments in which the baseline packets 127 were split for encoding into overlay packets 127, generating the baseline packets 127 may include reassembling the baseline packets 127 from the payloads of multiple overlay packets 127.
  • the edge node 114a communicates the baseline packets 127 to a destination network address.
  • the baseline packets 121 may be communicated to a server application 117 (FIG. 1A) or other application executed in a server computing environment 101 (FIG. 1A) communicatively coupled to the edge node 114a.
  • the edge node 114a may serve as a router, switch, access point, modem, or other networking device in communication with the server computing environment 101.
  • the server computing environment 101 may serve as the edge node 114a, and the baseline packets 121 are received by a process or application distinct from a server application 117 or other application executed within the server computing environment 101. After forwarding the baseline packets 121 to the destination network address, the process ends.
  • FIG. 4 shown is a flowchart that provides one example of the operation of a portion of the controller application 137 according to various embodiments. It is understood that the flowchart of FIG. 4 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the controller application 137 as described herein. As an alternative, the flowchart of FIG. 4 may be viewed as depicting an example of elements of a method implemented in the controller computing environment 134 (FIG. IB) according to one or more embodiments.
  • FIG. IB controller computing environment 134
  • the controller application 137 receives a request from a client 104 (FIG. IB) to generate a plurality of routes between an edge node 1 14b (FIG. IB) communicatively coupled to the client 104 and an edge node 1 14a (FIG. IB) via the overlay network 107 (FIG. IB).
  • the controller application 137 queries relay nodes l l la-n (FIG. IB) of the overlay network for relay data 141 (FIG. IB).
  • the controller application uses the relay data 141 , in box 407, the controller application generates routes between the edge node 1 14b and 114a.
  • the routes may be generated to minimize a latency between the edge node 1 14a and 1 14b. In other embodiments, the routes may be generated to avoid or preferably avoid the use of relay nodes l l l a-n having a pending workload meeting or exceeding a threshold, or a capacity meeting or falling below another threshold.
  • the controller application 137 communicates indications of the routes. In some embodiments, this includes communicating an indication of the routes to the edge node 1 14b, such that overlay packets 127 may be encoded with an indication of a respective route. This allows the overlay packets 127 (FIG. IB) to be communicated to relay nodes l l l a-n with an indication of a communications path through the overlay network 107.
  • this may include communicating, to the edge node 1 14b, an indication of one or more first relay nodes l l la-n in the path.
  • the controller application 137 may also communicate to relay nodes 11 la-n included in the routes an indication of a respective subsequent relay node l l l a-n to which overlay packets 127 should be forwarded.
  • the edge node 114b and relay nodes l l l a-n are only instructed of a next node to which overlay packets 127 should be forwarded.
  • the forwarding instructions to the relay nodes l l l a-n may be specific to overlay packets 127 to or from a particular client 104, client application 124, or destination edge node 114a.
  • relay nodes l l l a-n would select the next relay node l l l a-n according to the received instructions and a source or destination of the overlay packets 127.
  • the server computing environment 101, client 104, and controller computing environment each include one or more computing devices 501a, 501b or 501c, respectively.
  • Each computing device 501a/b/c includes at least one processor circuit, for example, having a processor 502a/b/c and a memory 504a/b/c, both of which are coupled to a local interface 507a/b/c.
  • each computing device 501a/b/c may comprise, for example, at least one server computer or like device.
  • the local interface 507a/b/c may comprise, for example, a data bus with an accompanying address/control bus or other bus structure as can be appreciated.
  • Stored in the memory 504a/b/c are both data and several components that are executable by the processor 502a/b/c.
  • stored in the memory 504a/b/c and executable by the processor 502a/b/c are server application 117, client application 124 or controller application 137, respectively, and potentially other applications.
  • Also stored in the memory 504a/b/c may be a data store and other data.
  • an operating system may be stored in the memory 504a/b/c and executable by the processor 502a/b/c.
  • a number of software components are stored in the memory 504a/b/c and are executable by the processor 502a/b/c.
  • executable means a program file that is in a form that can ultimately be run by the processor 502a/b/c.
  • Examples of executable programs may be, for example, a compiled program that can be translated into machine code in a format that can be loaded into a random access portion of the memory 504a/b/c and run by the processor 502a/b/c, source code that may be expressed in proper format such as object code that is capable of being loaded into a random access portion of the memory 504a/b/c and executed by the processor 502a/b/c, or source code that may be interpreted by another executable program to generate instructions in a random access portion of the memory 504a/b/c to be executed by the processor 502a/b/c, etc.
  • An executable program may be stored in any portion or component of the memory 504a/b/c including, for example, random access memory (RAM), read-only memory (ROM), hard drive, solid-state drive, USB flash drive, memory card, optical disc such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
  • RAM random access memory
  • ROM read-only memory
  • hard drive solid-state drive
  • USB flash drive Universal Serial Bus flash drive
  • memory card such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
  • CD compact disc
  • DVD digital versatile disc
  • the memory 504a/b/c is defined herein as including both volatile and nonvolatile memory and data storage components. Volatile components are those that do not retain data values upon loss of power. Nonvolatile components are those that retain data upon a loss of power.
  • the memory 504a/b/c may comprise, for example, random access memory (RAM), read-only memory (ROM), hard disk drives, solid-state drives, USB flash drives, memory cards accessed via a memory card reader, floppy disks accessed via an associated floppy disk drive, optical discs accessed via an optical disc drive, magnetic tapes accessed via an appropriate tape drive, and/or other memory components, or a combination of any two or more of these memory components.
  • the RAM may comprise, for example, static random access memory (SRAM), dynamic random access memory (DRAM), or magnetic random access memory (MRAM) and other such devices.
  • the ROM may comprise, for example, a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other like memory device.
  • the processor 502a/b/c may represent multiple processors 502a/b/c and/or multiple processor cores and the memory 504a/b/c may represent multiple memories 504a/b/c that operate in parallel processing circuits, respectively.
  • the local interface 507a/b/c may be an appropriate network that facilitates communication between any two of the multiple processors 502a/b/c, between any processor 502a/b/c and any of the memories 504a/b/c, or between any two of the memories 504a/b/c, etc.
  • the local interface 507a/b/c may comprise additional systems designed to coordinate this communication, including, for example, performing load balancing.
  • the processor 502a/b/c may be of electrical or of some other available construction.
  • each block may represent a module, segment, or portion of code that comprises program instructions to implement the specified logical function(s).
  • the program instructions may be embodied in the form of source code that comprises human- readable statements written in a programming language or machine code that comprises numerical instructions recognizable by a suitable execution system such as a processor in a computer system or other system.
  • the machine code may be converted from the source code, etc.
  • each block may represent a circuit or a number of interconnected circuits to implement the specified logical function(s).
  • FIGS. 2, 3 and 4 show a specific order of execution, it is understood that the order of execution may differ from that which is depicted. For example, the order of execution of two or more blocks may be scrambled relative to the order shown. Also, two or more blocks shown in succession in FIGS. 2, 3 and 4 may be executed concurrently or with partial concurrence. Further, in some embodiments, one or more of the blocks shown in FIGS. 2, 3 and 4 may be skipped or omitted. In addition, any number of counters, state variables, warning semaphores, or messages might be added to the logical flow described herein, for purposes of enhanced utility, accounting, performance measurement, or providing troubleshooting aids, etc. It is understood that all such variations are within the scope of the present disclosure.
  • any logic or application described herein that comprises software or code can be embodied in any non-transitory computer-readable medium for use by or in connection with an instruction execution system such as, for example, a processor 502a/b/c in a computer system or other system.
  • the logic may comprise, for example, statements including instructions and declarations that can be fetched from the computer-readable medium and executed by the instruction execution system.
  • a "computer-readable medium" can be any medium that can contain, store, or maintain the logic or application described herein for use by or in connection with the instruction execution system.
  • the computer-readable medium can comprise any one of many physical media such as, for example, magnetic, optical, or semiconductor media.
  • a suitable computer-readable medium would include, but are not limited to, magnetic tapes, magnetic floppy diskettes, magnetic hard drives, memory cards, solid-state drives, USB flash drives, or optical discs.
  • the computer-readable medium may be a random access memory (RAM) including, for example, static random access memory (SRAM) and dynamic random access memory (DRAM), or magnetic random access memory (MRAM).
  • the computer-readable medium may be a read-only memory (ROM), a programmable readonly memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other type of memory device.
  • any logic or application described herein may be implemented and structured in a variety of ways.
  • one or more applications described may be implemented as modules or components of a single application.
  • one or more applications described herein may be executed in shared or separate computing devices or a combination thereof.
  • a plurality of the applications described herein may execute in the same computing device 501 a/b/c, or in multiple computing devices in the same server computing environment 101 , client 104 or controller computing environment 134.
  • terms such as “application,” “service,” “system,” “engine,” “module,” and so on may be interchangeable and are not intended to be limiting.
  • Disjunctive language such as the phrase "at least one of X, Y, or Z," unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y, or Z, or any combination thereof (e.g. , X, Y, and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y, or at least one of Z to each be present.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

Disclosed are various embodiments for a parallel multipath routing architecture. An edge node of an overlay network receives packets to be communicated to a destination network address. The received packets are encoded for communication through multiple paths of the overlay network leading to a second edge node communicatively coupled to the destination network address. The packets are duplicated across the multiple paths to minimize the likelihood of packet loss. The second edge node discards duplicated instances of the packets and regenerates the original packets for forwarding to the destination network address.

Description

PARALLEL MULTIPATH ROUTING ARCHITECTURE
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims priority to U.S. Application Serial No. 62/325,134, filed April 20, 2016, which is incorporated herein by reference in its entirety.
BACKGROUND
[0002] Consumers of internet access may participate in activities that can be negatively impacted by packet loss. For example, players of networked video games, stock traders, and participants in activities where time or speed is essential may be negatively affected due to lost packets that must be resent.
BRIEF DESCRIPTION OF THE DRAWINGS
[0003] Many aspects of the present disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, with emphasis instead being placed upon clearly illustrating the principles of the disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views.
[0004] FIG. 1A is a drawing of a networked environment according to various embodiments of the present disclosure.
[0005] FIG. IB is a drawing of a networked environment according to various embodiments of the present disclosure.
[0006] FIG. 2 is a flowchart illustrating one example of functionality implemented as portions of an edge node 114b in the networked environment of FIGS. 1A or FIG. IB according to various embodiments of the present disclosure.
[0007] FIG. 3 is a flowchart illustrating one example of functionality implemented as portions of an edge node 114a in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
[0008] FIG. 4 is a flowchart illustrating one example of functionality implemented as portions of a controller application 137 executed in a controller computing environment in the networked environment of FIG. IB according to various embodiments of the present disclosure. [0009] FIG. 5A is a schematic block diagram that provides one example illustration of a server computing environment employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
[0010] FIG. 5B is a schematic block diagram that provides one example illustration of a client employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure
[0011] FIG. 5C is a schematic block diagram that provides one example illustration of a controller computing environment employed in the networked environment of FIG. 1A or FIG. IB according to various embodiments of the present disclosure.
DETAILED DESCRIPTION
[0012] Internet access consumers may experience performance degradation due to packet loss. For example, a player in a networked video game may experience packet loss between the player and a server. Additional resources must be used in order to resend a lost packet between the player and server, which may introduce lag, speed degradation, and other issues that may affect many players. As another example, stock traders communicating with a trading system may experience delays in consummating transactions due to additional resource expenditure in resending lost packets.
[0013] A parallel multipath architecture implements an overlay network on existing Internet Protocol network frameworks. The overlay network includes edge nodes communicatively coupled to communication endpoints. Relay nodes are communicatively coupled to the edge nodes and other relay nodes. Packets from a source communication endpoint are modified at an edge node to include metadata including sequence numbers, destination network address information, or other data. The edge node forwards the packets along multiple parallel paths of relay nodes within the overlay network. An edge node communicatively coupled to the destination communication endpoint removes the added metadata, discards duplicate packets, and forwards the sequenced payloads to the destination. By implementing multiple parallel paths between a source and destination, packet loss is reduced as a packet must be lost along each path in which it is sent in order to be lost. In the following discussion, a general description of the system and its components is provided, followed by a discussion of the operation of the same.
[0014] With reference to FIG. 1A, shown is a networked environment 100a according to various embodiments. The networked environment 100a includes a server computing environment 101, a client 104, which are in data communication with each other via an overlay network 107. The overlay network 107 includes, for example, the Internet, wired networks, wireless networks, or other suitable networks, etc., or any combination of two or more such networks. For example, such networks may comprise satellite networks, cable networks, Ethernet networks, and other types of networks.
[0015] To this end, the overlay network 107 comprises one or more relay nodes 11 la-n and edge nodes 114a/b, which can include a subset of network components or nodes of a network upon which the overlay network 107 lies. Relay nodes 11 la-n are communicatively coupled to other relay nodes 11 la-n and/or to edge nodes 114a/b. Edge nodes 114a/b are communicatively coupled to relay nodes 11 la-n or network source or destination endpoints, such as the server computing environment 101 and client 104. Relay nodes 11 la-n may correspond to data centers, network locations, routers, communications nexus, or other network components communicatively coupled to other relay nodes 11 la-n and edge nodes 114a/b via an overlay network approach. For example, relay nodes 11 la-n may be communicatively coupled to other relay nodes 11 la-n or edge nodes 114a/b using tunneling, including Transmission Control Protocol (TCP) over Internet Protocol (IP) /Universal Datagram Protocol (UDP) tunneling, UDP over IP/UDP tunneling, Secure Shell (SSH) tunneling, Virtual Private Networks (VPNs), or other approaches as can be appreciated.
[0016] The edge nodes 114a/b may include dedicated networking devices, such as routers, switches, or other devices configured to perform the operations of edge nodes 114a/b as will be described below. The edge nodes 114a/b may also include software, applications, services, or other functionality configured to perform the operations of edge nodes 114a/b and executed in one or more computing devices. Although, in this example embodiment, the edge nodes 114a/b are shown as distinct from the server computing environment 101 and client 104, it is understood that the edge nodes 114a/b may also include components or functionality executed within the server computing environment 101 or client 104. As a non- limiting example, a client 104 may be configured to execute an application facilitating the operations of an edge node 114b as can be appreciated.
[0017] The server computing environment 101 may comprise, for example, a server computer or any other system providing computing capability. Alternatively, the server computing environment 101 may employ a plurality of computing devices that may be arranged, for example, in one or more server banks or computer banks or other arrangements. Such computing devices may be located in a single installation or may be distributed among many different geographical locations. For example, the server computing environment 101 may include a plurality of computing devices that together may comprise a hosted computing resource, a grid computing resource and/or any other distributed computing arrangement. In some cases, the server computing environment 101 may correspond to an elastic computing resource where the allotted capacity of processing, network, storage, or other computing- related resources may vary over time.
[0018] Various applications and/or other functionality may be executed in the server computing environment 101 according to various embodiments. The components executed on the server computing environment 101, for example, include a server application 117, and other applications, services, processes, systems, engines, or functionality not discussed in detail herein. The server application 117 is executed to communicate baseline packets 121 of data to and receive baseline packets 121 of data from a client 104 via the overlay network 107.
[0019] The client 104 is representative of a plurality of client devices that may be coupled to the overlay network 107. The client 104 may comprise, for example, a processor-based system such as a computer system. Such a computer system may be embodied in the form of a desktop computer, a laptop computer, personal digital assistants, cellular telephones, smartphones, set-top boxes, music players, web pads, tablet computer systems, game consoles, electronic book readers, or other devices with like capability. The client 104 may include a display. The display may comprise, for example, one or more devices such as liquid crystal display (LCD) displays, gas plasma-based flat panel displays, organic light emitting diode (OLED) displays, electrophoretic ink (E ink) displays, LCD projectors, or other types of display devices, etc.
[0020] The client 104 may be configured to execute various applications such as a client application 124 and/or other applications. The client application 124 may be executed in a client 104, for example, to access network content served up by the server computing environment 101 and/or other servers, thereby rendering a user interface on the display. To this end, the client application 124 may comprise, for example, a browser, a dedicated application, etc. , and the user interface may comprise a network page, an application screen, etc. The client 104 may be configured to execute applications beyond the client application 124 such as, for example, email applications, social networking applications, word processors, spreadsheets, and/or other applications. The client application 124 may be configured, for example, to generate baseline packets 121 for communication to the server computing environment 101. The client application 124 may also be configured to access data of baseline packets 121 received from the server computing environment 101 to perform its functionality. [0021] Next, a general description of the operation of the various components of the networked environment 100a is provided. To begin, a client application 124 generates one or more baseline packets 121 for communication to a server application 117 executed in the server computing environment 101. The baseline packets 121 may be encoded, for example, for communication according to Transmission Control Protocol (TCP), Universal Datagram Protocol (UDP), Internet Protocol (IP), or another protocol. The baseline packets 121 are then obtained by an edge node 1 14b communicatively coupled to the client 104.
[0022] In some embodiments, the edge node 1 14b may comprise a router, switch, server, or other dedicated network component distinct from the client 104. In such an embodiment, obtaining the baseline packets 121 may include intercepting or otherwise capturing the baseline packets 121 as forwarded by the client 104. In other embodiments, the client 104 may be considered the edge node 1 14b. In such an embodiment, the client 104 may be configured to execute an application or service to intercept or access the baseline packets 121 prior to forwarding to the overlay network 107.
[0023] After obtaining the baseline packets 121, the edge node 114b generates overlay packets 127 from the baseline packets 121. In some embodiments, the baseline packets 121 are included in a data payload of the overlay packets 127. In other embodiments, generating the overlay packets 127 may include splitting the baseline packets 121 into smaller portions of data. For example, baseline packets 121 may be split such that the resulting overlay packets 127 conform to a Maximum Transmission Unit (MTU) size of a link in the overlay network 107, or by another approach. In such an embodiment, the split portions of baseline packets 121 may correspond to respective data payloads of overlay packets 127.
[0024] Generating the overlay packets 127 may also include generating metadata 131 for the respective overlay packets 127. Metadata 131 may include, for example, sequence numbers corresponding to a sequence of the encoded baseline packets 121. Metadata 131 may also include routing data such as, for example, a network address of the client 104, an originating port of the client 104, a network address or port of the server application 1 17 or server computing environment 101 to which the baseline packets 121 will be communicated, a network address or port of an edge node 114a to which the overlay packets 127 will be communicated, and potentially other data. Metadata 131 may also include, for example, authentication data, session identifiers, unique identifiers, or other data facilitating an authentication or identification of a client 104. Generating overlay packets 127 may also include applying an encryption algorithm to all or a portion of the baseline packet 121 , overlay packet 127, or performing other operations. [0025] The edge node 1 14b then forwards the overlay packets 127 to one or more relay nodes l l l a-n in the overlay network 107. In some embodiments, this may include performing a path finding operation applied to relay nodes l l la-n. For example, the edge nodes 1 14a/b and relay nodes l l la-n may be encoded or considered as a weighted graph. The weights of the graph may be based on network performance data such as latency between linked nodes, capacity of relay nodes l l la-n, or other data. The latency data, capacity data, or other data upon which a path is determined may be obtained by querying respective relay nodes 1 1 l a-n, obtained from a central data repository, monitoring service, data feed, or obtained by another approach. The edge node 114b may then determine a plurality of paths of the graph according to a lowest combined latency of the path, a lowest combined burden of the path, or other criteria.
[0026] After determining a plurality of paths, the edge node 1 14b then communicates the overlay packets 127 to a first relay node l l l a-n in each of the respective paths. The relay nodes l l la-n then forward the overlay packets 127 to the edge node 1 14a via other relay nodes l l l a-n in the overlay network 107. In some embodiments, the edge node 1 14b may have been configured to encode an indication of a routing path in the overlay packets 127. In such an embodiment, the relay nodes l l l a-n would forward the overlay packets 127 according to the indicated routing path. In other embodiments, the relay nodes 11 la-n may be configured to determine an optimal path to the edge node 1 14a similar to the approaches described above. In such an embodiment, the relay nodes l l l a-n would then forward the overlay packets 127 to the next relay node 11 la-n in a determined path.
[0027] Upon arrival at the edge node 114a, the edge node 1 14a generates the baseline packets 121 from the overlay packets 127. As the overlay packets 127 are sent along multiple paths in the overlay network 107, the edge node 1 14a may receive duplicate copies of one or more of the overlay packets 127. Accordingly, in some embodiments, this may include discarding duplicate overlay packets 127 as determined by a sequence number encoded in metadata 131 , by applying a hashing algorithm to the overlay packets 127, or by another approach.
[0028] Generating the baseline packets 121 from the overlay packets 127 may also include removing metadata 131, packet headers, or other data added to the baseline packets 121 by the edge node 114b to generate the overlay packets 127. In embodiments in which the baseline packets 121 were split into components of reduced size, this may also include combining or reassembling the baseline packets 121. The baseline packets 121 may also be regenerated from the overlay packets 127 by another approach. The edge node 114a then forwards the baseline packets 121 to the server application 117.
[0029] Although the operation described above discusses baseline packets 121 generated by a client application 124 for communication to a server application 1 17, it is understood that baseline packets 121 may be similarly generated by a server application 1 17 for communication to the client application 124. In such an embodiment, the edge node 114a would generate overlay packets 127 from the baseline packets 121 for forwarding to an edge node 114b via relay nodes 1 1 l a-n.
[0030] With reference to FIG. IB, shown is a networked environment 100b according to various embodiments. The networked environment 100b includes components similar to the networked environment 100a of FIG. 1A, including a server computing environment 101, a client 104, which are in data communication with each other via an overlay network 107. Such components may perform operations similar to those set forth with respect to the discussion of FIG. 1A. Additionally, the networked environment 100b includes a controller computing environment 134 communicatively coupled to the overlay network 107.
[0031] The controller computing environment 134 may comprise, for example, a server computer or any other system providing computing capability. Alternatively, the controller computing environment 134 may employ a plurality of computing devices that may be arranged, for example, in one or more server banks or computer banks or other arrangements. Such computing devices may be located in a single installation or may be distributed among many different geographical locations. For example, the controller computing environment 134 may include a plurality of computing devices that together may comprise a hosted computing resource, a grid computing resource and/or any other distributed computing arrangement. In some cases, the controller computing environment 134 may correspond to an elastic computing resource where the allotted capacity of processing, network, storage, or other computing-related resources may vary over time.
[0032] Various applications and/or other functionality may be executed in the controller computing environment 134 according to various embodiments. The components executed on the controller computing environment 134, for example, include a controller application 137, and other applications, services, processes, systems, engines, or functionality not discussed in detail herein. The controller application 137 is executed to query relay nodes l l l a-n for relay data 141 indicating an operational status of a respective relay node l l l a-n. The relay data 141 may indicate, for example, a latency between the respective relay node l l l a-n and another relay node l l l a-n, edge node 114a/b, or other component of the overlay network 107. The relay data 141 may also indicate a current pending workload or capacity of the respective relay node l l la-n, or other data. Using the relay data 141, the controller application 137 may then determine an optimal route or portion of a route between edge nodes 1 14a/b.
[0033] Next, a general description of the operation of the various components of the networked environment 100b is provided. To begin, a client 104 communicates a request to the controller application 137 to establish a route between a corresponding edge node 1 14b and an edge node 114a of a server computing environment 101. In response to the request, the controller application 137 queries one or more relay nodes 114a-n for their relay data 141. Using the relay data 141 , the controller application 137 uses a path finding or graph search algorithm to generate a plurality of routes between the edge nodes 114a/b according to the relay data 141. In some embodiments, the routes may be generated to minimize a latency between the edge node 1 14a and 114b. In other embodiments, the routes may be generated to avoid or preferably avoid the use of relay nodes l l la-n having a pending workload meeting or exceeding a threshold, or a capacity meeting or falling below another threshold.
[0034] After generating the routes, in some embodiments, the controller application 137 communicates an indication of the routes to the edge node 1 14b, such that overlay packets 127 may be encoded with an indication of a respective route. This allows the overlay packets 127 to be communicated to relay nodes l l l a-n with an indication of a communications path through the overlay network 107.
[0035] In other embodiments, the controller application 137 communicates, to the edge node 1 14b, an indication of one or more first relay nodes l l l a-n in the path. In such an embodiment, the controller application 137 may also communicate to relay nodes l l la-n included in the routes an indication of a respective subsequent relay node l l la-n to which overlay packets 127 should be forwarded. Thus, the edge node 1 14b and relay nodes l l l a-n are only instructed a next node to which overlay packets 127 should be forwarded. In some embodiments, the forwarding instructions to the relay nodes l l l a-n may be specific to overlay packets 127 to or from a particular client 104, client application 124, or destination edge node 1 14a. Thus, relay nodes l l l a-n would select a next relay node 1 1 la-n according to the received instructions and a source or destination of the overlay packets 127.
[0036] In further embodiments, the controller application 137 may repeatedly query the relay nodes l l l a-n for relay data 141 at a predefined interval, in response to a request, or according to other criteria. For example, an edge node 1 14a/b, relay node l l l a-n, or other component of the networked environment 100b may detect a network status such as a network component outage, a latency meeting or exceeding a threshold, or another event. The detecting component may then communicate a request to the controller application 137 to requery the relay nodes l l la-n for relay data 141. In such embodiments, the controller application 137 may then update the edge node 114b or relay nodes l l la-n with updated routes generated according to the updated relay data 141.
[0037] Although the operation described above discusses a controller application 137 generating routes between an edge node 114b and edge node 114a, it is understood that the controller application 137 may also generate routes between an edge node 114b and 114a under a similar approach, thereby facilitating communications between a server application 117 and client application 127.
[0038] Referring next to FIG. 2, shown is a flowchart that provides one example of the operation of a portion of the edge node 114b according to various embodiments. It is understood that the flowchart of FIG. 2 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of the edge node 114b as described herein. As an alternative, the flowchart of FIG. 2 may be viewed as depicting an example of elements of a method implemented in the edge node 114b according to one or more embodiments.
[0039] Beginning with box 201, the edge node 114b receives baseline packets 121 (FIG. 1A) for communication to a network destination address via the overlay network 107 (FIG. 1A). In some embodiments, the baseline packets 121 may be received from a client application 124 (FIG. 1A) or other application executed in a client 104 (FIG. 1A) communicatively coupled to the edge node 114b. For example, the edge node 114b may serve as a router, switch, access point, modem, or other networking device in communication with the client 104. In other embodiments, the client 104 may serve as the edge node 114b, and the baseline packets 121 are received by a process or application distinct from a client application 124 or other application executed within the client 104.
[0040] Next, in box 204, the edge node 114b generates overlay packets 127 (FIG. 2) from the received baseline packets 121. In some embodiments, this includes generating the overlay packets 127 as having a data payload including a corresponding baseline packet 121. In other embodiments, the edge node 114b may split the baseline packets 121 into multiple data portions, with each data portion serving as a data payload for an overlay packet 127. In such an embodiment, the baseline packets 121 may be split such that the overlay packet 127 having the payload of the resulting data portions satisfy a Maximum Transmission Unit (MTU) size or other threshold. The MTU size may be defined according to a relay node l l la-n (FIG. 1A) or other component of the overlay network 107.
[0041] Generating the overlay packets 127 may also include generating metadata 131 (FIG. 1A) for the overlay packets 127. Such metadata 131 can include sequence numbers or other identifying information, integrity or verification data such as a hash value, cyclical redundancy check value, or other data. The metadata 131 can also include login credentials or other identifying information corresponding to a client 104 or a user account. The metadata 131 may also include indications of relay nodes l l la-n or paths of the overlay network 107 through which the packets should be communicated, as will be described in more detail below. The metadata 131 can also include other data. Generating the overlay packets may further include applying an encryption algorithm to the baseline packets 121 or other portions of the overlay packets 127.
[0042] Next, in box 207, the edge node 114b selects one or more relay nodes l l la-n to which the overlay packets 127 will be communicated. In some embodiments, this may include performing a path finding operation on relay nodes l l la-n to select a plurality of paths of the overlay network 107 leading to an edge node 114a (FIG. 1A). The path finding operation may be performed based on a weighted graph of relay nodes l l la-n, with weights based on capacity or latency of relay nodes l l la-n, computational overhead, or other criteria. In such an embodiment, the edge node 114b may then select a plurality of paths based on the path finding operation. The edge node 114b would then select the relay nodes l l la-n as the first relay node l l la-n for each of the selected paths. In such an embodiment, the edge node 114b may encode an indication of a selected path for a corresponding duplicate set of overlay packets 127, indicating through which path the corresponding overlay packets 127 should be communicated. In other embodiments, the edge node 114b may select one or more relay nodes l l la-n based on the capacity, latency, nearness, or other criteria of communicatively adjacent relay nodes l l la-n. After selecting the relay nodes l l la-n, in box 211, the edge node 114b communicates a copy of the overlay packets 127 to each of the selected relay nodes 11 la-n, after which the process ends.
[0043] Moving on to FIG. 3, shown is a flowchart that provides one example of the operation of a portion of the edge node 114a according to various embodiments. It is understood that the flowchart of FIG. 3 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the portion of the edge node 114a as described herein. As an alternative, the flowchart of FIG. 3 may be viewed as depicting an example of elements of a method implemented in the edge node 114a according to one or more embodiments.
[0044] Beginning with box 301, the edge node 114a receives overlay packets 127 (FIG. 1A) for communication to a network destination address via one or more relay nodes l l la-n (FIG. 1A) of the overlay network 107 (FIG. 1A). As multiple copies of a given overlay packet 127 have been communicated via multiple paths of the overlay network 107 as described above, the edge node 114a discards duplicate instances of the received overlay packets 127 in box 304. This may include, for example, tracking sequence numbers, hash values or other identifiers encoded in overlay packet 127 metadata 131. A duplicate overlay packet 127 may be identified by determining if the identifier of an overlay packet 127 corresponds to a previously received overlay packet 127. Additionally, the edge node 114a may discard those packets failing a data verification, integrity or validation operation. The duplicate overlay packets 127 may then be deleted or otherwise discarded such that they are not forwarded to a destination address, as will be described below.
[0045] After discarding duplicate overlay packets 127, the edge node 114a generates baseline packets 121 (FIG. 1A) from the retained overlay packets 127 in box 307. This may include applying a decryption algorithm to the overlay packets 127 or the pay loads of the overlay packets 127. In embodiments in which the baseline packets 127 were split for encoding into overlay packets 127, generating the baseline packets 127 may include reassembling the baseline packets 127 from the payloads of multiple overlay packets 127.
[0046] Next, in box 311, the edge node 114a communicates the baseline packets 127 to a destination network address. In some embodiments, the baseline packets 121 may be communicated to a server application 117 (FIG. 1A) or other application executed in a server computing environment 101 (FIG. 1A) communicatively coupled to the edge node 114a. For example, the edge node 114a may serve as a router, switch, access point, modem, or other networking device in communication with the server computing environment 101. In other embodiments, the server computing environment 101 may serve as the edge node 114a, and the baseline packets 121 are received by a process or application distinct from a server application 117 or other application executed within the server computing environment 101. After forwarding the baseline packets 121 to the destination network address, the process ends.
[0047] Turning now to FIG. 4, shown is a flowchart that provides one example of the operation of a portion of the controller application 137 according to various embodiments. It is understood that the flowchart of FIG. 4 provides merely an example of the many different types of functional arrangements that may be employed to implement the operation of the controller application 137 as described herein. As an alternative, the flowchart of FIG. 4 may be viewed as depicting an example of elements of a method implemented in the controller computing environment 134 (FIG. IB) according to one or more embodiments.
[0048] Beginning with box 401, the controller application 137 receives a request from a client 104 (FIG. IB) to generate a plurality of routes between an edge node 1 14b (FIG. IB) communicatively coupled to the client 104 and an edge node 1 14a (FIG. IB) via the overlay network 107 (FIG. IB). Next, in box 404, the controller application 137 queries relay nodes l l la-n (FIG. IB) of the overlay network for relay data 141 (FIG. IB). Using the relay data 141 , in box 407, the controller application generates routes between the edge node 1 14b and 114a.
[0049] In some embodiments, the routes may be generated to minimize a latency between the edge node 1 14a and 1 14b. In other embodiments, the routes may be generated to avoid or preferably avoid the use of relay nodes l l l a-n having a pending workload meeting or exceeding a threshold, or a capacity meeting or falling below another threshold.
[0050] After generating the routes, in box 41 1, the controller application 137 communicates indications of the routes. In some embodiments, this includes communicating an indication of the routes to the edge node 1 14b, such that overlay packets 127 may be encoded with an indication of a respective route. This allows the overlay packets 127 (FIG. IB) to be communicated to relay nodes l l l a-n with an indication of a communications path through the overlay network 107.
[0051] In other embodiments this may include communicating, to the edge node 1 14b, an indication of one or more first relay nodes l l la-n in the path. In such an embodiment, the controller application 137 may also communicate to relay nodes 11 la-n included in the routes an indication of a respective subsequent relay node l l l a-n to which overlay packets 127 should be forwarded. Thus, the edge node 114b and relay nodes l l l a-n are only instructed of a next node to which overlay packets 127 should be forwarded. In some embodiments, the forwarding instructions to the relay nodes l l l a-n may be specific to overlay packets 127 to or from a particular client 104, client application 124, or destination edge node 114a. Thus, relay nodes l l l a-n would select the next relay node l l l a-n according to the received instructions and a source or destination of the overlay packets 127.
[0052] After communicating the indications of routes, the process ends. However, it is understood that the operations of boxes 404, 407 and 411 may be repeated to receive updated relay data 141 for the relay nodes l l la-n, and then indicate updated routes to the relay nodes l l la-n.
[0053] With reference to FIGS. 5A, 5B and 5C, shown are schematic block diagrams of the server computing environment 101, client 104 and controller computing environment 134 according to an embodiment of the present disclosure. The server computing environment 101, client 104, and controller computing environment each include one or more computing devices 501a, 501b or 501c, respectively. Each computing device 501a/b/c includes at least one processor circuit, for example, having a processor 502a/b/c and a memory 504a/b/c, both of which are coupled to a local interface 507a/b/c. To this end, each computing device 501a/b/c may comprise, for example, at least one server computer or like device. The local interface 507a/b/c may comprise, for example, a data bus with an accompanying address/control bus or other bus structure as can be appreciated.
[0054] Stored in the memory 504a/b/c are both data and several components that are executable by the processor 502a/b/c. In particular, stored in the memory 504a/b/c and executable by the processor 502a/b/c are server application 117, client application 124 or controller application 137, respectively, and potentially other applications. Also stored in the memory 504a/b/c may be a data store and other data. In addition, an operating system may be stored in the memory 504a/b/c and executable by the processor 502a/b/c.
[0055] It is understood that there may be other applications that are stored in the memory 504a/b/c and are executable by the processor 502a/b/c as can be appreciated. Where any component discussed herein is implemented in the form of software, any one of a number of programming languages may be employed such as, for example, C, C++, C#, Objective C, Java®, JavaScript®, Perl, PHP, Visual Basic®, Python®, Ruby, Flash®, or other programming languages.
[0056] A number of software components are stored in the memory 504a/b/c and are executable by the processor 502a/b/c. In this respect, the term "executable" means a program file that is in a form that can ultimately be run by the processor 502a/b/c. Examples of executable programs may be, for example, a compiled program that can be translated into machine code in a format that can be loaded into a random access portion of the memory 504a/b/c and run by the processor 502a/b/c, source code that may be expressed in proper format such as object code that is capable of being loaded into a random access portion of the memory 504a/b/c and executed by the processor 502a/b/c, or source code that may be interpreted by another executable program to generate instructions in a random access portion of the memory 504a/b/c to be executed by the processor 502a/b/c, etc. An executable program may be stored in any portion or component of the memory 504a/b/c including, for example, random access memory (RAM), read-only memory (ROM), hard drive, solid-state drive, USB flash drive, memory card, optical disc such as compact disc (CD) or digital versatile disc (DVD), floppy disk, magnetic tape, or other memory components.
[0057] The memory 504a/b/c is defined herein as including both volatile and nonvolatile memory and data storage components. Volatile components are those that do not retain data values upon loss of power. Nonvolatile components are those that retain data upon a loss of power. Thus, the memory 504a/b/c may comprise, for example, random access memory (RAM), read-only memory (ROM), hard disk drives, solid-state drives, USB flash drives, memory cards accessed via a memory card reader, floppy disks accessed via an associated floppy disk drive, optical discs accessed via an optical disc drive, magnetic tapes accessed via an appropriate tape drive, and/or other memory components, or a combination of any two or more of these memory components. In addition, the RAM may comprise, for example, static random access memory (SRAM), dynamic random access memory (DRAM), or magnetic random access memory (MRAM) and other such devices. The ROM may comprise, for example, a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other like memory device.
[0058] Also, the processor 502a/b/c may represent multiple processors 502a/b/c and/or multiple processor cores and the memory 504a/b/c may represent multiple memories 504a/b/c that operate in parallel processing circuits, respectively. In such a case, the local interface 507a/b/c may be an appropriate network that facilitates communication between any two of the multiple processors 502a/b/c, between any processor 502a/b/c and any of the memories 504a/b/c, or between any two of the memories 504a/b/c, etc. The local interface 507a/b/c may comprise additional systems designed to coordinate this communication, including, for example, performing load balancing. The processor 502a/b/c may be of electrical or of some other available construction.
[0059] Although the various systems described herein may be embodied in software or code executed by general purpose hardware as discussed above, as an alternative the same may also be embodied in dedicated hardware or a combination of software/general purpose hardware and dedicated hardware. If embodied in dedicated hardware, each can be implemented as a circuit or state machine that employs any one of or a combination of a number of technologies. These technologies may include, but are not limited to, discrete logic circuits having logic gates for implementing various logic functions upon an application of one or more data signals, application specific integrated circuits (ASICs) having appropriate logic gates, field-programmable gate arrays (FPGAs), or other components, etc. Such technologies are generally well known by those skilled in the art and, consequently, are not described in detail herein.
[0060] The flowcharts of FIGS. 2, 4 and 4 show the functionality and operation of an implementation of portions of the edge nodes 114a/b or controller computing environment 134. If embodied in software, each block may represent a module, segment, or portion of code that comprises program instructions to implement the specified logical function(s). The program instructions may be embodied in the form of source code that comprises human- readable statements written in a programming language or machine code that comprises numerical instructions recognizable by a suitable execution system such as a processor in a computer system or other system. The machine code may be converted from the source code, etc. If embodied in hardware, each block may represent a circuit or a number of interconnected circuits to implement the specified logical function(s).
[0061] Although the flowcharts of FIGS. 2, 3 and 4 show a specific order of execution, it is understood that the order of execution may differ from that which is depicted. For example, the order of execution of two or more blocks may be scrambled relative to the order shown. Also, two or more blocks shown in succession in FIGS. 2, 3 and 4 may be executed concurrently or with partial concurrence. Further, in some embodiments, one or more of the blocks shown in FIGS. 2, 3 and 4 may be skipped or omitted. In addition, any number of counters, state variables, warning semaphores, or messages might be added to the logical flow described herein, for purposes of enhanced utility, accounting, performance measurement, or providing troubleshooting aids, etc. It is understood that all such variations are within the scope of the present disclosure.
[0062] Also, any logic or application described herein that comprises software or code can be embodied in any non-transitory computer-readable medium for use by or in connection with an instruction execution system such as, for example, a processor 502a/b/c in a computer system or other system. In this sense, the logic may comprise, for example, statements including instructions and declarations that can be fetched from the computer-readable medium and executed by the instruction execution system. In the context of the present disclosure, a "computer-readable medium" can be any medium that can contain, store, or maintain the logic or application described herein for use by or in connection with the instruction execution system. [0063] The computer-readable medium can comprise any one of many physical media such as, for example, magnetic, optical, or semiconductor media. More specific examples of a suitable computer-readable medium would include, but are not limited to, magnetic tapes, magnetic floppy diskettes, magnetic hard drives, memory cards, solid-state drives, USB flash drives, or optical discs. Also, the computer-readable medium may be a random access memory (RAM) including, for example, static random access memory (SRAM) and dynamic random access memory (DRAM), or magnetic random access memory (MRAM). In addition, the computer-readable medium may be a read-only memory (ROM), a programmable readonly memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or other type of memory device.
[0064] Further, any logic or application described herein may be implemented and structured in a variety of ways. For example, one or more applications described may be implemented as modules or components of a single application. Further, one or more applications described herein may be executed in shared or separate computing devices or a combination thereof. For example, a plurality of the applications described herein may execute in the same computing device 501 a/b/c, or in multiple computing devices in the same server computing environment 101 , client 104 or controller computing environment 134. Additionally, it is understood that terms such as "application," "service," "system," "engine," "module," and so on may be interchangeable and are not intended to be limiting.
[0065] Disjunctive language such as the phrase "at least one of X, Y, or Z," unless specifically stated otherwise, is otherwise understood with the context as used in general to present that an item, term, etc., may be either X, Y, or Z, or any combination thereof (e.g. , X, Y, and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y, or at least one of Z to each be present.
[0066] It should be emphasized that the above-described embodiments of the present disclosure are merely possible examples of implementations set forth for a clear understanding of the principles of the disclosure. Many variations and modifications may be made to the above-described embodiment(s) without departing substantially from the spirit and principles of the disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure and protected by the following claims.

Claims

CLAIMS Therefore, the following is claimed:
1. A system, comprising:
at least one first computing device configured to at least:
receive a plurality of baseline packets addressed to at least one destination address;
generate a plurality of overlay packets from the plurality of baseline packets, the plurality of overlay packets comprising metadata to facilitate a communication of the plurality of overlay packets via a plurality of paths of an overlay network to the at least one destination address;
communicate the plurality of overlay packets to each of at least one first relay node corresponding to each of the plurality of paths of the overlay network; at least one second computing device configured to at least:
receive the plurality of overlay packets from the overlay network; generate the plurality of baseline packets from the plurality of overlay packets; and
communicate the plurality of baseline packets to the at least one destination network address.
2. The system of claim 1, wherein generating the plurality of overlay packets comprises splitting the plurality of baseline packets into a plurality of data portions, wherein the plurality of overlay packets comprise a payload corresponding to a respective one of the plurality of data portions.
3. The system of claim 2, wherein the plurality of baseline packets are split based at least in part on the plurality of overlay packets conforming to a Maximum Transmission Unit size of at least a portion of the overlay network.
4. The system of claim 1, wherein the at least one second computing device is configured to discard a duplicated one of the plurality of overlay packets received from the overlay network.
5. The system of claim 4, wherein the at least one second computing device is configured to identify the duplicated one of the plurality of overlay packets based at least in part on a sequence number or a hashing algorithm applied to the plurality of overlay packets.
6. The system of claim 1, wherein the at least one first relay node is included in a plurality of relay nodes of the overlay network, and the at least one computing device is further configured to at least:
identify the plurality of paths based at least in part on a path finding operation applied to the plurality of relay nodes; and
select the at least one first relay node according to the identified plurality of paths.
7. The system of claim 5, wherein the at least one computing device is further configured to at least encode an indication of a corresponding one of the plurality of paths in the metadata.
8. The system of claim 1, further comprising at least one third computing device configured to at least:
obtain, from a plurality of relay nodes of the overlay network, relay data; generate, based at least in part on the relay data, the plurality of paths of the overlay network; and
communicate, to at least the at least one first computing device, an indication of at least the at least one first relay node corresponding to each of the plurality of paths of the overlay network.
9. The system of claim 8, wherein the at least one third computing device is further configured to at least communicate, to a subset of the plurality of relay nodes included in the plurality of paths, an indication of a respective next one of the subset of the plurality of relay nodes included in the plurality of paths.
10. The system of claim 9, wherein the at least one third computing device is further configured to at least repeat, at a predefined interval or in response to a request, the obtaining of the relay data and the communicating of the indication of the respective next one of the subset of the plurality of relay nodes included in the plurality of paths.
11. A method, comprising:
receiving, by at least one first computing device, a plurality of baseline packets addressed to at least one destination address;
generating, by the at least one first computing device, a plurality of overlay packets from the plurality of baseline packets, the plurality of overlay packets comprising metadata to facilitate a communication of the plurality of overlay packets via a plurality of paths of an overlay network to the at least one destination address; communicating, by the at least one first computing device, the plurality of overlay packets to each of at least one first relay node corresponding to each of the plurality of paths of the overlay network;
receiving, by at least one second computing device, the plurality of overlay packets from the overlay network;
generating, by the at least one second computing device, the plurality of baseline packets from the plurality of overlay packets; and
communicating, by the at least one second computing device, the plurality of baseline packets to the at least one destination network address.
12. The method of claim 11 , wherein generating the plurality of overlay packets comprises splitting the plurality of baseline packets into a plurality of data portions, wherein the plurality of overlay packets comprise a payload corresponding to a respective one of the plurality of data portions.
13. The method of claim 12, wherein the plurality of baseline packets are split based at least in part on the plurality of overlay packets conforming to a Maximum Transmission Unit size of at least a portion of the overlay network.
14. The method of claim 1 1 , further comprising discarding, by the at least one second computing device, a duplicated one of the plurality of overlay packets received from the overlay network.
15. The method of claim 14, further comprising identifying, by the at least one second computing device, the duplicated one of the plurality of overlay packets based at least in part on a sequence number or a hashing algorithm applied to the plurality of overlay packets.
16. The method of claim 1 1, wherein the at least one first relay node is included in a plurality of relay nodes of the overlay network, and the method further comprises:
identifying, by the at least one first computing device, the plurality of paths based at least in part on a path finding operation applied to the plurality of relay nodes; and
selecting, by the at least one first computing device, the at least one first relay node according to the identified plurality of paths.
17. The method of claim 16, wherein the method further comprises encoding, by the at least one first computing device, an indication of a corresponding one of the plurality of paths in the metadata.
18. The method of claim 1 1 , wherein the at least one first relay node is configured to select a next relay node from a plurality of relay nodes based at least in part on at least one of latency data or capacity data for the plurality of relay nodes.
19. The method of claim 1 1 , further comprising generating, by the at least one first computing device, the plurality of baseline packets.
20. The method of claim 11 , wherein the plurality of baseline packets are generated by a second process executed in another computing device communicatively coupled to the at least one first computing device.
PCT/US2017/028504 2016-04-20 2017-04-20 Parallel multipath routing architecture WO2017184807A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201662325134P 2016-04-20 2016-04-20
US62/325,134 2016-04-20

Publications (1)

Publication Number Publication Date
WO2017184807A1 true WO2017184807A1 (en) 2017-10-26

Family

ID=60116341

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2017/028504 WO2017184807A1 (en) 2016-04-20 2017-04-20 Parallel multipath routing architecture

Country Status (1)

Country Link
WO (1) WO2017184807A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9992101B2 (en) 2014-11-24 2018-06-05 Taric Mirza Parallel multipath routing architecture
WO2020014115A1 (en) * 2018-07-09 2020-01-16 Replixio Ltd. System and method for data transmission in distributed computing environments

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090073921A1 (en) * 2007-09-19 2009-03-19 At&T Services Inc. Data forwarding in hybrid mesh networks
US20140297785A1 (en) * 2013-04-02 2014-10-02 Cisco Technology, Inc. Offload Operations for Overlay Networks
US20150188823A1 (en) * 2013-12-03 2015-07-02 Akamai Technologies, Inc. Virtual private network (VPN)-as-a-service with load-balanced tunnel endpoints
US20150358668A1 (en) * 2012-11-19 2015-12-10 Videolink Llc Internet-Based Video Delivery System
US20150358348A1 (en) * 2014-06-04 2015-12-10 Aaa Internet Publishing, Inc. Method of DDos and Hacking Protection for Internet-Based Servers Using a Private Network of Internet Servers by Executing Computer-Executable Instructions Stored On a Non-Transitory Computer-Readable Medium
US20160072669A1 (en) * 2008-11-12 2016-03-10 Teloip Inc. System, apparatus and method for providing a virtual network edge and overlay with virtual control plane
EP2996289A1 (en) * 2014-09-15 2016-03-16 Nokia Solutions and Networks GmbH & Co. KG Method, system and apparatus for establishing a data connection with parallel communication channels
US20160149804A1 (en) * 2014-11-24 2016-05-26 Taric Mirza Parallel multipath routing architecture

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090073921A1 (en) * 2007-09-19 2009-03-19 At&T Services Inc. Data forwarding in hybrid mesh networks
US20160072669A1 (en) * 2008-11-12 2016-03-10 Teloip Inc. System, apparatus and method for providing a virtual network edge and overlay with virtual control plane
US20150358668A1 (en) * 2012-11-19 2015-12-10 Videolink Llc Internet-Based Video Delivery System
US20140297785A1 (en) * 2013-04-02 2014-10-02 Cisco Technology, Inc. Offload Operations for Overlay Networks
US20150188823A1 (en) * 2013-12-03 2015-07-02 Akamai Technologies, Inc. Virtual private network (VPN)-as-a-service with load-balanced tunnel endpoints
US20150358348A1 (en) * 2014-06-04 2015-12-10 Aaa Internet Publishing, Inc. Method of DDos and Hacking Protection for Internet-Based Servers Using a Private Network of Internet Servers by Executing Computer-Executable Instructions Stored On a Non-Transitory Computer-Readable Medium
EP2996289A1 (en) * 2014-09-15 2016-03-16 Nokia Solutions and Networks GmbH & Co. KG Method, system and apparatus for establishing a data connection with parallel communication channels
US20160149804A1 (en) * 2014-11-24 2016-05-26 Taric Mirza Parallel multipath routing architecture

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9992101B2 (en) 2014-11-24 2018-06-05 Taric Mirza Parallel multipath routing architecture
WO2020014115A1 (en) * 2018-07-09 2020-01-16 Replixio Ltd. System and method for data transmission in distributed computing environments
US10826819B2 (en) 2018-07-09 2020-11-03 Replixio Ltd. System and method for data transmission in distributed computing environments

Similar Documents

Publication Publication Date Title
US9992101B2 (en) Parallel multipath routing architecture
US10581884B2 (en) Channel data encapsulation system and method for use with client-server data channels
EP3353997B1 (en) Technologies for offloading data object replication and service function chain management
US10630710B2 (en) Systems and methods of stateless processing in a fault-tolerant microservice environment
CN106537880B (en) Caching data in an information centric network architecture
CN108259347B (en) Message transmission method and device
US9369435B2 (en) Method for providing authoritative application-based routing and an improved application firewall
CN113326228B (en) Message forwarding method, device and equipment based on remote direct data storage
EP2995042A1 (en) Data plane learning of bi-directional service chains
US11765081B2 (en) Offloads for multicast virtual network packet processing in a network interface card
US8887280B1 (en) Distributed denial-of-service defense mechanism
US20230208721A1 (en) Determining network topology based on packet traffic
US10171354B2 (en) Communication processing system, communication processing apparatus, communication processing method, and storage medium
CN110545230B (en) Method and device for forwarding VXLAN message
US10033593B2 (en) Using timestamps to analyze network topologies
WO2017184807A1 (en) Parallel multipath routing architecture
WO2016154841A1 (en) Load balancing
US10367725B2 (en) Network programming
WO2016122562A1 (en) Replicating network communications
CN114268630A (en) Method, device and equipment for realizing random load balance access based on static ARP (Address resolution protocol) table entry
CN116405437A (en) Message processing method and message processing system

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17786603

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 17786603

Country of ref document: EP

Kind code of ref document: A1