WO2006124009A2 - Hardware filtering support for denial-of-service attacks - Google Patents

Hardware filtering support for denial-of-service attacks Download PDF

Info

Publication number
WO2006124009A2
WO2006124009A2 PCT/US2005/007059 US2005007059W WO2006124009A2 WO 2006124009 A2 WO2006124009 A2 WO 2006124009A2 US 2005007059 W US2005007059 W US 2005007059W WO 2006124009 A2 WO2006124009 A2 WO 2006124009A2
Authority
WO
WIPO (PCT)
Prior art keywords
packet
data packet
network node
cpu
received data
Prior art date
Application number
PCT/US2005/007059
Other languages
French (fr)
Other versions
WO2006124009A3 (en
Inventor
John Kenneth Stacy
Trevor Garner
Martin W. Hughes
William R. Lee
Original Assignee
Cisco Technology, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cisco Technology, Inc. filed Critical Cisco Technology, Inc.
Priority to EP05857867.5A priority Critical patent/EP1754349B1/en
Priority to CA2559251A priority patent/CA2559251C/en
Priority to CN2005800090740A priority patent/CN101421991B/en
Priority to AU2005326185A priority patent/AU2005326185A1/en
Publication of WO2006124009A2 publication Critical patent/WO2006124009A2/en
Publication of WO2006124009A3 publication Critical patent/WO2006124009A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • H04L49/901Buffering arrangements using storage descriptor, e.g. read or write pointers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/32Flow control; Congestion control by discarding or delaying data units, e.g. packets or frames
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • H04L49/9042Separate storage for different parts of the packet, e.g. header and payload
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/90Buffering arrangements
    • H04L49/9047Buffering arrangements including multiple buffers, e.g. buffer pools
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1408Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic by monitoring network traffic
    • H04L63/1416Event detection, e.g. attack signature detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1441Countermeasures against malicious traffic
    • H04L63/1458Denial of Service

Definitions

  • This invention relates generally to network communications, and, more specifi- cally, to a system and method for filtering denial-of-service traffic in an intermediate network node, such as a router.
  • a computer network is a geographically distributed collection of interconnected subnetworks for transporting data between nodes, such as computers.
  • a local area net- work is an example of such a subnetwork.
  • the network's topology is defined by an arrangement of client nodes that communicate with one another, typically through one or more intermediate network nodes, such as a router or switch.
  • a client node is an endstation node that is configured to originate or terminate communications over the network.
  • an intermediate network node is a node that facilitates routing data between client nodes. Communications between nodes are typically effected by exchanging discrete packets of data according to predefined protocols.
  • a protocol consists of a set of rules defining how the nodes interact with each other.
  • Each data packet typically comprises "payload” data prepended (“encapsu- lated”) by at least one network header formatted in accordance with a network communication protocol.
  • the network headers include information that enables the client nodes and intermediate nodes to efficiently route the packet through the computer network.
  • a packet's network headers include at least a data-link (layer 2) header, an internetwork (layer 3) header and a transport (layer 4) header, as defined by the Open Systems Interconnection (OSI) Reference Model.
  • the OSI Reference Model is generally described in more detail in Section 1.1 of the reference book entitled Interconnections Second Edition, by Radia Perlman, published September 1999, which is hereby incorporated by reference as though fully set forth herein.
  • the data-link header provides information for transmitting the packet over a particular physical link (i.e., a communication medium), such as a point-to-point link, Ethernet link, wireless link, optical link, etc.
  • a particular physical link i.e., a communication medium
  • the data-link header may specify a pair of "source” and "destination” network interfaces that are connected by the physical link.
  • a network interface contains the mechanical, electrical and signaling circuitry and logic used to couple a network node to one or more physical links.
  • a network interface is often associated with a hardware-specific address, known as a media access control (MAC) address. Accordingly, the source and destination network interfaces in the data-link header are typically represented as source and destination MAC addresses.
  • the data-link header may also store flow control, frame synchronization and error checking information used to manage data transmissions over the physical link.
  • the internetwork header provides information defining the packet's logical path (or "virtual circuit") through the computer network. Notably, the path may span multi- pie physical links.
  • the internetwork header may be formatted according to the Internet Protocol (IP), which specifies IP addresses of both a source and destination node at the end points of the logical path.
  • IP Internet Protocol
  • the packet may "hop" from node to node along its logical path until it reaches the client node assigned to the destination IP address stored in the packet's internetwork header.
  • the source and destination MAC addresses in the packet's data-link header may be updated, as necessary.
  • the source and destination IP addresses typically remain unchanged as the packet is transferred from link to link in the network.
  • the transport header provides information for ensuring that the packet is reliably transmitted from the source node to the destination node.
  • the transport header typically includes, among other things, source and destination port numbers that respectively identify particular software applications executing in the source and destination nodes. More specifically, the packet is generated in the source node by the application assigned to the source port number. Then, the packet is forwarded to the destination node and directed to the application assigned to the destination port number.
  • the transport header also may include error-checking information (i.e., a checksum) and other data-flow control information. For instance, in connection-oriented transport pro- tocols such as the Transmission Control Protocol (TCP), the transport header may store sequencing information that indicates the packet's relative position in a transmitted stream of data packets.
  • TCP Transmission Control Protocol
  • a dataflow is a stream of data packets that is communicated from a source node to a destination node.
  • Each packet in the flow satisfies a set of predetermined criteria, e.g., based on the packet's contents, size or relative position (i.e., temporal or spatial) in the data flow.
  • An intermediate network node may be configured to perform "flow-based" routing operations so as to route each packet in a data flow in the same manner.
  • the intermediate node typically receives data packets in the flow and forwards the packets in accordance with predetermined routing information that is distributed using a protocol, such as the Open Shortest Path First (OSPF) protocol.
  • OSPF Open Shortest Path First
  • the intermediate node need only perform one forwarding decision for the entire data flow, e.g., based on the first packet received in the flow. Thereafter, the intermediate node forwards packets in the data flow based on the flow's previously determined routing information (i.e., adjacency information). In this way, the intermediate node consumes fewer resources, such as processor bandwidth and processing time, than if it performed a separate forwarding decision for every packet in the data flow.
  • the intermediate network node may implement a hash table which stores packet-related information used to classify received packets into their corresponding data flows.
  • the hash table is typically organized as a table of linked lists, where each list may be indexed by the result of applying a conventional hash function to "signature" information.
  • a signature is a set of values that remain constant for every packet in a data flow. For example, assume each packet in a first data flow stores the same pair of source and destination IP address values. In this case, a signature for the first data flow may be generated based on the values of these source and destination IP addresses. Likewise, a different signature may be generated for a second data flow whose packets store a different set of source and destination IP addresses than packets in the first data flow.
  • a data flow's signature information is not limited to IP addresses and may include other information, such as TCP port numbers, IP version numbers and so forth.
  • Each linked list in the hash table contains one or more entries, and each linked- list entry stores information corresponding to a particular data flow.
  • Such information may include, inter alia, the data flow's associated signature information and a dataflow identifier ("flow ID").
  • the flow ID identifies the particular data flow and also may be used to locate routing information associated with the data flow.
  • the intermediate network node may maintain a data structure that maps flow ID values to the memory locations of their corresponding routing information, e.g., stored in the node's local or internal memory. Alternatively, the flow ID values may directly incorporate the memory locations of their data flows' routing information.
  • signature information is extracted from the packet's network headers and hashed using a conventional hash function, such as a cyclic redundancy check (CRC) function.
  • CRC cyclic redundancy check
  • the resultant hash value is used to index a hash-table entry which, in turn, references a linked list. Entries in the linked list are accessed sequentially until a "matching" entry is found storing the extracted signature.
  • the entry's stored flow ID value is used to associate the received packet with a data flow and the packet is routed in accordance with that flow.
  • the intermediate network node typically receives a large number of data flows from various sources, including client nodes and other intermediate nodes. Each source may be responsible for establishing one or more data flows with the intermediate node. To optimize use of its processing bandwidth, the intermediate node may process the received flows on a prioritized basis. That is, as packets are received at the intermediate node, they are identified as belonging to, e.g., a high or low priority data flow. Packets in the high-priority flow may be processed by the intermediate node in advance of the low-priority packets, even if the low-priority packets were received before the high-priority packets.
  • DoS attacks have become fairly common techniques for disabling access to resources and/or services in an intermediate network node.
  • a DoS attack corresponds to a data flow of "malicious" packets which, when processed by the intermediate network node, deprive non-malicious packets (i.e., non-DoS packets) access to certain resources and/or services in the node.
  • the DoS packets may be sent from a single source or may be coordinated among a plurality of sources. This latter case is often referred to as a distributed DoS (DDoS) attack.
  • a computer hacker may launch a DDoS attack through a multitude of compromised endstations that transmit data packets to a target intermediate node, thereby overwhelming the interme- diate node's processing bandwidth.
  • DoS attacks typically involve sending large quantities of a specific type of network traffic, such as packets formatted in accordance with the Internet Control Message Protocol (ICMP) or the Internet Group Management Protocol (IGMP), to the intermediate network node.
  • ICMP Internet Control Message Protocol
  • IGMP Internet Group Management Protocol
  • the DoS packets are pre-pended by a complex arrangement of network headers.
  • the targeted intermediate network node becomes overburdened not only by the large quantity of received DoS packets, but also by the consumption of resources required to process them. Since the intermediate node's resources become overly consumed processing these malicious packets, other non-malicious packets sent to the intermediate node are often dropped or discarded. Accordingly, different types of intermediate network nodes attempt to prevent DoS attacks in various ways.
  • the high-end "core” routers and switches typically have enough processing bandwidth to process both malicious DoS packets as well as non-malicious packets.
  • the high-end routers and switches are designed to handle large amounts of network traffic, e.g., on a network "backbone.” Consequently, the malicious packets can be processed at the rate at which they are received, i.e., at "line” rate.
  • These high- end intermediate nodes thus rely primarily on hardware forwarding solutions that cannot become over-subscribed while identifying and removing the malicious DoS packets. As a result, a substantial portion of processing bandwidth in the central processing unit(s) (CPU) executing the software may be consumed identifying and removing DoS packets.
  • Another disadvantage of this solution is that the routing or switching software becomes more complex by the inclusion of code for filtering the DoS packets from the received data flows.
  • the "mid-range" routers and switches unlike their high-end counterparts, typi- cally become oversubscribed as a result of a DoS attack.
  • These intermediate nodes are usually enterprise or LAN routers/switches that manage a relatively large number of data flows.
  • the mid- range routers and switches In order to identify and remove DoS traffic (i.e., data packets), the mid- range routers and switches typically utilize software executing on a centralized CPU or on a network processor supporting a general-purpose CPU. Like the software in the high-end routers and switches, the software in the mid-range routers and switches con- sumes an excessive amount of processing bandwidth and complexity for thwarting the DoS attack.
  • Hardware support for prioritization of ingress traffic is sometimes implemented in the mid-range routers and switches when the "problem" DoS traffic can be filtered and put on a low-priority queue serviced by the software.
  • the hardware filtering is typically implemented as a simple table lookup on data link (layer 2) or internetwork (layer 3) information contained in the received data packets.
  • the table lookup may be performed using a content addressable memory (CAM), such as a ternary CAM (TCAM).
  • CAM content addressable memory
  • TCAM ternary CAM
  • the low-end "access" routers and switches are typically single CPU systems that process a relatively small amount of network traffic and are therefore more susceptible to a DoS attack than the above ⁇ noted mid-range and high-end intermediate network nodes. There is only one CPU in the low-end routers and switches, so CPU bandwidth is typically not consumed pre-processing or prioritizing in-coming data packets.
  • Such pre-processing would require the software executing on the CPU to process each received packet twice (prioritize and route) thus consuming an unacceptable amount of processing resources. Therefore, the low-end routers and switches usually only filter received data packets (if at all) using simple lookup tables or TCAMs that are not able to identify complex DoS packet encapsulations.
  • the intermediate network node There is generally a need for an intermediate network node that can identify and remove DoS traffic without consuming an excessive amount of processing resources or bandwidth in the node. Further, the intermediate node should identify and remove DoS traffic having encapsulations of any arbitrary complexity.
  • the malicious DoS packets should be removed from the intermediate node without affecting the processing of non-malicious packets, such as low-priority non-malicious packets.
  • the present invention overcomes the disadvantages of the prior art by providing a system and method for automatically identifying and removing malicious data packets, such as denial-of-service (DoS) packets, in an intermediate network node before the packets can be forwarded to a central processing unit (CPU) in the node.
  • DoS denial-of-service
  • the CPU's processing bandwidth is therefore not consumed identifying and removing the malicious packets from the system memory.
  • processing of the malicious packets is essentially "off-loaded" from the CPU, thereby enabling the CPU to process non- malicious packets in a more efficient manner.
  • the invention identifies malicious packets having complex encapsulations that can not be identi- fied using traditional techniques, such as TCAMs or lookup tables.
  • the invention may be employed in an intermediate network node configured to perform as a high-end, mid-range or access router. Further, the invention may be used in conjunction with conventional access control lists (ACL) and/or intrusion detection systems (IDS) in order to delete DoS traffic whose encapsulations are too complex to be filtered by the ACL or IDS. More generally, the invention may be configured, either automatically or manually, to filter malicious packets having encapsulations of any arbitrary complexity. The malicious data packets are preferably identified using hash-based flow identification, although other flow classification techniques may be employed. The invention is illustratively implemented in a hardware assist device in the intermediate network node. However, it is also expressly contemplated that the invention may be embodied by other combinations of hardware and/or software. BRIEF DESCRIPTION OF THE DRAWINGS
  • Fig. 1 is a schematic block diagram of a computer network comprising a collection of interconnected subnetworks and nodes, including an intermediate network node;
  • Fig. 2 is a schematic block diagram of an illustrative intermediate network node that may be used in accordance with the present invention
  • Fig. 3 is a schematic block diagram of a system controller that may be implemented in the illustrative intermediate network node in Fig. 2;
  • Fig. 4 is a schematic block diagram of an exemplary buffer descriptor that may be enqueued in the ingress and/or egress descriptor rings of the present invention
  • Fig. 5 is a schematic block diagram of an illustrative hardware assist (HWA) module that may be used to automatically identify and remove malicious data packets in accordance with the present invention
  • Fig. 6 is a schematic block diagram of an exemplary hash table that may be searched by the HWA module for identifying a data flow associated with a received data packet; and Figs. 7A-B are a flow chart illustrating a sequence of steps that may be performed for automatically identifying and removing malicious data packets in accordance with the present invention.
  • Fig. 1 is a block diagram of a computer network 100 comprising a collection of interconnected subnetworks and nodes.
  • the nodes may comprise computers including end nodes 130 and 140, such as a sending end node 120 and a receiving end node 150, and an intermediate network node 200, the latter of which may be a switch or router.
  • the subnetworks 105, 110 included within network 100 are preferably local area networks (LANs) interconnected by the intermediate node 200, although the networks may comprise other communication links, such as wide area networks. Communication among the nodes coupled to the LANs is typically effected by exchanging discrete packets 160 among the nodes.
  • LANs local area networks
  • the sending node 120 generates a data packet 160 by encapsulating "payload" data within headers, such as conventional data link and internetwork s headers, as the data passes through different layers of a protocol stack.
  • the packet is then transmitted over the network to the intermediate node 200 which facilitates the flow of the data packet through the network by routing it to the proper receiving node 150.
  • the node 200 receives the packet at one of its network interfaces and renders a forwarding decision for the packet based on a destination end node specified io by the packet's internetwork header.
  • the packet's data link header is modified in accordance with the forwarding decision and the packet is transmitted over an appropriate subnetwork coupled to the intermediate network node.
  • Fig. 2 is a schematic block diagram of an intermediate node 200 that may be advantageously used with the present invention.
  • the node comprises a plurality of is network interfaces 210, a system controller 300, a central processing unit (CPU) 230 and a memory 250. Data is received by the network interfaces 210, each of which is coupled to at least one network or subnetwork, such as LANs 105 and 110.
  • the network interfaces contain the mechanical, electrical and signaling circuitry that enables the intermediate network node 200 to communicate over physical links connected to
  • ATM asynchronous transfer mode
  • ⁇ networks synchronous optical networks (SONET), wireless networks, frame relay networks, Ethernet networks, Fiber Distributed Data Interface (FDDI) networks, etc.
  • SONET synchronous optical networks
  • FDDI Fiber Distributed Data Interface
  • the system controller 300 is coupled to each network interface 210, the CPU 230 (i.e., a processor) and the memory 250 by different local buses in the intermediate
  • the system controller may be coupled to the network interfaces 210 by respective peripheral component interconnect (PCI) buses, whereas the controller may be coupled to the memory 250 by a plurality of high-speed connections, such as HyperTransport bus links.
  • PCI peripheral component interconnect
  • the controller 300 therefore functions as a "bridge" for transferring data from one local bus to another. That is, the controller re-
  • the system controller ceives data over a first local bus, e.g., coupled to a network interface 210, and converts the data to a format that may be transmitted over a second local bus, e.g., coupled to the memory 250.
  • the system controller may also include other functionality, such as ap- plication-specific circuitry or logic.
  • the controller 300 may be embodied in hardware as a field-programmable gate array (FPGA) or an application-specific integrated circuit (ASIC), although the controller's functionality alternatively may be implemented in various combinations of hardware and/or software.
  • the memory 250 comprises a plurality of storage locations that are addressable by the CPU 230 and the network interfaces 210 via the system controller 300.
  • the memory comprises a form of random access memory (RAM) that is generally cleared by a power cycle or other reboot operation (e.g., it is a "volatile" memory).
  • the memory 250 may comprise dynamic random access memory (DRAM) and/or synchronous DRAM (SDRAM) storage locations adapted to store program code and data structures accessible to the CPU 230. It will be apparent to those skilled in the art that the memory 250 may also comprise other memory means, including various computer-readable media, for storing program instructions and data structures pertaining to the operation of the intermediate network node 200.
  • a router operating system 260 portions of which are typically resident in the memory 250 and executed by the CPU 230, functionally organizes the intermediate network node 200 by, inter alia, invoking network operations in support of software processes executing on the intermediate node.
  • the IOSTM operating system by Cisco Systems, me. is one example of a router operating system 260.
  • the operating system may perform routing operations on data packets 160 received by the network interfaces 210.
  • a portion of the memory 250 may be organized as a buffer "pool" 240 of data buffers 242 configured to store received data packets. For example, each buffer may be configured to store up to a fixed amount, e.g., 2 kilobytes, of the received data packet 160.
  • a received packet 160 is transferred from a network interface 210 to one or more of the buffers 242.
  • the CPU 230 executing the router operating system 260 renders a forwarding decision for the received packet based on routing information 270 stored in the memory 250.
  • One or more data structures such as the hash table 600, may be stored in the memory to facilitate the operating system's forwarding decision.
  • the hash table 600 may be used to identify a data flow associated with the received packet, and the routing information 270 may store adjacency information associated with the identified flow.
  • the packet's network headers are modi- fied in accordance with the adjacency information associated with the packet's identified data flow.
  • Fig. 3 is a schematic block diagram of the system controller 300 that may be implemented in the illustrative intermediate network node 200.
  • the system controller comprises a plurality of first local bus (PCI) interfaces 310, a memory controller 320, a CPU bus interface 330, a bus controller 340, an on-chip memory 350 and a hardware assist (HWA) module 500 interconnected by a device bus 360.
  • PCI first local bus
  • each PCI interface 310 includes circuitry and logic configured to send and receive data over a PCI bus coupled to a network interface 210.
  • PCI in- terfaces 310 alternatively may be substituted by respective controllers that communicate over other types of buses, such as Industry Standard Architecture (ISA) buses, Extended ISA (EISA) buses, etc.
  • ISA Industry Standard Architecture
  • EISA Extended ISA
  • Data received at a network interface 210 is forwarded over a PCI bus to a PCI interface 310, which frames the received data so it may be transferred over the device bus 360.
  • the PCI interface may receive data from the bus 360 and reformat the data for transmission over the PCI bus.
  • the memory controller 320 comprises circuitry and logic configured to transfer data from the memory 250 over the second local bus to the device bus 360, and vice versa.
  • the CPU 230 may forward a memory address (or range of addresses) to the CPU bus interface 330.
  • the memory address may be accompanied by a CPU instruction to read or write data at that memory address.
  • the CPU bus interface 330 transmits the memory address and its corresponding CPU instruction over the device bus 360 to the memory controller 320.
  • the memory controller writes or retrieves data at the specified memory address, in accordance with the CPU instruction.
  • the bus controller 340 comprises circuitry and logic that, inter alia, implements an arbitration policy for coordinating access to the device bus 360.
  • the controller 340 prevents two or more entities, such as the PCI interfaces 310, memory controller 320, etc., from attempting to access the bus 360 at substantially the same time.
  • the bus controller 340 may be configured to grant or deny access to the bus 360 based on a predefined arbitration protocol.
  • one or more functions normally performed by the CPU 230 executing the router operating system 260 may be "off-loaded" to the HWA module 500.
  • the HWA module may be configured to filter denial-of-service (DoS) packets received at the intermediate network node 200, before the packets can be processed by the CPU.
  • DoS denial-of-service
  • the processing bandwidth of the CPU 230 is not consumed identifying and removing "malicious" DoS traffic, as in previous implementations.
  • the HWA module is configured to identify DoS packets having arbitrarily complex encapsulations that otherwise can not be identified using traditional techniques, such as TCAMs or lookup ta- bles.
  • the on-chip memory 350 comprises a set of addressable memory locations resident on the system controller 300.
  • the on-chip memory may be a form of volatile memory, such as static RAM (SRAM), or a form of erasable non- volatile memory, such as Flash memory.
  • SRAM static RAM
  • Flash memory erasable non- volatile memory
  • the illustrative on-chip memory 350 is situated in the sys- tern controller 300, it is also expressly contemplated that the on-chip memory may reside in a separate memory module coupled to the system controller, or the contents of the on-chip memory (or a portion thereof) may be incorporated into the "main" memory 250.
  • the on-chip memory 350 stores, among other things, one or more "ingress” de- scriptor rings 352 (i.e., circular f ⁇ rst-in, first-out (FIFO) queues), one or more "egress” descriptor rings 354 and a "delete” descriptor ring 356.
  • Each network interface 210 is associated with at least one ingress ring 352 in the on-chip memory 350.
  • the packet data is forwarded over an appropriate PCI bus, through the system controller 300, to an available data buffer 242 in the memory 250.
  • a memory reference i.e., a "descriptor" to the data buffer is then enqueued in the ingress descriptor ring 352 associated with the network interface 210 that received the packet data. Packet data is stored and descriptors are enqueued in this manner until the network interface 210 determines that an entire packet 160 has been received or an error has occurred. Accordingly, the network interface's ingress ring 352 stores an ordered list of descriptors corresponding to the order in which the packet data is received at the interface 210. After the packet 160 has been received, the network interface 210 notifies the HWA module 500 which ingress descriptor ring 352 contains the received packet's descriptors.
  • the HWA module dequeues the descriptors and processes the received packet data to determine the packet's associated data flow.
  • the HWA module performs hash-based flow classification that enables it to classify packets having arbitrarily complex encapsulations. Having identified the packet's associated data flow, the HWA module enqueues the packet's descriptors in an appropriate egress descriptor ring 354, selected based on the identified data flow.
  • the CPU 230 is notified, e.g., by a hardware interrupt, when the packet's de- scriptors have been enqueued on one of its egress descriptor rings 354. In response, the CPU dequeues the descriptors from its egress descriptor ring and renders a forwarding decision for the received data packet 160.
  • each of the egress descriptor rings 354 corresponds to a different destination and/or priority level that may be associated with the identified data flow.
  • the egress rings 354 may be separate high-priority and low-priority egress descriptor rings, respectively corresponding to high-priority and low-priority data flows processed by the CPU 230.
  • a set of one or more high-priority and/or low-priority egress descriptor rings 354 may be associated with each CPU 230 in the plurality of CPUs.
  • the HWA module 500 determines that the received data packet 160 is a DoS packet or other type of malicious packet, e.g., based on the packet's identified data flow, the HWA module enqueues the packet's descriptors on the delete egress ring 356 or on a first in, first out queue (not shown) of "free" buffer descriptors, i.e., descriptors whose referenced buffers 242 are available to store new packet data, associated with the network interface 210.
  • descriptors placed on the delete ring 356 or on the free-buffer FIFO are "recycled" (i.e., reused) before they can be forwarded to the CPU 230.
  • Fig. 4 illustrates an exemplary descriptor 400 that may be enqueued in at least one of the ingress and egress descriptor rings 352-356.
  • the descriptor contains, inter alia, a set of flags 410, a buffer pointer 420 as well as other information 430 associated with the descriptor.
  • the flags 410 may include, for example, an ownership flag 412, a start-of-packet (SOP) flag 414, an end-of-packet (EOP) flag 416 and an error (ERR) flag 418.
  • the ownership flag 412 stores a value that indicates which device is granted ownership (i.e., control) of the descriptor 400. For instance, the ownership flag may equal a first value to indicate that a network interface 210 is given exclusive access to the descriptor 400; the flag 412 may equal a second value when control of the descriptor is transferred to the HWA module 500.
  • the SOP flag 414 may be set equal to a value that indicates whether the de- scriptor's referenced data buffer 242 contains the start of the received data packet 160.
  • the EOP flag 416 may be set equal to a value that indicates whether the descriptor's referenced data buffer contains the end of the packet 160.
  • the ERR flag 418 may be set equal to a value indicating whether an error has been detected in the received packet data, e.g., as a result of a cyclic redundancy check (CRC) or the like.
  • CRC cyclic redundancy check
  • the descriptor 400 may include a flag (not shown) whose value indicates whether or not the descriptor's referenced packet data is part of a new data flow.
  • the buffer pointer 420 stores a value that indicates the memory location, e.g., in the memory 250, of the descriptor's corresponding data buffer 242.
  • Other status and configuration information 430 also may be included in the descriptor 400 to tailor the descriptor for a particular intermediate-node implementation.
  • the descriptor's other information 430 may include the amount of packet data stored in the descriptor's referenced data buffer 242.
  • Fig. 5 is a schematic block diagram of the illustrative HWA module 500 which is adapted to filter DoS packets before the packets can be forwarded to the CPU 230.
  • the HWA module 500 includes, among other things, a direct memory access (DMA) controller 510 and a flow classifier 520 that can collectively filter denial-of-service (DoS) traffic received at the intermediate network node 200. More specifically, the DMA controller 510 can access received packet data, e.g., stored in one or more data buffers 242 in the memory 250. The flow classifier 520 uses the received packet data and descriptors to identify a data flow associated with the packet. Unlike previous implementations, if the data flow identified by the flow classifier corresponds to a DoS data flow, then the DMA controller can "drop" (i.e., discard) the DoS packet before it is forwarded to the CPU 230. By filtering the DoS traffic in this manner, the CPU band- width and other processing resources are not consumed processing the DoS packets.
  • DMA direct memory access
  • DoS denial-of-service
  • the DMA controller 510 includes, inter alia, one or more free-buffer FIFOs 512, an ingress descriptor FIFO 514, a packet-header buffer 516, an egress descriptor FIFO 518. For every network interface 210 coupled to the system controller 300, the DMA controller 510 maintains a separate free-buffer FIFO 512. Each free-buffer FIFO stores a list of free buffer pointers 420 that have been allocated to its associated network interface. The CPU 230 may forward, via device bus 360, a predetermined number of free buffer pointers 420 to the DMA controller for storage in each interface's associated free-buffer FIFO 512.
  • the set of free-buffer pointers allocated to each network interface 210 corresponds to a respective set of data buffers 242 allocated for that interface.
  • the free buffer pointers 420 and their associated flag values 410 may be stored in ingress descriptors 400 which, in turn, may be stored in an ingress descriptor ring 352.
  • the ownership flag values 412 are set equal to values that indicate that the descriptors (and thus their corresponding free buffer pointers) are "owned" by the network interface.
  • the DMA controller may send an interrupt to the CPU 230 that requests additional free buffer pointers for that interface.
  • the low-water threshold value is preferably programmable and thus tunable for the system performance of the intermediate network node 200.
  • the CPU 230 responds to the interrupt by enqueueing the requested free- buffer pointers in the appropriate free-buffer FIFO 512.
  • the ingress descriptor FIFO 514 is adapted to store buffer descriptors retrieved from one or more of the ingress descriptor rings 352. Initially, descriptors are enqueued on the interface's ingress descriptor ring(s) 352, until each entry in the ingress descriptor ring contains a free buffer descriptor. Then, as packet data is received at the network interface, the interface forwards the received data to data buffers 242 refer- enced by the descriptors enqueued at the "head" of its ingress descriptor ring 352.
  • the network interface 210 modifies the ownership flag value 412 in the buffer's corresponding descriptor so as to transfer ownership of that descriptor to the DMA controller 510.
  • the interface also may modify other information, such as flag values 410, in the descriptor as well. This process is repeated until the end of the packet has been received, e.g., as indicated by an EOP flag value in the received data packet 160, or an error is detected.
  • the network interface 210 Having received the data packet 160, the network interface 210 sends an inter- rupt to the DMA controller 510 specifying which ingress descriptor ring 352 stores the packet's descriptors. For instance, each of the ingress rings 352 may be assigned a unique identification (ID) value that may be incorporated into the interrupt.
  • the controller 510 interacts with the device bus 360 to retrieve the packet's descriptors and store the descriptors in the ingress descriptor FIFO 514.
  • the DMA controller dequeues descriptors from the ingress ring 352, beginning with the descriptor whose SOP flag 414 value corresponds to the beginning of the data packet 160, until the packet's last descriptor has been dequeued, as indicated by the value of its EOP flag 416 or ERR flag 418.
  • the dequeued descriptors are then re-enqueued into appropriate entries in the ingress descriptor FIFO 514.
  • the DMA controller 510 replaces the dequeued descriptor with a free buffer descriptor containing a free buffer pointer obtained from the network interface's free-buffer FIFO 512.
  • the DMA controller 510 may also retrieve some (or all) of the received packet's header information, i.e., stored in the data buffers 242 referenced by the packet's descriptors. For example, information contained in the packet's data link, internetwork and/or transport layer headers may be retrieved by the controller and stored in the packet header buffer 516. In an illustrative embodiment, one or more selected packet headers are retrieved by the DMA controller in their entirety and stored in the packet header buffer 516.
  • the DMA controller 510 After retrieving the packet's descriptors and header information, the DMA controller 510 extracts the packet's descriptors from the ingress descriptor FIFO 514, the packet header information from the packet header buffer 516 and the ingress descriptor ring ID value from the received interrupt, and forwards this information to a packet- identifier engine 522 in the flow classifier 520.
  • the packet-identifier engine contains logic and circuitry configured to determine the type of packet 160 received by the net- work interface, based on the information forwarded from the DMA controller. For instance, based on this information, the engine 552 may determine whether the packet is formatted in accordance with a predetermined network protocol, such as the IGMP or ICMP protocol. Having identified the packet type, the packet-identifier engine further identifies a predetermined set of fields in the packet headers, descriptors and/or ingress ring ID value from which a set of "signature" information may be extracted.
  • the predetermined set of fields are forwarded from the packet-identifier engine 522 to a signature-extraction engine 524 in the flow classifier 520.
  • the signature- extraction engine is adapted to extract the signature information from the predetermined locations identified by the packet-identifier engine.
  • the signature information extracted by the engine 524 may include, among other things, source or destination TCP port numbers, source or destination IP addresses, protocol identifiers and so forth.
  • the extracted signature information is then input to a hash-entry address generator 530 in the flow classifier.
  • the hash-entry address generator includes a hash- function unit 532 that applies a predetermined hash function to the received signature information, thereby generating an «-bit resultant hash value.
  • the hash function unit 532 may be configured to apply other hash functions, such as the Message Digest 5 function, to the signa- ture information.
  • the hash value generated by the hash-function unit 532 may be forwarded to a bit-mask unit 534 in the address generator 530.
  • the m bits selected by the bit- mask unit may function as a hash-table index that uniquely identifies a specific entry in a hash table having 2 m entries.
  • the index may be converted to a memory address of its indexed hash-table entry, e.g., located in the memory 250. For example, assuming each hash-table entry is four bytes wide, the hash-table index times four may be added to the 5 hash table's base memory address 536 to derive the indexed hash-table entry's memory address.
  • the generated memory address is forwarded from the hash-entry address generator 530 to a linked-list walker 526.
  • the linked-list walker contains circuitry and logic for searching a linked list that begins at the forwarded hash-table entry memory
  • the linked-list walker extracts packet-related information, such as a data flow ID value and a destination egress ring ID value, from the matching entry.
  • the flow ID value corresponds to the packet's associated data flow and the des-
  • I 5 tination egress ring ID value corresponds to an egress descriptor ring 354 in which the packet's descriptors should be enqueued for further processing.
  • the packet's flow ID value corresponds to a DoS data flow
  • the packet's destination egress ring ID value corresponds to the delete ring 356.
  • linked-list walker (or another component of the HWA module 500) may modify the contents of one or more of the packet's descriptors to indicate that the received data packet 160 belongs to the new data flow.
  • the packet's descriptors are asso-
  • miss descriptor egress ring stores descriptors whose referenced packet data has been associated with a new data flow.
  • the linked-list walker 536 forwards the packet descriptors, flow ID value and 30 destination egress ring ID value to an egress packet manager 528.
  • the egress packet manager is responsible for reformatting the descriptors from an ingress descriptor for- mat to an egress descriptor format, if necessary. That is, the egress packet manager reformats the packet's descriptors if at least some of the ingress and egress descriptor rings 352-356 are configured to store descriptors having different formats. For example, the ingress and egress descriptor formats may include different flag values 410 or 5 other information 430.
  • the egress packet manager 528 then transfers the packet descriptors, flow ID value and destination egress ring ID value to an egress descriptor FIFO 518 in the DMA controller 510.
  • the DMA controller 510 When the DMA controller 510 identifies that the packet descriptors have been enqueued in the egress descriptor FIFO 518, the DMA controller forwards the descrip-
  • the DMA controller then sends the CPU 230 an interrupt indicating on which egress descriptor ring the descriptors are enqueued. Then, the CPU dequeues the packet descriptors from the appropriate destination egress ring 354 and renders a forwarding de-
  • the DMA controller identifies that the destination egress ring ID value associated with the packet descriptors corresponds to the delete egress ring 356, the DMA controller "recycles" the descriptors by removing the buffer pointers 420 within these descriptors and adding the removed buffer pointers to the free-buffer FIFO 512.
  • the packet 160 has been identified as belonging to a malicious data flow, such as a
  • DoS data flow, and its associated buffer pointers are returned to the network interface's pool of free buffer pointers rather than forwarding the packet's descriptors to the CPU 230. As such, CPU bandwidth and processing resources are not consumed processing the received DoS packet 160.
  • the delete egress ring 356 serves as an "overflow" queue for the free-buffer FIFO 512, thereby preventing possible stalling on the egress data path while the DMA controller waits for new entries to become available in the FIFO 512.
  • Buffer pointers stored in descriptors enqueued on the delete egress ring 356 later may be transferred to the appropriate free-buffer FIFO 512.
  • the CPU 230 may dequeue descriptors from the delete egress ring 356 when entries become available in the free-buffer FIFO 512.
  • the buffer pointers 420 in the dequeued descriptors then may be "recycled" by the CPU by enqueueing the buffer point- ers on the free-buffer FIFO.
  • the DMA controller 510 may transfer the descriptors from the delete egress ring to the free-buffer FIFO as entries become available in the FIFO.
  • Fig. 6 is a schematic block diagram of the hash table 600 configured to store a plurality of linked lists which may be searched by the linked-list walker 526.
  • the hash table contains 2 m hash-table entries 610, each of which is associated with a unique hash-table index 620 and is configured to store a list pointer value 630 referencing the memory location, e.g., in the memory 250, of a corresponding linked list.
  • the hash-table entries instead may be configured to directly store the first entry of their referenced linked lists.
  • a hash-table entry's list pointer value 630 may equal a predetermined "NULL" value if its referenced list does not contain any list entries 600, i.e., its referenced linked list is "empty.”
  • each "non-empty" linked list includes one or more entries 650 which store, inter alia, signature information 652 as well as a flow ID value 654 and a destination egress ring ID value 656.
  • Each destination egress ring ID value cor- responds to a respective one of the egress descriptor rings 354-356 in the on-chip memory 350. More specifically, a predetermined destination egress ring ID value 656 (e.g., "3”) corresponds to the delete egress ring 356 and the remaining destination egress ring ID values (e.g., "1" and "2”) correspond to respective egress descriptor rings 354.
  • the linked-list walker 526 locates a linked list in the hash table 600 using the list pointer 630 contained in the hash-table entry 610 whose memory address was generated by the hash-entry address generator 530. Then, the linked-list walker sequen- tially traverses ("walks") the list's linked-list entries 650 until it identifies a matching entry that contains the packet's signature information 652 or until the end of the list is reached.
  • Figure 7 is a flow chart illustrating a sequence of steps that may be performed in the intermediate network node 200 for identifying DoS packets before they can be forwarded to the CPU 230.
  • the sequence starts at step 700 and proceeds to step 704 where a packet data is received at a network interface 210.
  • the packet data is forwarded to the system controller 300, which in turn forwards the packet data to one or more buffers 242 in the memory 250, at step 712.
  • the network interface enqueues packet descriptors corresponding to the data buffers 242 storing the received packet data in an ingress descriptor ring 352 associated with the interface.
  • the network interface 210 determines whether the end of the packet 160 has been received. If the end of the packet has not been received, the sequence returns to step 704. Otherwise, at step 724, the network interface 210 sends an interrupt to the DMA controller 510 in the HWA module 500.
  • the interrupt includes, among other things, an ingress descriptor ring ID value that indicates which ingress descriptor ring 352 stores the received packet's descriptors.
  • the DMA controller retrieves the packet's descriptors from the ingress descriptor ring.
  • the DMA con- troller 510 replaces the dequeued descriptors with free buffer descriptors acquired from a free-buffer FIFO 512 associated with the network interface that received the packet.
  • the DMA controller also retrieves one or more of the received packet's headers, such as its data link, internetwork or transport layer headers.
  • the DMA controller 510 forwards the packet's descriptors, header data and the ingress descriptor ring ID value to a flow classifier 520 in the HWA module 500.
  • a packet-identifier engine 522 in the flow classifier identifies the type of data packet 160 received at the network interface 210.
  • signature information is extracted from a predetermined set of fields in the packet's descriptors and headers, based on the identified packet type.
  • the signature informa- tion may include TCP port number, IP addresses, protocol versions and so forth.
  • the extracted signature information is forwarded to a hash-entry address generator 530, in which a hash-function unit 532 calculates a hash of the signature infor- mation, e.g., using a CRC 32 hash function or the like.
  • the hash of the signature information is used to create an index in the hash table 600.
  • the index may be added to a hash-table base address 536 to generate the memory address, e.g., in the memory 250, of a hash-table entry 610.
  • the generated hash-entry address is forwarded to a linked-list walker 526 which traverses a linked list referenced by a list pointer 630 stored in the indexed hash-table entry 610. The list is searched until a matching list entry 650 is found containing the packet's extracted signature information 652 or the end of the list is reached.
  • the linked-list walker determines whether the received data packet 160 corresponds to a new data flow, i.e., the end of the list is reached without finding a matching linked-list entry 650.
  • the packet's descriptors are returned to the DMA controller 510 which forwards the descriptors to the CPU 230 for processing.
  • the CPU performs conventional routing operations for the received packet 160 in accordance with the router operating system 260.
  • the CPU 230 executing the router operating system may add a new linked-list entry 650 to the hash table 600 corresponding to the new data flow.
  • the added linked-list entry may specify a new flow ID value 654 that has been assigned to the flow by the operating system 260.
  • the linked-list walker 526 identifies a flow ID value 654 and a destination egress ring ID value 656 stored at predetermined locations in the matching list entry 650.
  • the linked-list walker then forwards the packet's descriptors to an egress packet manager 528 in the flow classifier 520 which reformats the descriptors, if necessary.
  • the egress packet manager forwards the packet's descriptors, flow ID value and destination egress ring ID value to an egress descriptor FIFO 518 in the DMA controller 510.
  • the DMA controller determines whether the destination egress ring ID value corresponds to the delete queue 356. If it does not, then at step 772, the controller 510 enqueues the packet's flow ID value 654 and descriptors on the destination egress ring 354 corresponding to the destination egress ring ID value 656. Then, at step 776, the DMA controller sends an inter- rupt to the CPU 230 to notify the CPU that the packet's descriptors and flow ID value have been enqueued on the egress descriptor ring 354.
  • step 796 the sequence ends at step 796. If, at step 768, the packet's associated destination egress ring ID value 656 corresponds to the delete queue 356, then at step 780, the DMA controller 510 determines whether there are enough available entries in the network interface's free-buffer FIFO 512 to store the buffer pointers 420 in the packet's descriptors. If so, the buffer pointers are removed from the descriptors and are enqueued on the free-buffer FIFO 512, at step 784; the sequence ends at step 796.
  • the DMA controller 510 enqueues the descriptors on the delete queue 356, at step 788.
  • the descriptors remain on the delete queue until their contained buffer pointers 420 can be transferred to the free-buffer FIFO, e.g., when entries become available in the FIFO, at step 792.
  • the CPU 230 executing the router operating system 260 may be responsible for dequeueing the descriptors from the delete queue 356 and transferring their buffer pointers to the free-buffer FIFO 512.
  • other hardware and/or software mechanisms may be used for this purpose. The sequence ends at step 796.
  • the novel technique for filtering ma- licious packets may be implemented in other network nodes besides intermediate network nodes.
  • the invention may be used in conjunction with conventional access control lists (ACL) and intrusion detection systems (IDS) to provide enhanced security against, e.g., DoS and DDoS attacks.
  • ACL access control lists
  • IDS intrusion detection systems
  • arbitrarily complex sets of signature information 652 can be stored in linked-list entries 650 whose associated data flows have been identified as malicious data flows. Identification of the malicious data flows may be determined manually, e.g., by a system administrator, and/or automatically, e.g., by the CPU 230 executing the router operating system 260.
  • ingress and egress descriptor rings 352-356 located in the on-chip memory 350
  • these descriptor rings alternatively may be stored in other memories in (or coupled to) the intermediate network node 200.
  • each network interface 210 is described having a corresponding ingress descriptor ring 352, it is also expressly contemplated that a network interface may be associated with more than one ingress descriptor ring, e.g., having different priority levels.
  • each destination such as the CPU 230, in the intermediate network node 200 is described having a corresponding egress descriptor ring 354, it is also expressly contemplated that a single destination in the node 200 may be associated with multiple egress descriptor rings, e.g., having different priority levels.
  • the signature information 652 associated with a received packet 160 is not limited to those values stored in fields of the packet's headers, e.g., and may be extracted from other portions of the packet's contents or other relevant packet information, such as which interface 210 received the packet. As described, the packet's extracted signature is compared with signature information 652 stored in the linked-list entries 650 until a matching list entry is located.
  • the linked-list entries alternatively may store the result of hashing the signature information 652.
  • a matching list entry is identified if its contained signature information 652 equals the result of hashing the packet's extracted signature information.
  • the technique is equally applicable for a plurality of different hash tables that are each configured as set forth in the illustrative embodiment.
  • a separate hash table 600 may be associated with each network interface 210 in the intermediate network node 200.
  • packets received at a particular network interface may be routed in accordance with flow ID values 654 stored in that network interface's associated hash table.
  • the hash table 600 may be replaced by a different searchable data structure, such as a search tree or the like, configured to store packets' signature information 652, flow ID values 654 and destination egress ring ID values 656.
  • the linked-list walker 526 is replaced with a search module that is adapted to search the searchable data structure.
  • teachings of this invention can be implemented as software, including a computer-readable medium having program instructions executing on a computer, hardware, firmware, or a combination thereof.
  • inventive technique therefore may be implemented in various combinations of hardware and/or software. Accordingly, this description is meant to be taken only by way of example and not to otherwise limit the scope of the invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Hardware Design (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

A system and method is provided for automatically identifying and removing malicious data packets, such as denial-of-service (DoS) packets, in an intermediate network node before the packets can be forwarded to a central processing unit (CPU) in the node. The CPU's processing bandwidth is therefore not consumed identifying and removing the malicious packets from the system memory. As such, processing of the malicious packets is essentially 'off-loaded' from the CPU, thereby enabling the CPU to process non-malicious packets in a more efficient manner. Unlike prior implementa­tions, the invention identifies malicious packets having complex encapsulations that can not be identified using traditional techniques, such as ternary content addressable memories (TCAM) or lookup tables.

Description

HARDWARE FILTERING SUPPORT FOR DENIAL-OF-SERVICE
ATTACKS
FIELD OF THE INVENTION
This invention relates generally to network communications, and, more specifi- cally, to a system and method for filtering denial-of-service traffic in an intermediate network node, such as a router.
BACKGROUND OF THE INVENTION
A computer network is a geographically distributed collection of interconnected subnetworks for transporting data between nodes, such as computers. A local area net- work (LAN) is an example of such a subnetwork. The network's topology is defined by an arrangement of client nodes that communicate with one another, typically through one or more intermediate network nodes, such as a router or switch. As used herein, a client node is an endstation node that is configured to originate or terminate communications over the network. In contrast, an intermediate network node is a node that facilitates routing data between client nodes. Communications between nodes are typically effected by exchanging discrete packets of data according to predefined protocols. In this context, a protocol consists of a set of rules defining how the nodes interact with each other.
Each data packet typically comprises "payload" data prepended ("encapsu- lated") by at least one network header formatted in accordance with a network communication protocol. The network headers include information that enables the client nodes and intermediate nodes to efficiently route the packet through the computer network. Often, a packet's network headers include at least a data-link (layer 2) header, an internetwork (layer 3) header and a transport (layer 4) header, as defined by the Open Systems Interconnection (OSI) Reference Model. The OSI Reference Model is generally described in more detail in Section 1.1 of the reference book entitled Interconnections Second Edition, by Radia Perlman, published September 1999, which is hereby incorporated by reference as though fully set forth herein. The data-link header provides information for transmitting the packet over a particular physical link (i.e., a communication medium), such as a point-to-point link, Ethernet link, wireless link, optical link, etc. To that end, the data-link header may specify a pair of "source" and "destination" network interfaces that are connected by the physical link. A network interface contains the mechanical, electrical and signaling circuitry and logic used to couple a network node to one or more physical links. A network interface is often associated with a hardware-specific address, known as a media access control (MAC) address. Accordingly, the source and destination network interfaces in the data-link header are typically represented as source and destination MAC addresses. The data-link header may also store flow control, frame synchronization and error checking information used to manage data transmissions over the physical link.
The internetwork header provides information defining the packet's logical path (or "virtual circuit") through the computer network. Notably, the path may span multi- pie physical links. The internetwork header may be formatted according to the Internet Protocol (IP), which specifies IP addresses of both a source and destination node at the end points of the logical path. Thus, the packet may "hop" from node to node along its logical path until it reaches the client node assigned to the destination IP address stored in the packet's internetwork header. After each hop, the source and destination MAC addresses in the packet's data-link header may be updated, as necessary. However, the source and destination IP addresses typically remain unchanged as the packet is transferred from link to link in the network.
The transport header provides information for ensuring that the packet is reliably transmitted from the source node to the destination node. The transport header typically includes, among other things, source and destination port numbers that respectively identify particular software applications executing in the source and destination nodes. More specifically, the packet is generated in the source node by the application assigned to the source port number. Then, the packet is forwarded to the destination node and directed to the application assigned to the destination port number. The transport header also may include error-checking information (i.e., a checksum) and other data-flow control information. For instance, in connection-oriented transport pro- tocols such as the Transmission Control Protocol (TCP), the transport header may store sequencing information that indicates the packet's relative position in a transmitted stream of data packets.
As used herein, a dataflow is a stream of data packets that is communicated from a source node to a destination node. Each packet in the flow satisfies a set of predetermined criteria, e.g., based on the packet's contents, size or relative position (i.e., temporal or spatial) in the data flow. An intermediate network node may be configured to perform "flow-based" routing operations so as to route each packet in a data flow in the same manner. The intermediate node typically receives data packets in the flow and forwards the packets in accordance with predetermined routing information that is distributed using a protocol, such as the Open Shortest Path First (OSPF) protocol. Because each packet in the flow is addressed to the same destination node, the intermediate node need only perform one forwarding decision for the entire data flow, e.g., based on the first packet received in the flow. Thereafter, the intermediate node forwards packets in the data flow based on the flow's previously determined routing information (i.e., adjacency information). In this way, the intermediate node consumes fewer resources, such as processor bandwidth and processing time, than if it performed a separate forwarding decision for every packet in the data flow.
In practice, the intermediate network node may implement a hash table which stores packet-related information used to classify received packets into their corresponding data flows. The hash table is typically organized as a table of linked lists, where each list may be indexed by the result of applying a conventional hash function to "signature" information. In this context, a signature is a set of values that remain constant for every packet in a data flow. For example, assume each packet in a first data flow stores the same pair of source and destination IP address values. In this case, a signature for the first data flow may be generated based on the values of these source and destination IP addresses. Likewise, a different signature may be generated for a second data flow whose packets store a different set of source and destination IP addresses than packets in the first data flow. Of course, those skilled in the art will appre- ciate that a data flow's signature information is not limited to IP addresses and may include other information, such as TCP port numbers, IP version numbers and so forth. - A -
Each linked list in the hash table contains one or more entries, and each linked- list entry stores information corresponding to a particular data flow. Such information may include, inter alia, the data flow's associated signature information and a dataflow identifier ("flow ID"). The flow ID identifies the particular data flow and also may be used to locate routing information associated with the data flow. To that end, the intermediate network node may maintain a data structure that maps flow ID values to the memory locations of their corresponding routing information, e.g., stored in the node's local or internal memory. Alternatively, the flow ID values may directly incorporate the memory locations of their data flows' routing information. When a packet is received by the intermediate network node, signature information is extracted from the packet's network headers and hashed using a conventional hash function, such as a cyclic redundancy check (CRC) function. The resultant hash value is used to index a hash-table entry which, in turn, references a linked list. Entries in the linked list are accessed sequentially until a "matching" entry is found storing the extracted signature. When a matching linked-list entry is located, the entry's stored flow ID value is used to associate the received packet with a data flow and the packet is routed in accordance with that flow.
The intermediate network node typically receives a large number of data flows from various sources, including client nodes and other intermediate nodes. Each source may be responsible for establishing one or more data flows with the intermediate node. To optimize use of its processing bandwidth, the intermediate node may process the received flows on a prioritized basis. That is, as packets are received at the intermediate node, they are identified as belonging to, e.g., a high or low priority data flow. Packets in the high-priority flow may be processed by the intermediate node in advance of the low-priority packets, even if the low-priority packets were received before the high-priority packets.
Denial-of-service (DoS) attacks have become fairly common techniques for disabling access to resources and/or services in an intermediate network node. A DoS attack corresponds to a data flow of "malicious" packets which, when processed by the intermediate network node, deprive non-malicious packets (i.e., non-DoS packets) access to certain resources and/or services in the node. The DoS packets may be sent from a single source or may be coordinated among a plurality of sources. This latter case is often referred to as a distributed DoS (DDoS) attack. For example, a computer hacker may launch a DDoS attack through a multitude of compromised endstations that transmit data packets to a target intermediate node, thereby overwhelming the interme- diate node's processing bandwidth.
DoS attacks typically involve sending large quantities of a specific type of network traffic, such as packets formatted in accordance with the Internet Control Message Protocol (ICMP) or the Internet Group Management Protocol (IGMP), to the intermediate network node. In many cases, the DoS packets are pre-pended by a complex arrangement of network headers. Thus, the targeted intermediate network node becomes overburdened not only by the large quantity of received DoS packets, but also by the consumption of resources required to process them. Since the intermediate node's resources become overly consumed processing these malicious packets, other non-malicious packets sent to the intermediate node are often dropped or discarded. Accordingly, different types of intermediate network nodes attempt to prevent DoS attacks in various ways.
The high-end "core" routers and switches typically have enough processing bandwidth to process both malicious DoS packets as well as non-malicious packets. In this context, the high-end routers and switches are designed to handle large amounts of network traffic, e.g., on a network "backbone." Consequently, the malicious packets can be processed at the rate at which they are received, i.e., at "line" rate. These high- end intermediate nodes thus rely primarily on hardware forwarding solutions that cannot become over-subscribed while identifying and removing the malicious DoS packets. As a result, a substantial portion of processing bandwidth in the central processing unit(s) (CPU) executing the software may be consumed identifying and removing DoS packets. Another disadvantage of this solution is that the routing or switching software becomes more complex by the inclusion of code for filtering the DoS packets from the received data flows.
The "mid-range" routers and switches, unlike their high-end counterparts, typi- cally become oversubscribed as a result of a DoS attack. These intermediate nodes are usually enterprise or LAN routers/switches that manage a relatively large number of data flows. In order to identify and remove DoS traffic (i.e., data packets), the mid- range routers and switches typically utilize software executing on a centralized CPU or on a network processor supporting a general-purpose CPU. Like the software in the high-end routers and switches, the software in the mid-range routers and switches con- sumes an excessive amount of processing bandwidth and complexity for thwarting the DoS attack.
Hardware support for prioritization of ingress traffic is sometimes implemented in the mid-range routers and switches when the "problem" DoS traffic can be filtered and put on a low-priority queue serviced by the software. However, because the num- ber of malicious packets typically becomes exorbitant during the DoS attack, the low- priority queue usually fills and is therefore tail-dropped, dropping both the malicious DoS traffic as well as non-malicious low-priority traffic. Furthermore, the hardware filtering is typically implemented as a simple table lookup on data link (layer 2) or internetwork (layer 3) information contained in the received data packets. The table lookup may be performed using a content addressable memory (CAM), such as a ternary CAM (TCAM). If the DoS attack traffic arrives via a complicated encapsulation, this table-based filtering cannot support these encapsulations and the DoS traffic is then forwarded to the software executing on the CPU. As a result, the hardware support does not prevent the CPU from being burdened with processing the DoS traffic. The low-end "access" routers and switches are typically single CPU systems that process a relatively small amount of network traffic and are therefore more susceptible to a DoS attack than the aboveηnoted mid-range and high-end intermediate network nodes. There is only one CPU in the low-end routers and switches, so CPU bandwidth is typically not consumed pre-processing or prioritizing in-coming data packets. Such pre-processing would require the software executing on the CPU to process each received packet twice (prioritize and route) thus consuming an unacceptable amount of processing resources. Therefore, the low-end routers and switches usually only filter received data packets (if at all) using simple lookup tables or TCAMs that are not able to identify complex DoS packet encapsulations. There is generally a need for an intermediate network node that can identify and remove DoS traffic without consuming an excessive amount of processing resources or bandwidth in the node. Further, the intermediate node should identify and remove DoS traffic having encapsulations of any arbitrary complexity. In addition, the malicious DoS packets should be removed from the intermediate node without affecting the processing of non-malicious packets, such as low-priority non-malicious packets.
SUMMARY OF THE INVENTION
The present invention overcomes the disadvantages of the prior art by providing a system and method for automatically identifying and removing malicious data packets, such as denial-of-service (DoS) packets, in an intermediate network node before the packets can be forwarded to a central processing unit (CPU) in the node. The CPU's processing bandwidth is therefore not consumed identifying and removing the malicious packets from the system memory. As such, processing of the malicious packets is essentially "off-loaded" from the CPU, thereby enabling the CPU to process non- malicious packets in a more efficient manner. Unlike prior implementations, the invention identifies malicious packets having complex encapsulations that can not be identi- fied using traditional techniques, such as TCAMs or lookup tables.
Advantageously, the invention may be employed in an intermediate network node configured to perform as a high-end, mid-range or access router. Further, the invention may be used in conjunction with conventional access control lists (ACL) and/or intrusion detection systems (IDS) in order to delete DoS traffic whose encapsulations are too complex to be filtered by the ACL or IDS. More generally, the invention may be configured, either automatically or manually, to filter malicious packets having encapsulations of any arbitrary complexity. The malicious data packets are preferably identified using hash-based flow identification, although other flow classification techniques may be employed. The invention is illustratively implemented in a hardware assist device in the intermediate network node. However, it is also expressly contemplated that the invention may be embodied by other combinations of hardware and/or software. BRIEF DESCRIPTION OF THE DRAWINGS
The above and further advantages of the invention may be better understood by referring to the following description in conjunction with the accompanying drawings in which like reference numerals indicate identically or functionally similar elements, of which:
Fig. 1 is a schematic block diagram of a computer network comprising a collection of interconnected subnetworks and nodes, including an intermediate network node;
Fig. 2 is a schematic block diagram of an illustrative intermediate network node that may be used in accordance with the present invention; Fig. 3 is a schematic block diagram of a system controller that may be implemented in the illustrative intermediate network node in Fig. 2;
Fig. 4 is a schematic block diagram of an exemplary buffer descriptor that may be enqueued in the ingress and/or egress descriptor rings of the present invention;
Fig. 5 is a schematic block diagram of an illustrative hardware assist (HWA) module that may be used to automatically identify and remove malicious data packets in accordance with the present invention;
Fig. 6 is a schematic block diagram of an exemplary hash table that may be searched by the HWA module for identifying a data flow associated with a received data packet; and Figs. 7A-B are a flow chart illustrating a sequence of steps that may be performed for automatically identifying and removing malicious data packets in accordance with the present invention.
DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS
Fig. 1 is a block diagram of a computer network 100 comprising a collection of interconnected subnetworks and nodes. The nodes may comprise computers including end nodes 130 and 140, such as a sending end node 120 and a receiving end node 150, and an intermediate network node 200, the latter of which may be a switch or router. The subnetworks 105, 110 included within network 100 are preferably local area networks (LANs) interconnected by the intermediate node 200, although the networks may comprise other communication links, such as wide area networks. Communication among the nodes coupled to the LANs is typically effected by exchanging discrete packets 160 among the nodes.
For example, the sending node 120 generates a data packet 160 by encapsulating "payload" data within headers, such as conventional data link and internetwork s headers, as the data passes through different layers of a protocol stack. The packet is then transmitted over the network to the intermediate node 200 which facilitates the flow of the data packet through the network by routing it to the proper receiving node 150. Specifically, the node 200 receives the packet at one of its network interfaces and renders a forwarding decision for the packet based on a destination end node specified io by the packet's internetwork header. The packet's data link header is modified in accordance with the forwarding decision and the packet is transmitted over an appropriate subnetwork coupled to the intermediate network node.
Fig. 2 is a schematic block diagram of an intermediate node 200 that may be advantageously used with the present invention. The node comprises a plurality of is network interfaces 210, a system controller 300, a central processing unit (CPU) 230 and a memory 250. Data is received by the network interfaces 210, each of which is coupled to at least one network or subnetwork, such as LANs 105 and 110. The network interfaces contain the mechanical, electrical and signaling circuitry that enables the intermediate network node 200 to communicate over physical links connected to
20 networks and subnetworks, including, inter alia, asynchronous transfer mode (ATM)
^ networks, synchronous optical networks (SONET), wireless networks, frame relay networks, Ethernet networks, Fiber Distributed Data Interface (FDDI) networks, etc.
The system controller 300 is coupled to each network interface 210, the CPU 230 (i.e., a processor) and the memory 250 by different local buses in the intermediate
25 network node 200. For instance, the system controller may be coupled to the network interfaces 210 by respective peripheral component interconnect (PCI) buses, whereas the controller may be coupled to the memory 250 by a plurality of high-speed connections, such as HyperTransport bus links. The controller 300 therefore functions as a "bridge" for transferring data from one local bus to another. That is, the controller re-
30 ceives data over a first local bus, e.g., coupled to a network interface 210, and converts the data to a format that may be transmitted over a second local bus, e.g., coupled to the memory 250. The system controller may also include other functionality, such as ap- plication-specific circuitry or logic. Illustratively, the controller 300 may be embodied in hardware as a field-programmable gate array (FPGA) or an application-specific integrated circuit (ASIC), although the controller's functionality alternatively may be implemented in various combinations of hardware and/or software. The memory 250 comprises a plurality of storage locations that are addressable by the CPU 230 and the network interfaces 210 via the system controller 300. The memory comprises a form of random access memory (RAM) that is generally cleared by a power cycle or other reboot operation (e.g., it is a "volatile" memory). For instance, the memory 250 may comprise dynamic random access memory (DRAM) and/or synchronous DRAM (SDRAM) storage locations adapted to store program code and data structures accessible to the CPU 230. It will be apparent to those skilled in the art that the memory 250 may also comprise other memory means, including various computer-readable media, for storing program instructions and data structures pertaining to the operation of the intermediate network node 200. A router operating system 260, portions of which are typically resident in the memory 250 and executed by the CPU 230, functionally organizes the intermediate network node 200 by, inter alia, invoking network operations in support of software processes executing on the intermediate node. The IOS™ operating system by Cisco Systems, me. is one example of a router operating system 260. The operating system may perform routing operations on data packets 160 received by the network interfaces 210. A portion of the memory 250 may be organized as a buffer "pool" 240 of data buffers 242 configured to store received data packets. For example, each buffer may be configured to store up to a fixed amount, e.g., 2 kilobytes, of the received data packet 160.
Operationally, a received packet 160 is transferred from a network interface 210 to one or more of the buffers 242. The CPU 230 executing the router operating system 260 renders a forwarding decision for the received packet based on routing information 270 stored in the memory 250. One or more data structures, such as the hash table 600, may be stored in the memory to facilitate the operating system's forwarding decision. For example, the hash table 600 may be used to identify a data flow associated with the received packet, and the routing information 270 may store adjacency information associated with the identified flow. In this case, the packet's network headers are modi- fied in accordance with the adjacency information associated with the packet's identified data flow.
Fig. 3 is a schematic block diagram of the system controller 300 that may be implemented in the illustrative intermediate network node 200. The system controller comprises a plurality of first local bus (PCI) interfaces 310, a memory controller 320, a CPU bus interface 330, a bus controller 340, an on-chip memory 350 and a hardware assist (HWA) module 500 interconnected by a device bus 360. In an illustrative embodiment, each PCI interface 310 includes circuitry and logic configured to send and receive data over a PCI bus coupled to a network interface 210. However, the PCI in- terfaces 310 alternatively may be substituted by respective controllers that communicate over other types of buses, such as Industry Standard Architecture (ISA) buses, Extended ISA (EISA) buses, etc. Data received at a network interface 210 is forwarded over a PCI bus to a PCI interface 310, which frames the received data so it may be transferred over the device bus 360. Conversely, the PCI interface may receive data from the bus 360 and reformat the data for transmission over the PCI bus.
The memory controller 320 comprises circuitry and logic configured to transfer data from the memory 250 over the second local bus to the device bus 360, and vice versa. For instance, the CPU 230 may forward a memory address (or range of addresses) to the CPU bus interface 330. The memory address may be accompanied by a CPU instruction to read or write data at that memory address. The CPU bus interface 330 transmits the memory address and its corresponding CPU instruction over the device bus 360 to the memory controller 320. In response, the memory controller writes or retrieves data at the specified memory address, in accordance with the CPU instruction. The bus controller 340 comprises circuitry and logic that, inter alia, implements an arbitration policy for coordinating access to the device bus 360. That is, the controller 340 prevents two or more entities, such as the PCI interfaces 310, memory controller 320, etc., from attempting to access the bus 360 at substantially the same time. To that end, the bus controller 340 may be configured to grant or deny access to the bus 360 based on a predefined arbitration protocol. According to the illustrative embodiment, one or more functions normally performed by the CPU 230 executing the router operating system 260 may be "off-loaded" to the HWA module 500. For instance, the HWA module may be configured to filter denial-of-service (DoS) packets received at the intermediate network node 200, before the packets can be processed by the CPU. As such, the processing bandwidth of the CPU 230 is not consumed identifying and removing "malicious" DoS traffic, as in previous implementations. Further, as discussed in more detail below, the HWA module is configured to identify DoS packets having arbitrarily complex encapsulations that otherwise can not be identified using traditional techniques, such as TCAMs or lookup ta- bles.
The on-chip memory 350 comprises a set of addressable memory locations resident on the system controller 300. The on-chip memory may be a form of volatile memory, such as static RAM (SRAM), or a form of erasable non- volatile memory, such as Flash memory. Although the illustrative on-chip memory 350 is situated in the sys- tern controller 300, it is also expressly contemplated that the on-chip memory may reside in a separate memory module coupled to the system controller, or the contents of the on-chip memory (or a portion thereof) may be incorporated into the "main" memory 250.
The on-chip memory 350 stores, among other things, one or more "ingress" de- scriptor rings 352 (i.e., circular fϊrst-in, first-out (FIFO) queues), one or more "egress" descriptor rings 354 and a "delete" descriptor ring 356. Each network interface 210 is associated with at least one ingress ring 352 in the on-chip memory 350. When packet data is received at a network interface 210, the packet data is forwarded over an appropriate PCI bus, through the system controller 300, to an available data buffer 242 in the memory 250. A memory reference (i.e., a "descriptor") to the data buffer is then enqueued in the ingress descriptor ring 352 associated with the network interface 210 that received the packet data. Packet data is stored and descriptors are enqueued in this manner until the network interface 210 determines that an entire packet 160 has been received or an error has occurred. Accordingly, the network interface's ingress ring 352 stores an ordered list of descriptors corresponding to the order in which the packet data is received at the interface 210. After the packet 160 has been received, the network interface 210 notifies the HWA module 500 which ingress descriptor ring 352 contains the received packet's descriptors. The HWA module dequeues the descriptors and processes the received packet data to determine the packet's associated data flow. In accordance with the il- lustrative embodiment, the HWA module performs hash-based flow classification that enables it to classify packets having arbitrarily complex encapsulations. Having identified the packet's associated data flow, the HWA module enqueues the packet's descriptors in an appropriate egress descriptor ring 354, selected based on the identified data flow. The CPU 230 is notified, e.g., by a hardware interrupt, when the packet's de- scriptors have been enqueued on one of its egress descriptor rings 354. In response, the CPU dequeues the descriptors from its egress descriptor ring and renders a forwarding decision for the received data packet 160.
Illustratively, each of the egress descriptor rings 354 corresponds to a different destination and/or priority level that may be associated with the identified data flow. For instance, among the egress rings 354 may be separate high-priority and low-priority egress descriptor rings, respectively corresponding to high-priority and low-priority data flows processed by the CPU 230. Moreover, in a multi-CPU implementation, a set of one or more high-priority and/or low-priority egress descriptor rings 354 may be associated with each CPU 230 in the plurality of CPUs. If the HWA module 500 determines that the received data packet 160 is a DoS packet or other type of malicious packet, e.g., based on the packet's identified data flow, the HWA module enqueues the packet's descriptors on the delete egress ring 356 or on a first in, first out queue (not shown) of "free" buffer descriptors, i.e., descriptors whose referenced buffers 242 are available to store new packet data, associated with the network interface 210. Advantageously, descriptors placed on the delete ring 356 or on the free-buffer FIFO are "recycled" (i.e., reused) before they can be forwarded to the CPU 230. As such, the data buffers 242 storing the malicious DoS packet data can be reclaimed by the HWA module 500 without having to consume the CPU's bandwidth or processing resources. Fig. 4 illustrates an exemplary descriptor 400 that may be enqueued in at least one of the ingress and egress descriptor rings 352-356. The descriptor contains, inter alia, a set of flags 410, a buffer pointer 420 as well as other information 430 associated with the descriptor. The flags 410 may include, for example, an ownership flag 412, a start-of-packet (SOP) flag 414, an end-of-packet (EOP) flag 416 and an error (ERR) flag 418. The ownership flag 412 stores a value that indicates which device is granted ownership (i.e., control) of the descriptor 400. For instance, the ownership flag may equal a first value to indicate that a network interface 210 is given exclusive access to the descriptor 400; the flag 412 may equal a second value when control of the descriptor is transferred to the HWA module 500.
The SOP flag 414 may be set equal to a value that indicates whether the de- scriptor's referenced data buffer 242 contains the start of the received data packet 160. Similarly, the EOP flag 416 may be set equal to a value that indicates whether the descriptor's referenced data buffer contains the end of the packet 160. The ERR flag 418 may be set equal to a value indicating whether an error has been detected in the received packet data, e.g., as a result of a cyclic redundancy check (CRC) or the like. Of course, other flag values besides those explicitly shown may be included in the set of flags 410. For instance, the descriptor 400 may include a flag (not shown) whose value indicates whether or not the descriptor's referenced packet data is part of a new data flow.
The buffer pointer 420 stores a value that indicates the memory location, e.g., in the memory 250, of the descriptor's corresponding data buffer 242. Other status and configuration information 430 also may be included in the descriptor 400 to tailor the descriptor for a particular intermediate-node implementation. For example, the descriptor's other information 430 may include the amount of packet data stored in the descriptor's referenced data buffer 242. Fig. 5 is a schematic block diagram of the illustrative HWA module 500 which is adapted to filter DoS packets before the packets can be forwarded to the CPU 230. The HWA module 500 includes, among other things, a direct memory access (DMA) controller 510 and a flow classifier 520 that can collectively filter denial-of-service (DoS) traffic received at the intermediate network node 200. More specifically, the DMA controller 510 can access received packet data, e.g., stored in one or more data buffers 242 in the memory 250. The flow classifier 520 uses the received packet data and descriptors to identify a data flow associated with the packet. Unlike previous implementations, if the data flow identified by the flow classifier corresponds to a DoS data flow, then the DMA controller can "drop" (i.e., discard) the DoS packet before it is forwarded to the CPU 230. By filtering the DoS traffic in this manner, the CPU band- width and other processing resources are not consumed processing the DoS packets.
The DMA controller 510 includes, inter alia, one or more free-buffer FIFOs 512, an ingress descriptor FIFO 514, a packet-header buffer 516, an egress descriptor FIFO 518. For every network interface 210 coupled to the system controller 300, the DMA controller 510 maintains a separate free-buffer FIFO 512. Each free-buffer FIFO stores a list of free buffer pointers 420 that have been allocated to its associated network interface. The CPU 230 may forward, via device bus 360, a predetermined number of free buffer pointers 420 to the DMA controller for storage in each interface's associated free-buffer FIFO 512. The set of free-buffer pointers allocated to each network interface 210 corresponds to a respective set of data buffers 242 allocated for that interface. The free buffer pointers 420 and their associated flag values 410 may be stored in ingress descriptors 400 which, in turn, may be stored in an ingress descriptor ring 352. As such, the ownership flag values 412 are set equal to values that indicate that the descriptors (and thus their corresponding free buffer pointers) are "owned" by the network interface. Notably, if the number of free buffer pointers 420 enqueued in a network interface's free-buffer FIFO 512 becomes less than a predetermined "low- water" level, the DMA controller may send an interrupt to the CPU 230 that requests additional free buffer pointers for that interface. The low-water threshold value is preferably programmable and thus tunable for the system performance of the intermediate network node 200. The CPU 230 responds to the interrupt by enqueueing the requested free- buffer pointers in the appropriate free-buffer FIFO 512.
The ingress descriptor FIFO 514 is adapted to store buffer descriptors retrieved from one or more of the ingress descriptor rings 352. Initially, descriptors are enqueued on the interface's ingress descriptor ring(s) 352, until each entry in the ingress descriptor ring contains a free buffer descriptor. Then, as packet data is received at the network interface, the interface forwards the received data to data buffers 242 refer- enced by the descriptors enqueued at the "head" of its ingress descriptor ring 352. When a data buffer 242 becomes filled, or the end of the packet is received, the network interface 210 modifies the ownership flag value 412 in the buffer's corresponding descriptor so as to transfer ownership of that descriptor to the DMA controller 510. Of course, the interface also may modify other information, such as flag values 410, in the descriptor as well. This process is repeated until the end of the packet has been received, e.g., as indicated by an EOP flag value in the received data packet 160, or an error is detected.
Having received the data packet 160, the network interface 210 sends an inter- rupt to the DMA controller 510 specifying which ingress descriptor ring 352 stores the packet's descriptors. For instance, each of the ingress rings 352 may be assigned a unique identification (ID) value that may be incorporated into the interrupt. In response to receiving the interrupt, the controller 510 interacts with the device bus 360 to retrieve the packet's descriptors and store the descriptors in the ingress descriptor FIFO 514. Operationally, the DMA controller dequeues descriptors from the ingress ring 352, beginning with the descriptor whose SOP flag 414 value corresponds to the beginning of the data packet 160, until the packet's last descriptor has been dequeued, as indicated by the value of its EOP flag 416 or ERR flag 418. The dequeued descriptors are then re-enqueued into appropriate entries in the ingress descriptor FIFO 514. For every descriptor dequeued from the ingress descriptor ring 352, the DMA controller 510 replaces the dequeued descriptor with a free buffer descriptor containing a free buffer pointer obtained from the network interface's free-buffer FIFO 512.
Additionally, the DMA controller 510 may also retrieve some (or all) of the received packet's header information, i.e., stored in the data buffers 242 referenced by the packet's descriptors. For example, information contained in the packet's data link, internetwork and/or transport layer headers may be retrieved by the controller and stored in the packet header buffer 516. In an illustrative embodiment, one or more selected packet headers are retrieved by the DMA controller in their entirety and stored in the packet header buffer 516. After retrieving the packet's descriptors and header information, the DMA controller 510 extracts the packet's descriptors from the ingress descriptor FIFO 514, the packet header information from the packet header buffer 516 and the ingress descriptor ring ID value from the received interrupt, and forwards this information to a packet- identifier engine 522 in the flow classifier 520. The packet-identifier engine contains logic and circuitry configured to determine the type of packet 160 received by the net- work interface, based on the information forwarded from the DMA controller. For instance, based on this information, the engine 552 may determine whether the packet is formatted in accordance with a predetermined network protocol, such as the IGMP or ICMP protocol. Having identified the packet type, the packet-identifier engine further identifies a predetermined set of fields in the packet headers, descriptors and/or ingress ring ID value from which a set of "signature" information may be extracted.
The predetermined set of fields are forwarded from the packet-identifier engine 522 to a signature-extraction engine 524 in the flow classifier 520. The signature- extraction engine is adapted to extract the signature information from the predetermined locations identified by the packet-identifier engine. For example, the signature information extracted by the engine 524 may include, among other things, source or destination TCP port numbers, source or destination IP addresses, protocol identifiers and so forth.
The extracted signature information is then input to a hash-entry address generator 530 in the flow classifier. The hash-entry address generator includes a hash- function unit 532 that applies a predetermined hash function to the received signature information, thereby generating an «-bit resultant hash value. For example, the hash function may be a conventional CRC-32 hash function that generates a 32-bit hash value (i.e., «=32). In alternate embodiments, the hash function unit 532 may be configured to apply other hash functions, such as the Message Digest 5 function, to the signa- ture information.
The hash value generated by the hash-function unit 532 may be forwarded to a bit-mask unit 534 in the address generator 530. The bit-mask unit selects m bits of the n received hash bits. For example, suppose the hash-function unit 532 generates a 32- bit hash value («=32). In this case, the bit-mask unit 534 may be configured to select eight (m=8) predetermined bits of a 32-bit hash value by ANDing this 32-bit value with a "mask" value equal to OxOOOOFFOO (in hexadecimal). The m bits selected by the bit- mask unit may function as a hash-table index that uniquely identifies a specific entry in a hash table having 2m entries. The index may be converted to a memory address of its indexed hash-table entry, e.g., located in the memory 250. For example, assuming each hash-table entry is four bytes wide, the hash-table index times four may be added to the 5 hash table's base memory address 536 to derive the indexed hash-table entry's memory address.
The generated memory address is forwarded from the hash-entry address generator 530 to a linked-list walker 526. The linked-list walker contains circuitry and logic for searching a linked list that begins at the forwarded hash-table entry memory
I0 address, until a linked-list entry is located containing the packet's extracted signature information or the end of the list is reached. If a "matching" list entry is found containing the signature information, the linked-list walker extracts packet-related information, such as a data flow ID value and a destination egress ring ID value, from the matching entry. The flow ID value corresponds to the packet's associated data flow and the des-
I5 tination egress ring ID value corresponds to an egress descriptor ring 354 in which the packet's descriptors should be enqueued for further processing. However, in accordance with the illustrative embodiment, if the packet's flow ID value corresponds to a DoS data flow, then the packet's destination egress ring ID value corresponds to the delete ring 356.
20 In the event that a matching linked-list entry is not found, the linked-list walker
526 may identify the received packet 160 as belonging to a new data flow. In this case, linked-list walker (or another component of the HWA module 500) may modify the contents of one or more of the packet's descriptors to indicate that the received data packet 160 belongs to the new data flow. In addition, the packet's descriptors are asso-
25 ciated with a destination egress ring ID value corresponding to a predetermined "miss descriptor" egress ring 354, e.g., in the on-chip memory 350. The miss descriptor egress ring stores descriptors whose referenced packet data has been associated with a new data flow.
The linked-list walker 536 forwards the packet descriptors, flow ID value and 30 destination egress ring ID value to an egress packet manager 528. The egress packet manager is responsible for reformatting the descriptors from an ingress descriptor for- mat to an egress descriptor format, if necessary. That is, the egress packet manager reformats the packet's descriptors if at least some of the ingress and egress descriptor rings 352-356 are configured to store descriptors having different formats. For example, the ingress and egress descriptor formats may include different flag values 410 or 5 other information 430. The egress packet manager 528 then transfers the packet descriptors, flow ID value and destination egress ring ID value to an egress descriptor FIFO 518 in the DMA controller 510.
When the DMA controller 510 identifies that the packet descriptors have been enqueued in the egress descriptor FIFO 518, the DMA controller forwards the descrip-
10 tors and their associated flow ID value, via the device bus 360, to the egress descriptor ring 356 corresponding to the descriptors' associated destination egress ring ID value. The DMA controller then sends the CPU 230 an interrupt indicating on which egress descriptor ring the descriptors are enqueued. Then, the CPU dequeues the packet descriptors from the appropriate destination egress ring 354 and renders a forwarding de-
I5 cision for the packet 160, in accordance with its identified data flow. When the CPU dequeues descriptors from the miss descriptor egress ring 354, the descriptors' referenced packet data is processed by the CPU as a new data flow, and a new linked-list entry 650 may be added for the data flow at an appropriate location in the hash-table 600.
20 If the DMA controller identifies that the destination egress ring ID value associated with the packet descriptors corresponds to the delete egress ring 356, the DMA controller "recycles" the descriptors by removing the buffer pointers 420 within these descriptors and adding the removed buffer pointers to the free-buffer FIFO 512. In this case, the packet 160 has been identified as belonging to a malicious data flow, such as a
25 DoS data flow, and its associated buffer pointers are returned to the network interface's pool of free buffer pointers rather than forwarding the packet's descriptors to the CPU 230. As such, CPU bandwidth and processing resources are not consumed processing the received DoS packet 160.
If the destination egress ring ID value corresponds to the delete egress ring 356 30 and there are not enough available entries in the free-buffer FIFO 512 to store the packet's buffer pointers 420, e.g., the CPU 230 has recently filled the FIFO 512 with free buffer pointers, then the packet's descriptors instead may be stored directly in the delete egress ring 356. As such, the delete egress ring 356 serves as an "overflow" queue for the free-buffer FIFO 512, thereby preventing possible stalling on the egress data path while the DMA controller waits for new entries to become available in the FIFO 512. Buffer pointers stored in descriptors enqueued on the delete egress ring 356 later may be transferred to the appropriate free-buffer FIFO 512. For instance, the CPU 230 may dequeue descriptors from the delete egress ring 356 when entries become available in the free-buffer FIFO 512. In this case, the buffer pointers 420 in the dequeued descriptors then may be "recycled" by the CPU by enqueueing the buffer point- ers on the free-buffer FIFO. Alternatively, the DMA controller 510 may transfer the descriptors from the delete egress ring to the free-buffer FIFO as entries become available in the FIFO.
Fig. 6 is a schematic block diagram of the hash table 600 configured to store a plurality of linked lists which may be searched by the linked-list walker 526. The hash table contains 2m hash-table entries 610, each of which is associated with a unique hash-table index 620 and is configured to store a list pointer value 630 referencing the memory location, e.g., in the memory 250, of a corresponding linked list. Alternatively, rather than store list pointer values 630, the hash-table entries instead may be configured to directly store the first entry of their referenced linked lists. A hash-table entry's list pointer value 630 may equal a predetermined "NULL" value if its referenced list does not contain any list entries 600, i.e., its referenced linked list is "empty."
On the other hand, each "non-empty" linked list includes one or more entries 650 which store, inter alia, signature information 652 as well as a flow ID value 654 and a destination egress ring ID value 656. Each destination egress ring ID value cor- responds to a respective one of the egress descriptor rings 354-356 in the on-chip memory 350. More specifically, a predetermined destination egress ring ID value 656 (e.g., "3") corresponds to the delete egress ring 356 and the remaining destination egress ring ID values (e.g., "1" and "2") correspond to respective egress descriptor rings 354. In operation, the linked-list walker 526 locates a linked list in the hash table 600 using the list pointer 630 contained in the hash-table entry 610 whose memory address was generated by the hash-entry address generator 530. Then, the linked-list walker sequen- tially traverses ("walks") the list's linked-list entries 650 until it identifies a matching entry that contains the packet's signature information 652 or until the end of the list is reached.
Figure 7 is a flow chart illustrating a sequence of steps that may be performed in the intermediate network node 200 for identifying DoS packets before they can be forwarded to the CPU 230. The sequence starts at step 700 and proceeds to step 704 where a packet data is received at a network interface 210. At step 708, the packet data is forwarded to the system controller 300, which in turn forwards the packet data to one or more buffers 242 in the memory 250, at step 712. At step 716, the network interface enqueues packet descriptors corresponding to the data buffers 242 storing the received packet data in an ingress descriptor ring 352 associated with the interface.
At step 720 the network interface 210 determines whether the end of the packet 160 has been received. If the end of the packet has not been received, the sequence returns to step 704. Otherwise, at step 724, the network interface 210 sends an interrupt to the DMA controller 510 in the HWA module 500. The interrupt includes, among other things, an ingress descriptor ring ID value that indicates which ingress descriptor ring 352 stores the received packet's descriptors. In response to receiving the interrupt, the DMA controller retrieves the packet's descriptors from the ingress descriptor ring. Notably, upon dequeueing descriptors from the ingress descriptor ring, the DMA con- troller 510 replaces the dequeued descriptors with free buffer descriptors acquired from a free-buffer FIFO 512 associated with the network interface that received the packet. The DMA controller also retrieves one or more of the received packet's headers, such as its data link, internetwork or transport layer headers. At step 728, the DMA controller 510 forwards the packet's descriptors, header data and the ingress descriptor ring ID value to a flow classifier 520 in the HWA module 500.
At step 732, a packet-identifier engine 522 in the flow classifier identifies the type of data packet 160 received at the network interface 210. At step 736, signature information is extracted from a predetermined set of fields in the packet's descriptors and headers, based on the identified packet type. For example, the signature informa- tion may include TCP port number, IP addresses, protocol versions and so forth. At step 740, the extracted signature information is forwarded to a hash-entry address generator 530, in which a hash-function unit 532 calculates a hash of the signature infor- mation, e.g., using a CRC 32 hash function or the like. The hash of the signature information is used to create an index in the hash table 600. At step 744, the index may be added to a hash-table base address 536 to generate the memory address, e.g., in the memory 250, of a hash-table entry 610. At step 748, the generated hash-entry address is forwarded to a linked-list walker 526 which traverses a linked list referenced by a list pointer 630 stored in the indexed hash-table entry 610. The list is searched until a matching list entry 650 is found containing the packet's extracted signature information 652 or the end of the list is reached. At step 752, the linked-list walker determines whether the received data packet 160 corresponds to a new data flow, i.e., the end of the list is reached without finding a matching linked-list entry 650. If the received packet is determined to belong to a new data flow, then, at step 756, the packet's descriptors are returned to the DMA controller 510 which forwards the descriptors to the CPU 230 for processing. At step 757, the CPU performs conventional routing operations for the received packet 160 in accordance with the router operating system 260. In this case, the CPU 230 executing the router operating system may add a new linked-list entry 650 to the hash table 600 corresponding to the new data flow. The added linked-list entry may specify a new flow ID value 654 that has been assigned to the flow by the operating system 260.
On the other hand, if at step 748 a linked-list entry with a matching signature is located, the linked-list walker 526 identifies a flow ID value 654 and a destination egress ring ID value 656 stored at predetermined locations in the matching list entry 650. The linked-list walker then forwards the packet's descriptors to an egress packet manager 528 in the flow classifier 520 which reformats the descriptors, if necessary. Next, at step 764, the egress packet manager forwards the packet's descriptors, flow ID value and destination egress ring ID value to an egress descriptor FIFO 518 in the DMA controller 510.
At step 768, after the descriptors, flow ID value and destination egress ring ID value have been enqueued in the egress descriptor FIFO 518, the DMA controller determines whether the destination egress ring ID value corresponds to the delete queue 356. If it does not, then at step 772, the controller 510 enqueues the packet's flow ID value 654 and descriptors on the destination egress ring 354 corresponding to the destination egress ring ID value 656. Then, at step 776, the DMA controller sends an inter- rupt to the CPU 230 to notify the CPU that the packet's descriptors and flow ID value have been enqueued on the egress descriptor ring 354. The sequence ends at step 796. If, at step 768, the packet's associated destination egress ring ID value 656 corresponds to the delete queue 356, then at step 780, the DMA controller 510 determines whether there are enough available entries in the network interface's free-buffer FIFO 512 to store the buffer pointers 420 in the packet's descriptors. If so, the buffer pointers are removed from the descriptors and are enqueued on the free-buffer FIFO 512, at step 784; the sequence ends at step 796. In contrast, if there are not enough available entries in the free-buffer FIFO 512 at step 780, then the DMA controller 510 enqueues the descriptors on the delete queue 356, at step 788. The descriptors remain on the delete queue until their contained buffer pointers 420 can be transferred to the free-buffer FIFO, e.g., when entries become available in the FIFO, at step 792. The CPU 230 executing the router operating system 260 may be responsible for dequeueing the descriptors from the delete queue 356 and transferring their buffer pointers to the free-buffer FIFO 512. However, those skilled in the art will appreciate that other hardware and/or software mechanisms may be used for this purpose. The sequence ends at step 796.
The foregoing has been a detailed description of illustrative embodiments of the invention. Various modifications and additions can be made without departing from the spirit and scope of the invention. For example, the novel technique for filtering ma- licious packets, such as DoS packets, may be implemented in other network nodes besides intermediate network nodes. Further, the invention may be used in conjunction with conventional access control lists (ACL) and intrusion detection systems (IDS) to provide enhanced security against, e.g., DoS and DDoS attacks. More specifically, the flexibility of the hash-based flow classification described herein enables detection of complex packet encapsulations that can not be detected using traditional ACL or IDS approaches. Thus, arbitrarily complex sets of signature information 652 can be stored in linked-list entries 650 whose associated data flows have been identified as malicious data flows. Identification of the malicious data flows may be determined manually, e.g., by a system administrator, and/or automatically, e.g., by the CPU 230 executing the router operating system 260.
While the illustrative embodiments describe the ingress and egress descriptor rings 352-356 located in the on-chip memory 350, those skilled in the art will appreci- ate that these descriptor rings alternatively may be stored in other memories in (or coupled to) the intermediate network node 200. In addition, although each network interface 210 is described having a corresponding ingress descriptor ring 352, it is also expressly contemplated that a network interface may be associated with more than one ingress descriptor ring, e.g., having different priority levels. Similarly, while each destination, such as the CPU 230, in the intermediate network node 200 is described having a corresponding egress descriptor ring 354, it is also expressly contemplated that a single destination in the node 200 may be associated with multiple egress descriptor rings, e.g., having different priority levels. It is also noted that the signature information 652 associated with a received packet 160 is not limited to those values stored in fields of the packet's headers, e.g., and may be extracted from other portions of the packet's contents or other relevant packet information, such as which interface 210 received the packet. As described, the packet's extracted signature is compared with signature information 652 stored in the linked-list entries 650 until a matching list entry is located. However, it is also contemplated that the linked-list entries alternatively may store the result of hashing the signature information 652. In this case, a matching list entry is identified if its contained signature information 652 equals the result of hashing the packet's extracted signature information. Although the inventive technique is described in terms of a single hash table
600, the technique is equally applicable for a plurality of different hash tables that are each configured as set forth in the illustrative embodiment. For instance, a separate hash table 600 may be associated with each network interface 210 in the intermediate network node 200. As such, packets received at a particular network interface may be routed in accordance with flow ID values 654 stored in that network interface's associated hash table. Moreover, the hash table 600 may be replaced by a different searchable data structure, such as a search tree or the like, configured to store packets' signature information 652, flow ID values 654 and destination egress ring ID values 656. In such an embodiment, the linked-list walker 526 is replaced with a search module that is adapted to search the searchable data structure. It is expressly contemplated that the teachings of this invention can be implemented as software, including a computer-readable medium having program instructions executing on a computer, hardware, firmware, or a combination thereof. The inventive technique therefore may be implemented in various combinations of hardware and/or software. Accordingly, this description is meant to be taken only by way of example and not to otherwise limit the scope of the invention.
What is claimed is:

Claims

CLAIMS 1. A method for a network node, which includes a central processing unit (CPU) configured to execute a router operating system, to filter malicious data packets re- ceived at the network node, the method comprising: receiving a data packet at the network node; performing hash-based flow classification on the received data packet to deter- mine whether the received data packet is a malicious data packet; and discarding the received data packet before the data packet can be forwarded to the CPU for processing by the router operating system, if the received data packet is determined to be a malicious data packet.
2. The method of claim 1, wherein the step of performing hash-based flow classi- fication further comprises: identifying a packet type associated with the received data packet; extracting a set of signature information corresponding to the identified packet type; and searching a hash table to locate the extracted set of signature information.
3. The method of claim 2, further comprising: configuring the hash table, either manually or automatically, to associate the set of signature information with a data flow; and determining whether the data flow associated with the set of signature informa- tion corresponds to a malicious data flow.
4. The method of claim 1, further comprising: associating the received data packet with a destination in the network node as a result of the hash-based flow classification.
5. The method of claim 4, further comprising: determining whether the destination associated with the received data packet is a predetermined destination associated with malicious data packets.
6. The method of claim 5, further comprising: in response to determining that the destination associated with the received data packet is the predetermined destination, performing the steps of: removing buffer pointers from a set of descriptors associated with the received data packet; and storing the removed buffer pointers on a queue of free buffer pointers.
7. The method of claim 6, further comprising: if the queue of free buffer pointers does not contain enough available entries to store the removed buffer pointers, storing the set of descriptors associated with the re- ceived data packet on a delete queue until enough entries become available in the queue of free buffer pointers.
8. The method of claim 6, further comprising: transferring free buffer pointers from the router operating system to the queue of free buffer pointers.
9. The method of claim 1, wherein the step of performing hash-based flow classi- fication is used in conjunction with an access control list or an intrusion detection sys- tern.
10. The method of claim 1 , wherein the network node is an intermediate network node.
11. A network node, comprising: a central processing unit (CPU) configured to execute instructions that imple- ment a router operating system; a network interface adapted to receive a data packet; a memory having a plurality of storage locations addressable by the CPU, the storage locations being configured to store: (i) at least a portion of the router operating system instructions, (ii) one or more data buffers for storing the received data packet, and (iii) a searchable data structure configured to store information as- soci- ated with the received data packet; and a system controller coupled to the memory and the CPU, the system controller including a hardware assist (HWA) module configured to discard malicious data pack- ets from the network node before the malicious data packets can be forwarded to the CPU for processing by the router operating system.
12. The network node of claim 11 , wherein the searchable data structure is a hash table.
13. The network node of claim 11 , wherein the HWA module includes a direct memory access (DMA) controller and a flow classifier.
14. The network node of claim 13, wherein the DMA controller includes: an ingress descriptor first in, first out (FIFO) queue configured to store a set of descriptors referencing the one or more data buffers in which the received data packet is stored; a packet-header buffer configured to store information contained in at least one packet header prepended to the received data packet; an egress descriptor FIFO configured to store the set of descriptors as well as a data flow identification (ID) value for identifying the data flow associated with the re- ceived data packet and a destination value for identifying a destination in the network node associated with the received data packet, the flow classifier searching the search- able data structure to locate the data flow ID value and the destination value; and a free-buffer FIFO containing a set of free buffer descriptors allocated for the network interface.
15. The network node of claim 13 , wherein the flow classifier includes : a packet-identifier engine configured to identify a packet type associated with the received data packet based on information received from the DMA controller; a signature-extraction engine configured to extract a set of signature information from a predetermined set of fields in the information received from the DMA control- ler, the predetermined set of fields being selected based on the packet type identified by the packet-identifier engine; an address generator configured to generate a memory address based on the set of signature information, the memory address corresponding to an entry in the search- able data structure; and a search module configured to search the searchable data structure to locate a flow ID value and a destination value associated with the received data packet.
16. The network node of claim 15, wherein the flow classifier further includes: an egress packet manager configured to reformat descriptors from an ingress descriptor format to an egress descriptor format.
17. The network node of claim 11 , wherein the network node is an intermediate net- work node.
18. A network node including a central processing unit (CPU) configured to exe- cute a router operating system, the network node comprising: means for receiving a data packet at the network node; means for performing hash-based flow classification on the received data packet to determine whether the received data packet is a malicious data packet; and means for discarding the received data packet before the data packet can be for- warded to the CPU for processing by the router operating system, if the received data packet is determined to be a malicious data packet.
19. A computer-readable media including instructions for execution by a processor, the instructions for a method of filtering malicious data packets received at a network node in which a central processing unit (CPU) is configured to execute a router operat- ing system, the method comprising: receiving a data packet at the network node; performing hash-based flow classification on the received data packet to deter- mine whether the received data packet is a malicious data packet; and discarding the received data packet before the data packet can be forwarded to the CPU for processing by the router operating system, if the received data packet is determined to be a malicious data packet.
PCT/US2005/007059 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks WO2006124009A2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP05857867.5A EP1754349B1 (en) 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks
CA2559251A CA2559251C (en) 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks
CN2005800090740A CN101421991B (en) 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks
AU2005326185A AU2005326185A1 (en) 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/811,195 US7411957B2 (en) 2004-03-26 2004-03-26 Hardware filtering support for denial-of-service attacks
US10/811,195 2004-03-26

Publications (2)

Publication Number Publication Date
WO2006124009A2 true WO2006124009A2 (en) 2006-11-23
WO2006124009A3 WO2006124009A3 (en) 2009-04-16

Family

ID=34989731

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/007059 WO2006124009A2 (en) 2004-03-26 2005-03-03 Hardware filtering support for denial-of-service attacks

Country Status (6)

Country Link
US (1) US7411957B2 (en)
EP (1) EP1754349B1 (en)
CN (1) CN101421991B (en)
AU (1) AU2005326185A1 (en)
CA (1) CA2559251C (en)
WO (1) WO2006124009A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101184095B (en) * 2007-12-06 2011-09-21 中兴通讯股份有限公司 Network anti-attack method and system based on strategy control listing of CPU
CN104394150A (en) * 2014-11-26 2015-03-04 大连梯耐德网络技术有限公司 System and method for implementing mimic security network architecture based on hardware reconfiguration
WO2022084625A1 (en) * 2020-10-22 2022-04-28 Orange Methods and devices for protecting a stream of packets

Families Citing this family (139)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8095508B2 (en) 2000-04-07 2012-01-10 Washington University Intelligent data storage and processing using FPGA devices
US7200105B1 (en) * 2001-01-12 2007-04-03 Bbn Technologies Corp. Systems and methods for point of ingress traceback of a network attack
US10572824B2 (en) 2003-05-23 2020-02-25 Ip Reservoir, Llc System and method for low latency multi-functional pipeline with correlation logic and selectively activated/deactivated pipelined data processing engines
EP2528000B1 (en) 2003-05-23 2017-07-26 IP Reservoir, LLC Intelligent data storage and processing using FPGA devices
US7602785B2 (en) * 2004-02-09 2009-10-13 Washington University Method and system for performing longest prefix matching for network address lookup using bloom filters
US7200692B2 (en) * 2004-03-10 2007-04-03 Cisco Technology, Inc. PVDM (packet voice data module) generic bus
US7117308B1 (en) * 2004-04-06 2006-10-03 Cisco Technology, Inc. Hypertransport data path protocol
US7111092B1 (en) * 2004-04-16 2006-09-19 Cisco Technology, Inc. Buffer management technique for a hypertransport data path protocol
US7669240B2 (en) * 2004-07-22 2010-02-23 International Business Machines Corporation Apparatus, method and program to detect and control deleterious code (virus) in computer network
US7626940B2 (en) * 2004-12-22 2009-12-01 Intruguard Devices, Inc. System and method for integrated header, state, rate and content anomaly prevention for domain name service
US7602731B2 (en) * 2004-12-22 2009-10-13 Intruguard Devices, Inc. System and method for integrated header, state, rate and content anomaly prevention with policy enforcement
US7600057B2 (en) * 2005-02-23 2009-10-06 Broadcom Corporation Method and system for configurable drain mechanism in two-way handshake system
JP2008532177A (en) 2005-03-03 2008-08-14 ワシントン ユニヴァーシティー Method and apparatus for performing biological sequence similarity searches
JP5008270B2 (en) * 2005-04-13 2012-08-22 ソニー株式会社 Information processing apparatus and information processing method
US7693050B2 (en) * 2005-04-14 2010-04-06 Microsoft Corporation Stateless, affinity-preserving load balancing
US20070011732A1 (en) * 2005-07-05 2007-01-11 Yang-Hung Peng Network device for secure packet dispatching via port isolation
US7580351B2 (en) * 2005-07-12 2009-08-25 Cisco Technology, Inc Dynamically controlling the rate and internal priority of packets destined for the control plane of a routing device
US7522521B2 (en) * 2005-07-12 2009-04-21 Cisco Technology, Inc. Route processor adjusting of line card admission control parameters for packets destined for the route processor
GB0518578D0 (en) * 2005-09-13 2005-10-19 Qinetiq Ltd Communications systems firewall
US7702629B2 (en) 2005-12-02 2010-04-20 Exegy Incorporated Method and device for high performance regular expression pattern matching
US7954114B2 (en) 2006-01-26 2011-05-31 Exegy Incorporated Firmware socket module for FPGA-based pipeline processing
US7668107B2 (en) * 2006-03-22 2010-02-23 Marvell Israel (M.I.S.L.) Ltd. Hardware implementation of network testing and performance monitoring in a network device
US8379841B2 (en) 2006-03-23 2013-02-19 Exegy Incorporated Method and system for high throughput blockwise independent encryption/decryption
US7636703B2 (en) * 2006-05-02 2009-12-22 Exegy Incorporated Method and apparatus for approximate pattern matching
US8453147B2 (en) * 2006-06-05 2013-05-28 Cisco Technology, Inc. Techniques for reducing thread overhead for systems with multiple multi-threaded processors
US8194662B2 (en) * 2006-06-08 2012-06-05 Ilnickl Slawomir K Inspection of data
US8041929B2 (en) 2006-06-16 2011-10-18 Cisco Technology, Inc. Techniques for hardware-assisted multi-threaded processing
US7921046B2 (en) 2006-06-19 2011-04-05 Exegy Incorporated High speed processing of financial information using FPGA devices
US7840482B2 (en) 2006-06-19 2010-11-23 Exegy Incorporated Method and system for high speed options pricing
US7773507B1 (en) 2006-06-30 2010-08-10 Extreme Networks, Inc. Automatic tiered services based on network conditions
US7817549B1 (en) * 2006-06-30 2010-10-19 Extreme Networks, Inc. Flexible flow-aging mechanism
US20080030903A1 (en) * 2006-08-03 2008-02-07 Sae Magnetics (H.K.) Ltd. Suspension having stress absorbing structure, head gimbal assembly and disk drive unit with the same
US8010966B2 (en) * 2006-09-27 2011-08-30 Cisco Technology, Inc. Multi-threaded processing using path locks
US7660793B2 (en) 2006-11-13 2010-02-09 Exegy Incorporated Method and system for high performance integration, processing and searching of structured and unstructured data using coprocessors
US8326819B2 (en) 2006-11-13 2012-12-04 Exegy Incorporated Method and system for high performance data metatagging and data indexing using coprocessors
JP4734223B2 (en) * 2006-11-29 2011-07-27 アラクサラネットワークス株式会社 Traffic analyzer and analysis method
KR100864888B1 (en) * 2007-02-12 2008-10-22 삼성전자주식회사 Routing System and Method for Managing Rule Entry of Ternary Content Addressable Memory
US8320249B2 (en) * 2007-03-07 2012-11-27 Broadcom Corporation Method and system for controlling network access on a per-flow basis
US20080240140A1 (en) * 2007-03-29 2008-10-02 Microsoft Corporation Network interface with receive classification
US8325633B2 (en) * 2007-04-26 2012-12-04 International Business Machines Corporation Remote direct memory access
US20080279190A1 (en) * 2007-05-10 2008-11-13 Raveendra Muniyappa Maintaining End-to-End Packet Ordering
US8037213B2 (en) * 2007-05-30 2011-10-11 International Business Machines Corporation Replenishing data descriptors in a DMA injection FIFO buffer
US8018951B2 (en) * 2007-07-12 2011-09-13 International Business Machines Corporation Pacing a data transfer operation between compute nodes on a parallel computer
US8478834B2 (en) 2007-07-12 2013-07-02 International Business Machines Corporation Low latency, high bandwidth data communications between compute nodes in a parallel computer
US8312541B2 (en) * 2007-07-17 2012-11-13 Cisco Technology, Inc. Detecting neighbor discovery denial of service attacks against a router
US8959172B2 (en) 2007-07-27 2015-02-17 International Business Machines Corporation Self-pacing direct memory access data transfer operations for compute nodes in a parallel computer
US20090031001A1 (en) * 2007-07-27 2009-01-29 Archer Charles J Repeating Direct Memory Access Data Transfer Operations for Compute Nodes in a Parallel Computer
EP2186250B1 (en) 2007-08-31 2019-03-27 IP Reservoir, LLC Method and apparatus for hardware-accelerated encryption/decryption
US7916728B1 (en) 2007-09-28 2011-03-29 F5 Networks, Inc. Lockless atomic table update
US20090119292A1 (en) * 2007-11-06 2009-05-07 Barracuda Inc. Peer to peer traffic control method and system
US9225545B2 (en) * 2008-04-01 2015-12-29 International Business Machines Corporation Determining a path for network traffic between nodes in a parallel computer
US9009350B2 (en) * 2008-04-01 2015-04-14 International Business Machines Corporation Determining a path for network traffic between nodes in a parallel computer
US8374986B2 (en) 2008-05-15 2013-02-12 Exegy Incorporated Method and system for accelerated stream processing
WO2009139170A1 (en) * 2008-05-16 2009-11-19 パナソニック株式会社 Attack packet detector, attack packet detection method, image receiver, content storage device, and ip communication device
US7948979B2 (en) * 2008-05-28 2011-05-24 Intel Corporation Programmable network interface card
GB2462493B (en) * 2008-08-13 2012-05-16 Gnodal Ltd Data processing
US7855967B1 (en) * 2008-09-26 2010-12-21 Tellabs San Jose, Inc. Method and apparatus for providing line rate netflow statistics gathering
US7966620B2 (en) * 2008-11-07 2011-06-21 Microsoft Corporation Secure network optimizations when receiving data directly in a virtual machine's memory address space
US20120095893A1 (en) 2008-12-15 2012-04-19 Exegy Incorporated Method and apparatus for high-speed processing of financial market depth data
US8880696B1 (en) 2009-01-16 2014-11-04 F5 Networks, Inc. Methods for sharing bandwidth across a packetized bus and systems thereof
US8880632B1 (en) 2009-01-16 2014-11-04 F5 Networks, Inc. Method and apparatus for performing multiple DMA channel based network quality of service
US8103809B1 (en) 2009-01-16 2012-01-24 F5 Networks, Inc. Network devices with multiple direct memory access channels and methods thereof
US8112491B1 (en) 2009-01-16 2012-02-07 F5 Networks, Inc. Methods and systems for providing direct DMA
EP2413547A4 (en) * 2009-03-26 2014-12-24 Nec Corp Route setting server, route setting method, and route setting program
US9313047B2 (en) 2009-11-06 2016-04-12 F5 Networks, Inc. Handling high throughput and low latency network data packets in a traffic management device
US20110145572A1 (en) * 2009-12-15 2011-06-16 Christensen Kenneth J Apparatus and method for protecting packet-switched networks from unauthorized traffic
US8544026B2 (en) * 2010-02-09 2013-09-24 International Business Machines Corporation Processing data communications messages with input/output control blocks
US9286298B1 (en) 2010-10-14 2016-03-15 F5 Networks, Inc. Methods for enhancing management of backup data sets and devices thereof
US8949453B2 (en) 2010-11-30 2015-02-03 International Business Machines Corporation Data communications in a parallel active messaging interface of a parallel computer
US9148376B2 (en) 2010-12-08 2015-09-29 AT&T Intellectual Property I, L.L.P. Method and system for dynamic traffic prioritization
JP6045505B2 (en) 2010-12-09 2016-12-14 アイピー レザボア, エルエルシー.IP Reservoir, LLC. Method and apparatus for managing orders in a financial market
CN102055627B (en) * 2011-01-04 2012-06-13 深信服网络科技(深圳)有限公司 Method and device for identifying peer-to-peer (P2P) application connection
US10135831B2 (en) 2011-01-28 2018-11-20 F5 Networks, Inc. System and method for combining an access control system with a traffic management system
US20120327956A1 (en) * 2011-06-24 2012-12-27 Altior Inc. Flow compression across multiple packet flows
US8949328B2 (en) 2011-07-13 2015-02-03 International Business Machines Corporation Performing collective operations in a distributed processing system
WO2013044248A1 (en) * 2011-09-23 2013-03-28 Futurewei Technologies, Inc. Method and apparatus to increase forwarding silicon functionality through packet manipulation
US8976647B2 (en) * 2011-11-08 2015-03-10 Futurewei Technologies, Inc. Hardware-based dynamic load balancing that avoids flow packet reordering statistically
US8930962B2 (en) 2012-02-22 2015-01-06 International Business Machines Corporation Processing unexpected messages at a compute node of a parallel computer
US9015852B2 (en) 2012-04-30 2015-04-21 Cisco Technology, Inc. Protecting address resolution protocol neighbor discovery cache against denial of service attacks
US10033837B1 (en) 2012-09-29 2018-07-24 F5 Networks, Inc. System and method for utilizing a data reducing module for dictionary compression of encoded data
US9519501B1 (en) * 2012-09-30 2016-12-13 F5 Networks, Inc. Hardware assisted flow acceleration and L2 SMAC management in a heterogeneous distributed multi-tenant virtualized clustered system
US10102260B2 (en) 2012-10-23 2018-10-16 Ip Reservoir, Llc Method and apparatus for accelerated data translation using record layout detection
US9633093B2 (en) 2012-10-23 2017-04-25 Ip Reservoir, Llc Method and apparatus for accelerated format translation of data in a delimited data format
EP2912579B1 (en) 2012-10-23 2020-08-19 IP Reservoir, LLC Method and apparatus for accelerated format translation of data in a delimited data format
US9270602B1 (en) 2012-12-31 2016-02-23 F5 Networks, Inc. Transmit rate pacing of large network traffic bursts to reduce jitter, buffer overrun, wasted bandwidth, and retransmissions
US10375155B1 (en) 2013-02-19 2019-08-06 F5 Networks, Inc. System and method for achieving hardware acceleration for asymmetric flow connections
US9554418B1 (en) 2013-02-28 2017-01-24 F5 Networks, Inc. Device for topology hiding of a visited network
US20140369363A1 (en) * 2013-06-18 2014-12-18 Xpliant, Inc. Apparatus and Method for Uniquely Enumerating Paths in a Parse Tree
US9864606B2 (en) 2013-09-05 2018-01-09 F5 Networks, Inc. Methods for configurable hardware logic device reloading and devices thereof
WO2015095000A1 (en) 2013-12-16 2015-06-25 F5 Networks, Inc. Methods for facilitating improved user authentication using persistent data and devices thereof
US9825884B2 (en) 2013-12-30 2017-11-21 Cavium, Inc. Protocol independent programmable switch (PIPS) software defined data center networks
WO2015164639A1 (en) 2014-04-23 2015-10-29 Ip Reservoir, Llc Method and apparatus for accelerated data translation
US10015143B1 (en) 2014-06-05 2018-07-03 F5 Networks, Inc. Methods for securing one or more license entitlement grants and devices thereof
US10616380B2 (en) 2014-06-19 2020-04-07 Cavium, Llc Method of handling large protocol layers for configurable extraction of layer information and an apparatus thereof
US9635146B2 (en) 2014-06-19 2017-04-25 Cavium, Inc. Method of using bit vectors to allow expansion and collapse of header layers within packets for enabling flexible modifications and an apparatus thereof
US9742694B2 (en) 2014-06-19 2017-08-22 Cavium, Inc. Method of dynamically renumbering ports and an apparatus thereof
US9531848B2 (en) 2014-06-19 2016-12-27 Cavium, Inc. Method of using generic modification instructions to enable flexible modifications of packets and an apparatus thereof
US9497294B2 (en) 2014-06-19 2016-11-15 Cavium, Inc. Method of using a unique packet identifier to identify structure of a packet and an apparatus thereof
US9516145B2 (en) 2014-06-19 2016-12-06 Cavium, Inc. Method of extracting data from packets and an apparatus thereof
US9473601B2 (en) 2014-06-19 2016-10-18 Cavium, Inc. Method of representing a generic format header using continuous bytes and an apparatus thereof
US9628385B2 (en) 2014-06-19 2017-04-18 Cavium, Inc. Method of identifying internal destinations of networks packets and an apparatus thereof
US9888033B1 (en) * 2014-06-19 2018-02-06 Sonus Networks, Inc. Methods and apparatus for detecting and/or dealing with denial of service attacks
US9531849B2 (en) 2014-06-19 2016-12-27 Cavium, Inc. Method of splitting a packet into individual layers for modification and intelligently stitching layers back together after modification and an apparatus thereof
US9961167B2 (en) 2014-06-19 2018-05-01 Cavium, Inc. Method of modifying packets to a generic format for enabling programmable modifications and an apparatus thereof
US10050833B2 (en) 2014-06-19 2018-08-14 Cavium, Inc. Method of reducing latency in a flexible parser and an apparatus thereof
US9438703B2 (en) 2014-06-19 2016-09-06 Cavium, Inc. Method of forming a hash input from packet contents and an apparatus thereof
US9729439B2 (en) 2014-09-26 2017-08-08 128 Technology, Inc. Network packet flow controller
WO2016055668A1 (en) * 2014-10-10 2016-04-14 Keelwit Laiseca Internet Security Solutions, S.L. Device and method for protection in communication networks
US9606781B2 (en) * 2014-11-14 2017-03-28 Cavium, Inc. Parser engine programming tool for programmable network devices
US9864582B2 (en) * 2014-11-14 2018-01-09 Cavium, Inc. Code processor to build orthogonal execution blocks for programmable network devices
CN105786733B (en) * 2014-12-26 2020-08-07 南京中兴新软件有限责任公司 Method and device for writing TCAM (ternary content addressable memory) entries
US9736184B2 (en) 2015-03-17 2017-08-15 128 Technology, Inc. Apparatus and method for using certificate data to route data
US10305819B2 (en) 2015-05-13 2019-05-28 Cisco Technology, Inc. Dynamic protection of shared memory used by output queues in a network device
US9866401B2 (en) * 2015-05-13 2018-01-09 Cisco Technology, Inc. Dynamic protection of shared memory and packet descriptors used by output queues in a network device
CN106302318A (en) 2015-05-15 2017-01-04 阿里巴巴集团控股有限公司 A kind of website attack defense method and device
US9729682B2 (en) 2015-05-18 2017-08-08 128 Technology, Inc. Network device and method for processing a session using a packet signature
US9800524B1 (en) * 2015-09-24 2017-10-24 Qlogic Corporation Switching methods and systems for a network interface card
US10942943B2 (en) 2015-10-29 2021-03-09 Ip Reservoir, Llc Dynamic field data translation to support high performance stream data processing
US9973528B2 (en) 2015-12-21 2018-05-15 Fortinet, Inc. Two-stage hash based logic for application layer distributed denial of service (DDoS) attack attribution
CN105591989B (en) * 2016-01-25 2019-12-20 盛科网络(苏州)有限公司 Chip implementation method for uploading protocol message to CPU
TWI618387B (en) * 2016-02-24 2018-03-11 Method for improving packet processing of virtual switch
US9916269B1 (en) * 2016-04-14 2018-03-13 Amazon Technologies, Inc. Packet queueing for network device
US10009282B2 (en) * 2016-06-06 2018-06-26 128 Technology, Inc. Self-protecting computer network router with queue resource manager
CN106131050B (en) * 2016-08-17 2022-12-09 裴志永 Data packet fast processing system
WO2018119035A1 (en) 2016-12-22 2018-06-28 Ip Reservoir, Llc Pipelines for hardware-accelerated machine learning
US10972453B1 (en) 2017-05-03 2021-04-06 F5 Networks, Inc. Methods for token refreshment based on single sign-on (SSO) for federated identity environments and devices thereof
US10212089B1 (en) * 2017-09-21 2019-02-19 Citrix Systems, Inc. Encapsulating traffic entropy into virtual WAN overlay for better load balancing
US10516695B1 (en) * 2017-09-26 2019-12-24 Amazon Technologies, Inc. Distributed denial of service attack mitigation in service provider systems
US11005884B2 (en) * 2017-09-29 2021-05-11 Intel Corporation Denial of service mitigation with two-tier hash
US11855898B1 (en) 2018-03-14 2023-12-26 F5, Inc. Methods for traffic dependent direct memory access optimization and devices thereof
EP3618389B1 (en) 2018-08-27 2020-10-07 Ovh Systems and methods for operating a networking device
PL3618355T3 (en) * 2018-08-27 2021-02-08 Ovh Systems and methods for operating a networking device
US11537716B1 (en) 2018-11-13 2022-12-27 F5, Inc. Methods for detecting changes to a firmware and devices thereof
US11681625B2 (en) * 2018-12-20 2023-06-20 Intel Corporation Receive buffer management
US11500737B2 (en) * 2019-05-21 2022-11-15 Mellanox Technologies, Ltd. Coherent capturing of shared-buffer status
CN112181755B (en) * 2019-07-02 2024-05-10 腾讯科技(深圳)有限公司 Method for determining parameter threshold in search service and related equipment
CN112882833B (en) * 2021-02-09 2022-06-17 广州思林杰科技股份有限公司 Data acquisition method and device, computer equipment and storage medium
US11962615B2 (en) 2021-07-23 2024-04-16 Bank Of America Corporation Information security system and method for denial-of-service detection
US20240297854A1 (en) * 2023-03-01 2024-09-05 Nio Technology (Anhui) Co., Ltd. Distributed function-specific buffer arrangement in a communication layer

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002027469A2 (en) 2000-09-25 2002-04-04 Crossbeam Systems, Inc. Flow scheduling and architecture for network application apparatus
US20020108059A1 (en) 2000-03-03 2002-08-08 Canion Rodney S. Network security accelerator
US20030123447A1 (en) 2001-12-31 2003-07-03 Tippingpoint Technologies, Inc. System and method for classifying network packets with packet content
US20040024915A1 (en) 2002-04-24 2004-02-05 Nec Corporation Communication controller and communication control method
US20040054925A1 (en) 2002-09-13 2004-03-18 Cyber Operations, Llc System and method for detecting and countering a network attack

Family Cites Families (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5509123A (en) * 1994-03-22 1996-04-16 Cabletron Systems, Inc. Distributed autonomous object architectures for network layer routing
US5754659A (en) * 1995-12-22 1998-05-19 General Instrument Corporation Of Delaware Generation of cryptographic signatures using hash keys
US6430183B1 (en) * 1997-09-18 2002-08-06 International Business Machines Corporation Data transmission system based upon orthogonal data stream mapping
US6657742B1 (en) * 1997-11-24 2003-12-02 Xerox Corporation System for printing facsimile jobs with a property profile
US6738814B1 (en) 1998-03-18 2004-05-18 Cisco Technology, Inc. Method for blocking denial of service and address spoofing attacks on a private network
US6522188B1 (en) * 1998-04-10 2003-02-18 Top Layer Networks, Inc. High-speed data bus for network switching
US6714553B1 (en) * 1998-04-15 2004-03-30 Top Layer Networks, Inc. System and process for flexible queuing of data packets in network switching
US6826616B2 (en) * 1998-10-30 2004-11-30 Science Applications International Corp. Method for establishing secure communication link between computers of virtual private network
US6301668B1 (en) 1998-12-29 2001-10-09 Cisco Technology, Inc. Method and system for adaptive network security using network vulnerability assessment
US6499107B1 (en) 1998-12-29 2002-12-24 Cisco Technology, Inc. Method and system for adaptive network security using intelligent packet analysis
US6487666B1 (en) 1999-01-15 2002-11-26 Cisco Technology, Inc. Intrusion detection signature analysis using regular expressions and logical operators
US6578147B1 (en) 1999-01-15 2003-06-10 Cisco Technology, Inc. Parallel intrusion detection sensors with load balancing for high speed networks
US6789116B1 (en) * 1999-06-30 2004-09-07 Hi/Fn, Inc. State processor for pattern matching in a network monitor device
US6950434B1 (en) * 1999-12-07 2005-09-27 Advanced Micro Devices, Inc. Arrangement for searching packet policies using multi-key hash searches in a network switch
US20040064737A1 (en) * 2000-06-19 2004-04-01 Milliken Walter Clark Hash-based systems and methods for detecting and preventing transmission of polymorphic network worms and viruses
US7328349B2 (en) * 2001-12-14 2008-02-05 Bbn Technologies Corp. Hash-based systems and methods for detecting, preventing, and tracing network worms and viruses
US6754662B1 (en) * 2000-08-01 2004-06-22 Nortel Networks Limited Method and apparatus for fast and consistent packet classification via efficient hash-caching
US8397284B2 (en) 2006-01-17 2013-03-12 University Of Maryland Detection of distributed denial of service attacks in autonomous system domains
US20070204344A1 (en) 2006-02-26 2007-08-30 Chun Xue Parallel Variable Length Pattern Matching Using Hash Table
US20070245417A1 (en) 2006-04-17 2007-10-18 Hojae Lee Malicious Attack Detection System and An Associated Method of Use
US7890612B2 (en) 2006-05-08 2011-02-15 Electro Guard Corp. Method and apparatus for regulating data flow between a communications device and a network
KR100875673B1 (en) * 2007-05-14 2008-12-24 주식회사 하이닉스반도체 On-die termination device and its calibration method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020108059A1 (en) 2000-03-03 2002-08-08 Canion Rodney S. Network security accelerator
WO2002027469A2 (en) 2000-09-25 2002-04-04 Crossbeam Systems, Inc. Flow scheduling and architecture for network application apparatus
US20030123447A1 (en) 2001-12-31 2003-07-03 Tippingpoint Technologies, Inc. System and method for classifying network packets with packet content
US20040024915A1 (en) 2002-04-24 2004-02-05 Nec Corporation Communication controller and communication control method
US20040054925A1 (en) 2002-09-13 2004-03-18 Cyber Operations, Llc System and method for detecting and countering a network attack

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
HUTCHINGS ET AL.: "Assisting network intrusion detection with reconfigurable hardware", FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES, 2002, PROCEEDINGS 10TH ANNUAL IEEE SYMPOSIUM, 22 April 2002 (2002-04-22)
See also references of EP1754349A4

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101184095B (en) * 2007-12-06 2011-09-21 中兴通讯股份有限公司 Network anti-attack method and system based on strategy control listing of CPU
CN104394150A (en) * 2014-11-26 2015-03-04 大连梯耐德网络技术有限公司 System and method for implementing mimic security network architecture based on hardware reconfiguration
CN104394150B (en) * 2014-11-26 2018-09-25 大连梯耐德网络技术有限公司 A kind of realization system and method for the mimicry security network infrastructure based on hardware reconstruction
WO2022084625A1 (en) * 2020-10-22 2022-04-28 Orange Methods and devices for protecting a stream of packets
FR3115648A1 (en) * 2020-10-22 2022-04-29 Orange Packet flow protection methods and devices

Also Published As

Publication number Publication date
CN101421991A (en) 2009-04-29
US7411957B2 (en) 2008-08-12
CN101421991B (en) 2011-04-06
AU2005326185A8 (en) 2008-07-31
CA2559251A1 (en) 2005-09-26
WO2006124009A3 (en) 2009-04-16
US20050213570A1 (en) 2005-09-29
CA2559251C (en) 2011-06-28
EP1754349A2 (en) 2007-02-21
AU2005326185A1 (en) 2006-10-05
EP1754349A4 (en) 2011-01-12
EP1754349B1 (en) 2020-02-12

Similar Documents

Publication Publication Date Title
US7411957B2 (en) Hardware filtering support for denial-of-service attacks
US12095882B2 (en) Accelerated network packet processing
US20050171937A1 (en) Memory efficient hashing algorithm
US10735325B1 (en) Congestion avoidance in multipath routed flows
US7522521B2 (en) Route processor adjusting of line card admission control parameters for packets destined for the route processor
US9813339B2 (en) Filtering and route lookup in a switching device
US7580351B2 (en) Dynamically controlling the rate and internal priority of packets destined for the control plane of a routing device
JP3734704B2 (en) Packet classification engine
US8799507B2 (en) Longest prefix match searches with variable numbers of prefixes
US7346059B1 (en) Header range check hash circuit
WO2018178906A1 (en) Flexible processor of a port extender device
US10097467B1 (en) Load balancing for multipath groups routed flows by re-associating routes to multipath groups
US10116567B1 (en) Load balancing for multipath group routed flows by re-routing the congested route
US8265072B2 (en) Frame switching device
US10069734B1 (en) Congestion avoidance in multipath routed flows using virtual output queue statistics
US11863467B2 (en) Methods and systems for line rate packet classifiers for presorting network packets onto ingress queues
US7397762B1 (en) System, device and method for scheduling information processing with load-balancing
US20190044873A1 (en) Method of packet processing using packet filter rules
CN110365667B (en) Attack message protection method and device and electronic equipment
US9152494B2 (en) Method and apparatus for data packet integrity checking in a processor

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 2005326185

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 200580009074.0

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2559251

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

ENP Entry into the national phase

Ref document number: 2005326185

Country of ref document: AU

Date of ref document: 20050303

Kind code of ref document: A

WWP Wipo information: published in national office

Ref document number: 2005326185

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 2005857867

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 2005857867

Country of ref document: EP