WO2005104437A1 - Weighted random scheduling - Google Patents
Weighted random scheduling Download PDFInfo
- Publication number
- WO2005104437A1 WO2005104437A1 PCT/US2005/006801 US2005006801W WO2005104437A1 WO 2005104437 A1 WO2005104437 A1 WO 2005104437A1 US 2005006801 W US2005006801 W US 2005006801W WO 2005104437 A1 WO2005104437 A1 WO 2005104437A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- outputs
- requests
- packets
- inputs
- request
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/25—Routing or path finding in a switch fabric
- H04L49/253—Routing or path finding in a switch fabric using establishment or release of connections between ports
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/25—Routing or path finding in a switch fabric
- H04L49/253—Routing or path finding in a switch fabric using establishment or release of connections between ports
- H04L49/254—Centralised controller, i.e. arbitration or scheduling
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/20—Support for services
- H04L49/205—Quality of Service based
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/30—Peripheral units, e.g. input or output ports
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L49/00—Packet switching elements
- H04L49/30—Peripheral units, e.g. input or output ports
- H04L49/3045—Virtual queuing
Definitions
- This invention especially relates to communications and computer systems; and more particularly, the invention relates to weighted random scheduling particularly applicable to packet switching systems.
- IP Internet Protocol
- Input queued switches typically store packets destined for different destinations in a single queue which can lead to blocking conditions.
- many network devices have at each input multiple virtual output queues (VoQs), with one or more queues corresponding to each output of the respective network device.
- This architecture requires the scheduling of the sending of packets from these multiple VoQs (in contrast to selecting the packet at the head of the single input queue).
- Prior systems use a combination of virtual output queues (VoQs) and scheduling algorithms based on maximal weight matching.
- the practical implementations of these scheduling algorithms are complex and expensive to implement in hardware at high speed, and is becoming more problematic as line rates continue to increase.
- SLIP is an algorithm for scheduling the sending of packets across an Nx N switch. In one implementation, the following three steps are performed:
- Each unmatched input sends a request to every output for which it has a queued cell.
- an unmatched output receives any requests, it chooses the one that appears next in a fixed, round-robin schedule starting from the highest selection priority element. The output notifies each input whether or not its request was granted. The pointer to the highest selection priority element of the round-robin schedule is incremented (modulo N) to one location beyond the granted input if and only if the grant is accepted in step 3 of the first iteration. The pointer is not incremented in subsequent iterations. 3. If an input receives a grant, it accepts the one that appears next in a fixed, round-robin schedule starting from the highest selection priority element. The pointer to the highest selection priority element of the round-robin schedule is incremented (modulo N) to one location beyond the accepted output.
- I-SLIP is a scheduling algorithm including multiple iterations of the SLIP algorithm to determine the scheduling of packets for each round of sending packets (rather than just one SLIP iteration.)
- Each output scheduler decides among the set of ordered, competing requests using a rotating selection priority. When a requesting input is granted and the input accepts that grant, the input will have the lowest selection priority at that output in the next cell time. Also, whatever input has the highest selection priority at an output will continue to be granted during each successive time slot until it is serviced. This ensures that a connection will not be starved: the highest selection priority connection at an output will always be accepted by an input in no more than N cell times. Moving the pointers not only prevents starvation, it tends to desynchronize the schedulers.
- Each of the outputs that matched in the previous time slot will have a different highest selection priority input. Thus, they will each grant to different inputs.
- both outputs may grant to the same input; in that case only one connection will be made in the first iteration.
- the successful output will increment its pointer and in the next cell time, the outputs will no longer contend: one will have moved on to grant to another input and the other will grant to the same input as before. This leads to a better match in the first iteration of the next cell time. This is because the output schedulers have become desynchronized (or "slipped") with respect to each other. This leads to high performance, even for a single iteration of SLIP.
- the algorithm tends to provide a fair allocation of bandwidth among competing connections and to be burst-reducing.
- the burst-reduction is simplest to understand under high load when all input queues are occupied: the algorithm will visit each competing connection in turn, so that even if a burst of cells for the same output arrives at the input, the burst will be spread out in time if there is competing traffic.
- An example implementation is described in Nicholas W. McKeown, "Method and
- One embodiment generates a request to send a packet to one of the outputs of the switch for each input of a switch by weighted randomly selecting one of the outputs to which the particular input has one or more packets to send.
- One of the requests is granted for each different one of the outputs for which one or more requests were generated. Packets are sent between the inputs and the output corresponding to the granted requests.
- a round consisting of generating the request, granting one of the requests, and sending packets is performed at least two times per packet time, with a packet time corresponding to the amount of time allocated to receive a packet at an input from an external source.
- the switch is configured to setup connections between the switch inputs and the outputs corresponding to the granted requests prior to the sending packets.
- the granting one of the requests includes randomly selecting one of the requests for each of the outputs for which one or more of the requests were generated.
- One embodiment performs one or more rounds of generating requests by weighted randomly selecting among all or a subset of the outputs, and selectively granting the requests corresponding to available bandwidth after granting one of the requests and before sending packets.
- the weighted random selection is weighted based on a number of packets to send to each of the outputs by a corresponding input of the inputs. In one embodiment, the weighted random selection is weighted based on a number of bytes to send to each of the outputs by a corresponding input of the inputs. In one embodiment, the weighted random selection is weighted based on the last times packets were sent from a corresponding input of the inputs to each of the outputs. In one embodiment, the weighted random selection is weighted based on classes of service associated with packets to send to each of the outputs by a corresponding input of the inputs.
- the weighted random selection is weighted based on at least two items from a group consisting of: (a) the number of packets to send to each of the outputs by a corresponding input of the inputs, (b) the number of bytes to send to each of the outputs by a corresponding input of the inputs, (c) the last times packets were sent from a corresponding input of the inputs to each of the outputs, and (d) classes of service associated with packets to send to each of the outputs by a corresponding input of the inputs.
- FIGs. 1A-C and 2A-B are block diagrams of a few of an unlimited number of embodiments using weighted random scheduling of packets;
- FIG. 3 is a block diagram of a component used in one embodiment for performing weighted random scheduling;
- FIG. 4A is a diagram illustrating the generation of requests and grants as performed in one embodiment;
- FIG. 4B is a diagram illustrating the generation of requests and grants as performed in one embodiment;
- FIG. 1A-C and 2A-B are block diagrams of a few of an unlimited number of embodiments using weighted random scheduling of packets;
- FIG. 3 is a block diagram of a component used in one embodiment for performing weighted random scheduling;
- FIG. 4A is a diagram illustrating the generation of requests and grants as performed in one embodiment;
- FIG. 4B is a diagram illustrating the generation of requests and grants as performed in one embodiment;
- FIG. 5 A is a flow diagram of a process used in one embodiment for performing weighted random scheduling of packets including determining requests based on a weighted random selection of destinations of pending packets at a switch input;
- FIG. 5B is a flow diagram of a process used in one embodiment for determining grants in response to received requests; and
- FIG. 6 is a flow diagram of a process used in one embodiment for performing weighted random scheduling of packets.
- Embodiments described herein include various elements and limitations, with no one element or limitation contemplated as being a critical element or limitation. Each of the claims individually recites an aspect of the invention in its entirety. Moreover, some embodiments described may include, but are not limited to, inter alia, systems, networks, integrated circuit chips, embedded processors, ASICs, methods, and computer-readable media containing instructions. One or multiple systems, devices, components, etc. may comprise one or more embodiments, which may include some elements or limitations of a claim being performed by the same or different systems, devices, components, etc.
- packet refers to packets of all types or any other units of information or data, including, but not limited to, fixed length cells and variable length packets, each of which may or may not be divisible into smaller packets or cells.
- packet as used herein also refers to both the packet itself or a packet indication, such as, but not limited to all or part of a packet or packet header, a data structure value, pointer or index, or any other part or direct or indirect identification of a packet or information associated therewith.
- these packets may contain one or more types of information, including, but not limited to, voice, data, video, and audio information.
- the term "item” is used generically herein to refer to a packet or any other unit or piece of information or data, a device, component, element, or any other entity.
- processing a packet typically refers to performing some steps or actions based on the packet contents (e.g., packet header or other fields), and such steps or action may or may not include modifying, storing, dropping, and/or forwarding the packet and/or associated data.
- system is used generically herein to describe any number of components, elements, sub-systems, devices, packet switch elements, packet switches, routers, networks, computer and/or communication devices or mechanisms, or combinations of components thereof.
- computer is used generically herein to describe any number of computers, including, but not limited to personal computers, embedded processing elements and systems, control logic, ASICs, chips, workstations, mainframes, etc.
- processing element is used generically herein to describe any type of processing mechanism or device, such as a processor, ASIC, field programmable gate array, computer, etc.
- device is used generically herein to describe any type of mechanism, including a computer or system or component thereof.
- task and “process” are used generically herein to describe any type of running program, including, but not limited to a computer process, task, thread, executing application, operating system, user process, device driver, native code, machine or other language, etc., and can be interactive and/or non-interactive, executing locally and/or remotely, executing in foreground and/or background, executing in the user and/or operating system address spaces, a routine of a library and/or standalone application, and is not limited to any particular memory partitioning technique.
- the term "identify” is used generically to describe any manner or mechanism for directly or indirectly ascertaining something, which may include, but is not limited to receiving, retrieving from memory, determining, defining, calculating, generating, etc.
- the terms “network” and “communications mechanism” are used generically herein to describe one or more networks, communications media or communications systems, including, but not limited to the Internet, private or public telephone, cellular, wireless, satellite, cable, local area, metropolitan area and/or wide area networks, a cable, electrical connection, bus, etc., and internal communications mechanisms such as message passing, interprocess communications, shared memory, etc.
- the term “message” is used generically herein to describe a piece of information which may or may not be, but is typically communicated via one or more communication mechanisms of any type.
- storage mechanism includes any type of memory, storage device or other mechanism for maintaining instructions or data in any format.
- Computer-readable medium is an extensible term including any memory, storage device, storage mechanism, and other storage and signaling mechanisms including interfaces and devices such as network interface cards and buffers therein, as well as any communications devices and signals received and transmitted, and other current and evolving technologies that a computerized system can interpret, receive, and/or transmit.
- the term “memory” includes any random access memory (RAM), read only memory (ROM), flash memory, integrated circuits, and/or other memory components or elements.
- storage device includes any solid state storage media, disk drives, diskettes, networked services, tape drives, and other storage devices.
- Memories and storage devices may store computer-executable instructions to be executed by a processing element and/or control logic, and data which is manipulated by a processing element and/or control logic.
- data structure is an extensible term referring to any data element, variable, data structure, database, and/or one or more organizational schemes that can be applied to data to facilitate interpreting the data or performing operations on it, such as, but not limited to memory locations or devices, sets, queues, trees, heaps, lists, linked lists, arrays, tables, pointers, etc.
- a data structure is typically maintained in a storage mechanism.
- pointer and link are used generically herein to identify some mechanism for referencing or identifying another element, component, or other entity, and these may include, but are not limited to a reference to a memory or other storage mechanism or location therein, an index in a data structure, a value, etc.
- the term "one embodiment” is used herein to reference a particular embodiment, wherein each reference to “one embodiment” may refer to a different embodiment, and the use of the term repeatedly herein in describing associated features, elements and/or limitations does not establish a cumulative set of associated features, elements and/or limitations that each and every embodiment must include, although an embodiment typically may include all these features, elements and/or limitations.
- the phrase “means for xxx” typically includes computer-readable medium containing computer-executable instructions for performing xxx.
- the terms “first,” “second,” etc. are typically used herein to denote different units (e.g., a first element, a second element). The use of these terms herein does not necessarily connote an ordering such as one unit or event occurring or coming before another, but rather provides a mechanism to distinguish between particular units.
- a singular tense of a noun is non-limiting, with its use typically including one or more of the particular thing rather than just one (e.g., the use of the word "memory” typically refers to one or more memories without having to specify “memory or memories,” or “one or more memories” or “at least one memory”, etc.).
- the phrases “based on x” and “in response to x” are used to indicate a minimum set of items x from which something is derived or caused, wherein "x" is extensible and does not necessarily describe a complete list of items on which the operation is performed, etc.
- Coupled to is used to indicate some level of direct or indirect connection between two elements or devices, with the coupling device or devices modifying or not modifying the coupled signal or communicated information.
- subset is used to indicate a group of all or less than all of the elements of a set.
- subtree is used to indicate all or less than all of a tree.
- or is used herein to identify a selection of one or more, including all, of the conjunctive items.
- transitional term "comprising,” which is synonymous with “including,” “containing,” or “characterized by,” is inclusive or open-ended and does not exclude additional, unrecited elements or method steps.
- One embodiment generates a request to send a packet to one of the outputs of the switch for each input of a switch by weighted randomly selecting one of the outputs to which the particular input has one or more packets to send.
- One of the requests is granted for each different one of the outputs for which one or more requests were generated. Packets are sent between the inputs and the output corresponding to the granted requests.
- a round consisting of generating the request, granting one of the requests, and sending packets is performed at least two times per packet time, with a packet time corresponding to the amount of time allocated to receive a packet at an input from an external source.
- the switch is configured to setup connections between the switch inputs and the outputs corresponding to the granted requests prior to the sending packets.
- the granting one of the requests includes randomly selecting one of the requests for each of the outputs for which one or more of the requests were generated.
- One embodiment performs one or more rounds of generating requests by weighted randomly selecting among all or a subset of the outputs, and selectively granting the requests corresponding to available bandwidth after granting one of the requests and before sending packets.
- the weighted random selection is weighted based on a number of packets to send to each of the outputs by a corresponding input of the inputs. In one embodiment, the weighted random selection is weighted based on a number of bytes to send to each of the outputs by a corresponding input of the inputs. In one embodiment, the weighted random selection is weighted based on the last times packets were sent from a corresponding input of the inputs to each of the outputs. In one embodiment, the weighted random selection is weighted based on classes of service associated with packets to send to each of the outputs by a corresponding input of the inputs.
- the weighted random selection is weighted based on at least two items from a group consisting of: (a) the number of packets to send to each of the outputs by a corresponding input of the inputs, (b) the number of bytes to send to each of the outputs by a corresponding input of the inputs, (c) the last times packets were sent from a corresponding input of the inputs to each of the outputs, and (d) classes of service associated with packets to send to each of the outputs by a corresponding input of the inputs.
- One embodiment includes multiple request generators and one or more grant arbiters coupled to the request generators.
- Each of the request generators is configured to weighted randomly generate a request for its associated input of multiple inputs of a switch, the request including a weighted random selection of one of the outputs of the switch.
- One or more grant arbiters are configured to grant requests received from the request generators such that one request for each requested output is granted.
- the inputs are configured to send packets corresponding to said granted requests.
- One embodiment includes a control to configure the switch to setup connections between said inputs and said outputs corresponding to said granted requests.
- one or more grant arbiters randomly select a request to grant for a particular output from said requests received for the particular output.
- inputs are configured to send packets from their respective virtual output queues corresponding to the granted requests.
- One embodiment includes multiple line cards and a switching board including the switch.
- the virtual output queues are located on the line cards.
- the virtual output queues are located on the switching board.
- the weighted random selection is weighted based on a number of packets or bytes to send to each of the outputs by a corresponding input of the inputs.
- the weighted random selection is weighted based on the last times packets were sent from a corresponding input of the inputs to each of the outputs.
- the weighted random selection is weighted based on classes of service associated with packets to send to each of the outputs by a corresponding input of the inputs.
- the weighted random selection is weighted on at least two items from a group consisting of: (a) the number of packets to send to each of the outputs by a corresponding input, (b) the number of bytes to send to each of the outputs by a corresponding input, (c) the last times packets were sent from a corresponding input to each of the outputs, and (d) the classes of service associated with packets to send to each of the outputs by a corresponding input.
- One embodiment for each particular input of a inputs, generates a first request including weighted randomly selecting one of the outputs. One of the first requests for each different one of the outputs for which one or more of said first requests was generated is granted.
- a second request is generated, which includes weighted randomly selecting one of the outputs (possibly limited to those not already have a request previously granted).
- One of said second requests is granted for each different one of the outputs not already having a corresponding first request granted and for which one or more of said second requests was generated.
- the weighted random selection is weighted based on the number of packets or bytes to send to each of the outputs by the corresponding input.
- packets are sent between said inputs and said outputs based said granted first and second requests.
- only two request phases are performed per packet time, said two request phases consisting of said generating said first requests and said generating said second requests.
- each particular request to grant of said first requests and said second requests for each particular output is randomly selected from all the respective said first requests or said second request for said particular output.
- One embodiment includes multiple request generators and one or more grant arbiters coupled to the request generators.
- Each of the request generators is configured to weighted randomly generate a first round request and as required a second round request for its associated input of a inputs of a switch, with the first round request including a first weighted random selection of one of a outputs and the second round request including a second weighted random selection of one of the outputs if the first round request was not granted.
- the grant arbiters are configured to grant requests received from the request generators such that one request for each requested output in the first round of requests is granted, and one request for each requested output in the second round of requests is granted if a grant for said requested output was not generated for a request in the first round of requests.
- the one or more grant arbiters randomly select a request to grant for a output from the set of requests received for the output.
- the weighted random generation is weighted based on the number of packets or bytes to send to each of the outputs by the corresponding input.
- One embodiment randomly picks one of the candidate virtual output queues / destination for each switch input, where the random selection is weighted by the length of each of the queues for the input.
- Scheduling requests are then sent to the chosen destination.
- Each destination after getting requests from all the candidate inputs, sends a grant back to the randomly chosen input. Inputs that get a grant back from the destination, will send a corresponding packet in the corresponding packet time.
- the two stage random selection process represents a single phase of the weighted random selection process. A second phase is performed for those switch inputs that have not received a grant and for those destinations for which a grant was not generated in the first phase (i.e., no request was generated during the first phase). In this manner, a minimal amount of scheduling overhead is incurred, and desired traffic throughput is received when these two scheduling phases are performed in the time allotted to send one or more packets through the switch.
- FIGs. 1A-C and 2A-B are block diagrams of a few of an unlimited number of embodiments using weighted random scheduling of packets.
- a switching mechanism e.g., a physical switching unit, a switching fabric, a crossbar, bus, or any other mechanism for switching packets, etc.
- System 100 schedules packets being sent among line cards 103-104 using weighted random scheduling.
- weighted random scheduling can be used in conjunction with a blocking or a non-blocking switch.
- IB illustrates a system 110 that uses weighted random scheduling of packets.
- Line cards 111-112 each contain virtual output queues in this exemplary embodiment, and switching board 120 includes a control with scheduler and memory 121, multiple buffers 122-123, and a switching mechanism 125.
- Line cards 111-112 generate the scheduling requests based on a weighted random selection.
- Packets are then sent to input buffers 122-123 according to the determined weighted random scheduling, with the particular weight applied being typically based on the occupancies of the virtual output queues and/or characteristics of the packets (e.g., number of packets or bytes, age of packets, quality of service, etc.).
- input buffers 122-123 can each store multiple packets, and thus previously weighted random scheduled packets can be queued while the scheduling is performed for subsequent packet times.
- Control with scheduler and memory 121 typically generates the grants based on the received scheduling requests, and controls the configuration of switching mechanism 125 so as to allow the scheduled packets to be sent through it.
- FIG. 1C illustrates a system 140 that uses weighted random scheduling of packets.
- Line cards 141-142 communicate packets with switching board 150 in this exemplary embodiment, and switching board 150 includes a control with scheduler and memory 151, virtual output queues 152-153, and a switching mechanism 155.
- the weighted random scheduling of packets is performed by control with scheduler and memory 151, with the particular weight applied being typically based on the occupancies of the virtual output queues 152-153 and/or characteristics of the packets (e.g., number of packets or bytes, age of packets, quality of service, etc.). Packets are then sent from virtual output queues 152-153 according to the determined weighted random scheduling, with the configuration of switching mechanism 155.
- FIG. 2A illustrates a system 200 that uses weighted random scheduling of packets.
- Packets 201 are received and placed in packet queues 202.
- the queue occupancies 211 are communicated to the weighted random request generator 210 for each switch input (also referred to herein as a switch input), which also communicated with grant generator(s) 220 (e.g., one grant generator, one grant generator for each destination, etc.).
- Grant generator(s) 220 communicate the determined switch configuration 221 to the switching mechanism 225 based on the generated grants 217, and the weighted random request generators 210 provide send instructions to packet queues 202 to send packets corresponding to the weighted random scheduled packets.
- Packets 203 are sent from packet queues 202, through the accordingly configured switching mechanism 204 to produce the switched packets 205.
- FIG. 2B illustrates a system 230 that uses weighted random scheduling of packets.
- Packets 231 are received and placed in packet queues 232.
- the queue occupancies 241 are communicated to the weighted random generator 210 for each switching mechanism input.
- Weighted random request and grant generator 250 uses one or more processes and/or customized control logic to perform the weighted random scheduling of packets, and to inform packet queues 232 of the packets to send (243) and switching mechanism 234 of the corresponding switch configuration 251.
- Packets 213 are sent from packet queues 232, through the accordingly configured switching mechanism 234 to produce the switched packets 235.
- component 300 is a block diagram of a component 300 used in one embodiment for performing weighted random scheduling, such as for generating weighted random requests and/or grants.
- component 300 performs one or more processes corresponding to one of the flow diagrams illustrated or otherwise described herein.
- system or component 300 includes a processing element 301, memory 302, storage devices 303, and an interface 304 for sending and receiving packets, items, and/or other information (e.g., queue occupancy, instructions to send packets, switch configuration to a communicatively coupled switch, etc.), which are typically coupled via one or more communications mechanisms 309 (shown as a bus for illustrative purposes.)
- memory 302 and/or storage devices 303 are used to store packets on which the weighted random scheduling of the packets is performed.
- Various embodiments of component 300 may include more or less elements. The operation of component 300 is typically controlled by processing element 301 using memory 302 and storage devices 303 to perform one or more tasks or processes.
- Memory 302 is one type of computer-readable media, and typically comprises random access memory (RAM), read only memory (ROM), flash memory, integrated circuits, and/or other memory components. Memory 302 typically stores computer-executable instructions to be executed by processing element 301 and/or data which is manipulated by processing element 301 for implementing functionality in accordance with an embodiment.
- Storage devices 303 are another type of computer-readable media, and typically comprise solid state storage media, disk drives, diskettes, networked services, tape drives, and other storage devices. Storage devices 303 typically store computer-executable instructions to be executed by processing element 301 and/or data which is manipulated by processing element 301 for implementing functionality in accordance with an embodiment.
- FIG. 4 A is a diagram illustrating the generation of requests and grants used in one embodiment.
- Weighted random request generators 400 determine a request for each switch input as indicated by processing 411. This selection is weighted based on the occupancies of the virtual output queues and/or characteristics of the packets (e.g., number of packets or bytes, age of packets, quality of service i.e., higher priority traffic is given greater weight, etc.). For example, assume there are four switch outputs, with eight packets queued at one input with one packet queued for the first output, two packets queued for the second output, one packet queued for the third output, and four packets queued for the fourth output. One embodiment uses a direct weighted proportion of the number of queued packets, while one embodiment uses different proportion calculation based on the occupancy of the virtual output queues.
- one embodiment uses a weight of one-eighth to send a packet from the queue corresponding to switch output/destination one; a weight of one-quarter to send a packet from the queue corresponding to switch output/destination two; a weight of one-half to send a packet from the queue corresponding to switch output/destination three; and a weight of one-eighth to send a packet from the queue corresponding to switch output/destination four.
- a random number i.e., a true random number, a pseudo random number, or any approximated random value
- one embodiment maps the percentages for each of the destination to a cumulative fractional value between zero and one, and a random value between zero and one is generated.
- destination one corresponds to 0.00 to 0.125
- destination two corresponds to 0.126 to 0.375
- destination three corresponds to 0.376 to 0.500
- destination four corresponds to 0.501 to 1.000.
- the determined weighted random requests 412 for each switch input are communicated to grant generator(s) 402, which randomly or otherwise select one of the requests for each destination for which a request was made (as illustrated by processing 413).
- These grants 414 are communicated to weighted random request generator(s) 400 (or in one embodiment to the control for sending packets and/or the control for configuring the switch).
- Packets are sent according to the grants as illustrated by reference number 431.
- typically one packet per grant is sent as that is the most that can be sent from one input to one output of a switch at one time.
- multiple packets can be sent from an input and/or received by an output at a time, so the number of requests generated for an input and the number of grants generated for an output is accordingly adjusted.
- FIG. 4B is similar to FIG. 4A with the addition used in one embodiment to perform multiple requests/grant rounds prior to sending packets corresponding to the generated grants.
- Weighted random request generators 400 determine a request for each switch input as indicated by processing 411.
- This selection is weighted based on the occupancies of the virtual output queues and/or characteristics of the packets (e.g., number of packets or bytes, age of packets, quality of service i.e., higher priority traffic is given greater weight, etc.).
- the determined weighted random requests 412 for each switch input are communicated to grant generator(s) 402, which randomly or otherwise select one of the requests for each destination for which a request was made (as illustrated by processing 413).
- These grants 414 are communicated to weighted random request generator(s) 400.
- Each request generator that did not receive a grant then performs an additional request phase, including determining their requests based on a weighted random selection based on the occupancies of the virtual output queues and/or characteristics of the packets in their queues (415).
- the additional requests may have been determined prior to receiving a grant 414.
- all destinations remain available for a request in the additional phase.
- grants 414 include an indication of the granted destinations, and thus these destinations can be removed from contention in identifying a destination for which to request during the second phase.
- the additional phase request determination excludes the destination requested in a prior phase (as it was granted to a different requester and is unavailable during the additional phase) from the weighted random selection in the additional phase.
- more than one packet can be sent from an input and/or received by an output in a packet time, and therefore the number of requests and/or grants are accordingly increased.
- the weighted random requests 416 for the switch inputs that can send more packets in a packet time are communicated to grant generator(s) 402, which then randomly or otherwise select (417) among the received requests 416 for the destinations which can accept additional packets in a packet time (e.g., those destinations which did not receive/grant a request during a prior phase in one embodiment).
- the additional phase grants 418 are then communicated to weighted random request generator(s) 400 (or in one embodiment to the control for sending packets and/or the control for configuring the switch).
- FIG. 5 A is a flow diagram of a process used in one embodiment for performing weighted random scheduling of packets by a request generator for a particular switch input.
- Processing begins with process block 500, and proceeds to process block 502, wherein a request is generated by identifying one of the possible destinations (i.e., identifying the destination/switch output, virtual output queue, or other corresponding value) based on a weighted random selection based on packet occupancies of the virtual output queues and/or characteristics of the packets (e.g., number of packets or bytes, age of packets, quality of service i.e., higher priority traffic is given greater weight, etc.).
- process block 504 if the request was granted, then in process block 506, a packet is sent to the destination from the virtual output queue corresponding to the destination. Processing is complete as indicated by process block 508.
- 5B is a flow diagram of a process used in one embodiment for determining grants from received requests. Processing begins with process block 520, and proceeds to process block 522, wherein zero or more requests are received for each particular destination, and one of these requests is randomly or otherwise selected for each particular destination, and a corresponding grant is communicated to the requester. In process block 524, the switch is configured, if required, based on the granted requests so the packets can be sent (e.g., in process block 506 of FIG. 5 A). Processing is complete as indicated by process block 526. FIG.
- FIG. 6 is a flow diagram of a process used in one embodiment for performing weighted random scheduling of packets in a packet time, with a packet time corresponding to the amount of time allocated to receive a packet at an input of a switch from an external source.
- Processing begins with process block 600, and proceeds to process block 602, wherein requests are generated for each switch input having one or more packets to send based on a weighted random selection.
- process block 604 one of the requests for each of the requested destinations is randomly or otherwise granted.
- the switch is configured, if required, to accommodate the granted requests, and the packets are sent from the corresponding inputs to the corresponding outputs.
- process block 608 if there are more request/grant/send packet rounds to be performed in the packet time, processing returns to process block 602. Otherwise, processing is complete as indicated by process block 610.
- process block 610 In view of the many possible embodiments to which the principles of our invention may be applied, it will be appreciated that the embodiments and aspects thereof described herein with respect to the drawings/figures are only illustrative and should not be taken as limiting the scope of the invention. For example and as would be apparent to one skilled in the art, many of the process block operations can be re-ordered to be performed before, after, or substantially concurrent with other operations. Also, many different forms of data structures could be used in various embodiments. The invention as described herein contemplates all such embodiments as may come within the scope of the following claims and equivalents thereof.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2005800104226A CN1938992B (en) | 2004-04-12 | 2005-03-02 | Method and device for weighted random scheduling |
EP05724365.1A EP1741229B1 (en) | 2004-04-12 | 2005-03-02 | Weighted random scheduling |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/822,908 US7061927B2 (en) | 2004-04-12 | 2004-04-12 | Weighted random scheduling particularly applicable to packet switching systems |
US10/822,908 | 2004-04-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005104437A1 true WO2005104437A1 (en) | 2005-11-03 |
Family
ID=35060476
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2005/006801 WO2005104437A1 (en) | 2004-04-12 | 2005-03-02 | Weighted random scheduling |
Country Status (4)
Country | Link |
---|---|
US (1) | US7061927B2 (en) |
EP (1) | EP1741229B1 (en) |
CN (1) | CN1938992B (en) |
WO (1) | WO2005104437A1 (en) |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7817643B2 (en) * | 2004-09-14 | 2010-10-19 | Maksim Azarov | System and method for varying the scheduling of real time protocol (RTP) packets |
US7499464B2 (en) * | 2005-04-06 | 2009-03-03 | Robert Ayrapetian | Buffered crossbar switch with a linear buffer to port relationship that supports cells and packets of variable size |
US7657537B1 (en) * | 2005-04-29 | 2010-02-02 | Netapp, Inc. | System and method for specifying batch execution ordering of requests in a storage system cluster |
US8547855B1 (en) * | 2006-03-21 | 2013-10-01 | Cisco Technology, Inc. | Method and apparatus to schedule multiple probes for active or passive monitoring of networks |
US7747734B2 (en) * | 2006-03-29 | 2010-06-29 | International Business Machines Corporation | Apparatus, system, and method for error assessment over a communication link |
US20070268825A1 (en) * | 2006-05-19 | 2007-11-22 | Michael Corwin | Fine-grain fairness in a hierarchical switched system |
US20080161105A1 (en) * | 2006-12-30 | 2008-07-03 | Cadillac Jack Inc. | Progressive jackpot gaming systems and methods |
US7839864B1 (en) * | 2007-05-21 | 2010-11-23 | Cisco Technology, Inc. | Priority propagation with a hierarchical asynchronous transfer mode (ATM) scheduler |
US8830837B2 (en) | 2010-12-17 | 2014-09-09 | Cisco Technology, Inc. | Dynamic synchronized scheduling in a computer network |
US9590820B1 (en) | 2011-09-02 | 2017-03-07 | Juniper Networks, Inc. | Methods and apparatus for improving load balancing in overlay networks |
US9331929B1 (en) | 2012-03-29 | 2016-05-03 | Juniper Networks, Inc. | Methods and apparatus for randomly distributing traffic in a multi-path switch fabric |
EP3170353B1 (en) * | 2014-07-17 | 2018-12-26 | Telefonaktiebolaget LM Ericsson (publ) | Method and network element for scheduling a communication device |
US11683276B2 (en) | 2021-05-20 | 2023-06-20 | Cisco Technology, Inc. | Quasi-output queue behavior of a packet switching device achieved using virtual output queue ordering independently determined for each output queue |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5500858A (en) | 1994-12-20 | 1996-03-19 | The Regents Of The University Of California | Method and apparatus for scheduling cells in an input-queued switch |
US6212182B1 (en) | 1996-06-27 | 2001-04-03 | Cisco Technology, Inc. | Combined unicast and multicast scheduling |
US6262986B1 (en) * | 1995-07-07 | 2001-07-17 | Kabushiki Kaisha Toshiba | Method and apparatus for packet scheduling using queue length and connection weight |
US6359861B1 (en) | 1997-10-08 | 2002-03-19 | Massachusetts Institute Of Technology | Method for scheduling transmissions in a buffered switch |
US20030031193A1 (en) | 2001-04-06 | 2003-02-13 | Andrews Daniel M. | Scalable weight-based terabit switch scheduling method |
US6757246B2 (en) * | 2001-08-14 | 2004-06-29 | Pts Corporation | Method and apparatus for weighted arbitration scheduling separately at the input ports and the output ports of a switch fabric |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS6262697A (en) * | 1985-09-13 | 1987-03-19 | Toshiba Corp | Data transmission control system |
US5923644A (en) * | 1996-10-03 | 1999-07-13 | The Board Of Trustees Of The Leland Stanford Junior University | Apparatus and method for processing multicast cells in an input-queued multicast switch |
US6628646B1 (en) | 1999-05-14 | 2003-09-30 | Nortel Networks Limited | Programmable multicast scheduling for a network device |
KR100382142B1 (en) | 2000-05-19 | 2003-05-01 | 주식회사 케이티 | Cell scheduling method of input and output buffered switch using simple iterative matching algorithm |
US20030072326A1 (en) * | 2001-08-14 | 2003-04-17 | Mehdi Alasti | Method and apparatus for parallel, weighted arbitration scheduling for a switch fabric |
US7184443B2 (en) | 2002-03-30 | 2007-02-27 | Cisco Technology, Inc. | Packet scheduling particularly applicable to systems including a non-blocking switching fabric and homogeneous or heterogeneous line card interfaces |
US7539199B2 (en) * | 2003-02-21 | 2009-05-26 | Gireesh Shrimali | Switch fabric scheduling with fairness and priority consideration |
-
2004
- 2004-04-12 US US10/822,908 patent/US7061927B2/en not_active Expired - Fee Related
-
2005
- 2005-03-02 CN CN2005800104226A patent/CN1938992B/en not_active Expired - Fee Related
- 2005-03-02 EP EP05724365.1A patent/EP1741229B1/en not_active Not-in-force
- 2005-03-02 WO PCT/US2005/006801 patent/WO2005104437A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5500858A (en) | 1994-12-20 | 1996-03-19 | The Regents Of The University Of California | Method and apparatus for scheduling cells in an input-queued switch |
US6262986B1 (en) * | 1995-07-07 | 2001-07-17 | Kabushiki Kaisha Toshiba | Method and apparatus for packet scheduling using queue length and connection weight |
US6212182B1 (en) | 1996-06-27 | 2001-04-03 | Cisco Technology, Inc. | Combined unicast and multicast scheduling |
US6359861B1 (en) | 1997-10-08 | 2002-03-19 | Massachusetts Institute Of Technology | Method for scheduling transmissions in a buffered switch |
US20030031193A1 (en) | 2001-04-06 | 2003-02-13 | Andrews Daniel M. | Scalable weight-based terabit switch scheduling method |
US6757246B2 (en) * | 2001-08-14 | 2004-06-29 | Pts Corporation | Method and apparatus for weighted arbitration scheduling separately at the input ports and the output ports of a switch fabric |
Also Published As
Publication number | Publication date |
---|---|
US7061927B2 (en) | 2006-06-13 |
US20050226263A1 (en) | 2005-10-13 |
EP1741229A4 (en) | 2011-01-12 |
CN1938992B (en) | 2011-09-07 |
CN1938992A (en) | 2007-03-28 |
EP1741229A1 (en) | 2007-01-10 |
EP1741229B1 (en) | 2017-10-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1741229B1 (en) | Weighted random scheduling | |
US7453898B1 (en) | Methods and apparatus for simultaneously scheduling multiple priorities of packets | |
US7630376B2 (en) | Distributed packet processing with ordered locks to maintain requisite packet orderings | |
US7606250B2 (en) | Assigning resources to items such as processing contexts for processing packets | |
US20050219564A1 (en) | Image forming device, pattern formation method and storage medium storing its program | |
US7212535B2 (en) | Scheduling items using mini-quantum values | |
AU2003218229B2 (en) | Packet scheduling mechanism | |
US7830903B2 (en) | Method and apparatus for scheduling packets and/or cells | |
US7350208B1 (en) | Method and apparatus for scheduling using a resource variable decreased by amounts corresponding to the efficiency of the resource | |
US20030043813A1 (en) | Distribution of weightings between port control system and switch cards of a packet switching device | |
US8018932B1 (en) | Methods and apparatus for sending a multicast packet from multiple network interfaces using the same MAC source address | |
US7408937B1 (en) | Methods and apparatus for identifying a variable number of items first in sequence from a variable starting position which may be particularly useful by packet or other scheduling mechanisms | |
Caminero et al. | MMR: A MultiMedia Router architecture to support integrated workloads |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 200580010422.6 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2005724365 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005724365 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 2005724365 Country of ref document: EP |