US20050207339A1 - Burst switching in a high capacity network - Google Patents

Burst switching in a high capacity network Download PDF

Info

Publication number
US20050207339A1
US20050207339A1 US11/124,656 US12465605A US2005207339A1 US 20050207339 A1 US20050207339 A1 US 20050207339A1 US 12465605 A US12465605 A US 12465605A US 2005207339 A1 US2005207339 A1 US 2005207339A1
Authority
US
United States
Prior art keywords
burst
time
input port
calendar
node
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/124,656
Inventor
Maged Beshai
Richard Vickers
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US11/124,656 priority Critical patent/US20050207339A1/en
Publication of US20050207339A1 publication Critical patent/US20050207339A1/en
Assigned to AFRL/RIJ reassignment AFRL/RIJ EXECUTIVE ORDER 9424, CONFIRMATORY LICENSE Assignors: NORTEL NETWORKS LIMITED
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/0001Selecting arrangements for multiplex systems using optical switching
    • H04Q11/0062Network aspects
    • H04Q11/0066Provisions for optical burst or packet networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/04Selecting arrangements for multiplex systems for time-division multiplexing
    • H04Q11/06Time-space-time switching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q3/00Selecting arrangements
    • H04Q3/0016Arrangements providing connection between exchanges
    • H04Q3/0062Provisions for network management
    • H04Q3/0091Congestion or overload control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/0001Selecting arrangements for multiplex systems using optical switching
    • H04Q11/0005Switch and router aspects
    • H04Q2011/0037Operation
    • H04Q2011/0039Electrical control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/0001Selecting arrangements for multiplex systems using optical switching
    • H04Q11/0062Network aspects
    • H04Q2011/0064Arbitration, scheduling or medium access control aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q11/00Selecting arrangements for multiplex systems
    • H04Q11/0001Selecting arrangements for multiplex systems using optical switching
    • H04Q11/0062Network aspects
    • H04Q2011/0088Signalling aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2213/00Indexing scheme relating to selecting arrangements in general and for multiplex systems
    • H04Q2213/1305Software aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2213/00Indexing scheme relating to selecting arrangements in general and for multiplex systems
    • H04Q2213/13103Memory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2213/00Indexing scheme relating to selecting arrangements in general and for multiplex systems
    • H04Q2213/13106Microprocessor, CPU
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2213/00Indexing scheme relating to selecting arrangements in general and for multiplex systems
    • H04Q2213/13164Traffic (registration, measurement,...)
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q2213/00Indexing scheme relating to selecting arrangements in general and for multiplex systems
    • H04Q2213/13166Fault prevention

Definitions

  • the present invention relates to data communication networks and, in particular, to burst switching in a high capacity network.
  • a source node sends a burst transfer request to a core node to indicate that a burst of data is coming, the size of the burst and the destination of the burst. Responsive to this burst transfer request, the core node configures a space switch to connect a link on which the burst will be received to a link to the requested burst destination.
  • the burst follows the burst transfer request after a predetermined time period (a scheduling time) and it is expected that, when the burst arrives at the core node, the space switch will have been properly configured by the core node.
  • the source node waits for a message from the core node, where the message acknowledges that the space switch in the core node is properly configured, before sending the burst.
  • Core nodes are used that do not have buffers to buffer incoming data. Core nodes without buffers are desirable because: it may not be possible to provide buffers without an expensive optical-electrical conversion at input and electrical-optical conversion at output of an optical space switch; and the core node may be distant from the source and sink (edge) nodes, therefore requiring remote buffer management in an edge-controlled network.
  • a burst may arrive at a core node before the space switch is properly configured and, if the core node does not include a buffer, the burst may be lost. Furthermore, until the source node fails to receive an acknowledgement of receipt of the burst from the burst destination, the fact that the burst has been lost at the core node is unknown to the source node. Having not received acknowledgement of receipt of the burst, the source node may then retransmit the burst.
  • the time delay involved in sending a burst transfer request and receiving an acceptance before sending a burst may be unacceptably high, leading to low network utilization.
  • burst switching is gaining popularity as a technique to transfer data in high-speed networks since it simplifies many of the control functions and does not require capacity to be reserved when it may not always be in use. Furthermore, burst switching reduces a need for characterizing the traffic. Clearly, a burst switching technique that allows for greater network utilization is desirable.
  • a novel burst scheduling technique allows efficient utilization of network resources.
  • Burst transfer requests are received at the space switch controller and pipelined such that the controller may determine a schedule for allowing the bursts, represented by the burst transfer requests, access to the space switch. According to the schedule, scheduling information is distributed to the sources of the burst transfer requests and to a controller of the space switch.
  • the novel burst scheduling technique allows for utilization of network resources that is more efficient than typical burst switching techniques, especially when the novel burst scheduling technique is used in combination with known time locking methods.
  • the novel burst scheduling technique enables the application of burst switching to wide coverage networks. Instead of handling burst requests one-by-one, burst requests are pipelined and the handling of the bursts is scheduled over a long future period.
  • a method of controlling a space switch to establish time-varying connections includes receiving a stream of burst transfer requests from a source node, each of the burst transfer requests including parameters specifying a requested connection and a duration for the requested connection, generating scheduling information for each of the burst transfer requests based on the parameters, transmitting the scheduling information to the source node and transmitting instructions to a slave controller for the space switch, where the instructions are based on the scheduling information and instruct the space switch to establish the requested connection.
  • a space switch master controller is provided for performing this method.
  • a software medium that permits a general purpose computer to carry out this method.
  • a method of generating scheduling information includes determining a next-available input port among a plurality of input ports and a time index at which the next-available input port will become available and, for each burst transfer request of a plurality of burst transfer requests received in relation to the next-available input port, and where each the each burst transfer request includes an identity of a burst and a destination for the burst: determining, from the destination for the burst, a corresponding output port among a plurality of output ports; determining a time gap, where the time gap is a difference between: the time index at which the next-available input port will become available; and a time index at which the corresponding output port will become available.
  • the method further includes selecting one of the plurality of burst transfer requests as a selected burst transfer request, where the selected burst transfer request has a minimum time gap of the plurality of burst transfer requests, selecting a scheduled time index, where the scheduled time index is one of the time index at which the next-available input port is available and the time index at which the corresponding output port is available and transmitting scheduling information for a burst identified by the selected burst transfer request, the scheduling information based on the scheduled time index.
  • a burst scheduler is provided for performing this method.
  • a software medium that permits a general purpose computer to carry out this method.
  • a core node in a data network includes a space switch, a plurality of input ports, a plurality of output ports and a slave controller for the space switch for receiving instructions from a master controller of the space switch, the instructions including specifications of temporary connections to establish between the plurality of input ports and the plurality of output ports and indications of timing with which to establish the connections.
  • a data network including a plurality of edge nodes, a plurality of core nodes, each core node of the plurality of core nodes including a space switch and a master controller for one the space switch in one the core node for: receiving a stream of burst transfer requests from one of the plurality of edge nodes, each of the burst transfer requests including parameters specifying a requested connection and a duration for the requested connection; generating scheduling information for each of the burst transfer requests based on the parameters; transmitting the scheduling information to the one of the plurality of edge nodes; and transmitting the instructions to a slave controller for the one the space switch, where the instructions are based on the scheduling information.
  • FIG. 1 schematically illustrates a hub and spoke network including a core node that may employ embodiments of the present invention
  • FIG. 2 illustrates the core node of FIG. 1 ;
  • FIG. 3 illustrates a master controller for use in the core node of FIG. 2 ;
  • FIG. 4 illustrates a burst scheduler for use in the space switch controller of FIG. 3 ;
  • FIG. 5A illustrates a data structure for use in an embodiment of the present invention
  • FIG. 5B illustrates an entry in the data structure of FIG. 5A ;
  • FIG. 6 illustrates a time-space map for use in an embodiment of the present invention
  • FIG. 7 illustrates an M-entry Map for use in an embodiment of the present invention
  • FIG. 8 illustrates steps of a burst scheduling method for use in an embodiment of the present invention
  • FIG. 9 illustrates steps of a map maintenance method for use in an embodiment of the present invention.
  • FIG. 10 illustrates an exemplary configuration of groups of ports of a space switch for parallel processing in an embodiment of the present invention
  • FIG. 11 illustrates a data structure adapted from the data structure in FIG. 5A for use in a parallel processing embodiment of the present invention
  • FIG. 12 illustrates a data network for use with an embodiment of the present invention
  • FIG. 13 illustrates an edge node for use in the data network of FIG. 12 ;
  • FIG. 14 illustrates an electronic core node for use in the data network of FIG. 12 ;
  • FIG. 15 illustrates a data network that is an adaptation of the data network of FIG. 12 wherein a core node and an edge node have been collocated;
  • FIG. 16 illustrates an edge node for use in the data network of FIG. 15 ;
  • FIG. 17 illustrates a master controller including a burst scheduler for use in the data network of FIG. 15 ;
  • FIG. 18 illustrates a core node for use in the data network of FIG. 15 ;
  • FIG. 19 illustrates a data network that is an adaptation of the data network of FIG. 15 wherein a second core node and a second edge node have been collocated;
  • FIG. 20 illustrates an edge node for use in the data network of FIG. 19 ;
  • FIG. 21 depicts a master time counter cycle and a calendar cycle for a master controller for use in an embodiment of the present invention
  • FIG. 22 illustrates scheduling of burst transfers and resultant changes in the state of a calendar in an embodiment of the present invention.
  • FIG. 23 illustrates a master controller including a burst scheduler and a circuit scheduler for use in the data network of FIG. 19 .
  • FIG. 1 illustrates a rudimentary “hub and spoke” data network 100 wherein a number of edge nodes 108 A, 108 B, 108 C, 108 D, 108 E, 108 F, 108 G, 108 H (referred to individually or collectively as 108 ) connect to each other via a core node 102 .
  • An edge node 108 includes a source node that supports traffic sources and a sink node that supports traffic sinks. Traffic sources and traffic sinks (not shown) are usually paired and each source node is usually integrated with a sink node with which it shares memory and control.
  • the core node 102 may be considered in greater detail in view of FIG. 2 , which illustrates an electronic core node.
  • the core node 102 includes N input ports 202 A, 202 B, 202 C, . . . , 202 N (referred to individually or collectively as 202 ) for receiving data from the edge nodes 108 of FIG. 1 .
  • Each of the N input ports 202 is connected to a corresponding buffer 204 A, 204 B, 204 C, . . . , 204 N (referred to individually or collectively as 204 ) that is connected to a corresponding port controller 206 A, 206 B, 206 C, . . . , 206 N (referred to individually or collectively as 206 ).
  • a space switch 212 directs input received from each of the buffers 204 to an appropriate one of M output ports 208 A, 208 B, 208 C, . . . , 208 M (referred to individually or collectively as 208 ) under control of a slave space switch controller 214 .
  • N number of inputs
  • M number of outputs
  • a master controller 210 is communicatively coupled to the port controllers 206 and the output ports 208 as well as to the slave space switch controller 214 .
  • Each of the control functions of the master controller 210 can be implemented in application-specific hardware, which is the preferred implementation when high speed is a requirement.
  • the master controller 210 may be loaded with burst scheduling and time locking software for executing methods exemplary of this invention from a software medium 224 which could be a disk, a tape, a chip or a random access memory containing a file downloaded from a remote source.
  • the master controller 210 includes a processor 302 .
  • the processor 302 maintains connections to a memory 304 , an input interface 306 , an output interface 308 , a switch interface 312 and a master time counter 314 .
  • the master controller 210 receives burst transfer requests from the port controllers 206 .
  • the master controller 210 may communicate with the output ports 208 to perform conventional operational and maintenance functions.
  • the processor 302 is also connected to a burst-scheduling kernel 310 . Based on the burst transfer requests received from the processor 302 , the burst-scheduling kernel 310 determines appropriate timing for switching at the space switch 212 .
  • the processor 302 passes scheduling information to the slave space switch controller 214 via the switch interface 312 .
  • the processor 302 also controls the timing of transmission of bursts, from the buffers 204 to the space switch 212 , by transmitting scheduling information to the port controllers 206 via the input interface 306 .
  • the burst-scheduling kernel 310 may now be described in view of FIG. 4 .
  • the burst-scheduling kernel 310 receives burst transfer requests from the processor 302 via a processor interface 402 and a burst parameter receiver 404 .
  • the burst parameter receiver 404 may, for instance, be implemented as a time slotted bus.
  • the parameters of these bursts are queued at a burst parameter queue 406 before being accessed by a burst-scheduling unit 408 .
  • Included in the burst-scheduling unit 408 may be a time-space map and a space-time map as well as comparators and selectors for generating scheduling information (co-ordination between these maps).
  • the maps are implemented in partitioned random-access memories. After generating scheduling information for a burst, the scheduling information is transferred to the processor 302 via a schedule transmitter 410 and the processor interface 402 .
  • an input port 202 A of core node 102 receives a burst from a subtending edge node 108 .
  • the burst is stored in the buffer 204 A.
  • Parameters indicating the size (e.g., two megabits) and destination (e.g., a particular edge node 108 B) of the burst are communicated from the port controller 206 A to the master controller 210 as a burst transfer request.
  • the burst-scheduling unit 408 of the master controller 210 executes a burst scheduling algorithm to generate scheduling information and communicates relevant parts of the generated scheduling information to the port controllers 206 .
  • the master controller 210 also communicates relevant parts of the generated scheduling information to the slave space switch controller 214 .
  • the buffer 204 A sends bursts to the space switch 212 .
  • a connection is established between the buffer 204 A and the output port 208 B, according to instructions received from the slave space switch controller 214 , such that the burst is successfully transferred from an edge node 108 associated with the traffic source to the edge node 108 associated with the traffic sink.
  • the burst transfer request is received by the input interface 306 and passed to the processor 302 .
  • the processor 302 then sends the burst transfer request to the burst-scheduling kernel 310 .
  • the burst transfer request is received at the processor interface 402 and the included burst parameters are extracted at the burst parameter receiver 404 .
  • the parameters are queued at the burst parameter queue 406 and subsequently stored at the burst-scheduling unit 408 in a data structure 500 ( FIG. 5A ).
  • the parameters are stored as an entry 506 in a record 504 , where the entry 506 is associated with the burst described by the received parameters.
  • Each record 504 has a plurality of entries 506 , and each entry 506 is associated with a burst waiting in a buffer 204 . As the number of bursts waiting in each buffer 204 may be different, the records 504 may be of varying sizes. As well, the plurality of entries 506 in each record 504 may be a linked list as will be described hereinafter.
  • the data structure 500 is made up of N records 504 , where each record 504 corresponds to one of the N input ports 202 ( FIG. 2 ). As illustrated in FIG. 5B , each entry 506 includes a destination field 508 for storing the destination parameter of the burst and a size field 510 for storing the transfer-time (size) parameter of the burst.
  • a generic memory device storing an array that has a time-varying number of data units must have a sufficient capacity to store the expected maximum number of data units. If several arrays, each having a time-varying number of data units, share the generic memory device, then the allocation of the expected maximum number of data units for each array may be considered wasteful.
  • the data structure 500 stores entries 506 containing parameters of burst transfer requests received from each of the input ports 202 .
  • the number of entries 506 for any particular input port 202 may vary violently with time, i.e., number of entries 506 for the particular input port 202 may have a high coefficient of variation.
  • interleaved linked lists are well known in the art and are not described here. Essentially, interleaved linked lists allow dynamic sharing of a memory by X (where X>1) data groupings using X insertion pointers and X removal pointers. Thus, the interleaved linked lists are addressed independently but they share the same memory device.
  • the number, X, of data groupings in the data structure 500 is at least equal to the number of input ports, N, though X may be higher than N if traffic classes are introduced. X may also be higher than N if data from a source node to a sink node uses multiple paths through different core nodes (as will be described hereinafter), since the data of each path must be identified.
  • the use of an interleaved linked list is preferred to the use of a memory structured to provide a fixed memory partition per traffic stream.
  • a traffic stream is an aggregation of traffic from a particular source edge node 108 to a particular destination edge node 108 , often resulting in a succession of bursts.
  • the burst-scheduling unit 408 maintains two other data structures, namely a calendar (i.e., a time-space map) 600 (see FIG. 6 ) and an M-element array (i.e., a space-time map) 700 (see FIG. 7 ).
  • a calendar i.e., a time-space map
  • M-element array i.e., a space-time map
  • the calendar 600 is divided into K time slots 604 ; indexed from 1 to K. Some of the time slots 604 in the calendar 600 contain identifiers 606 of input ports 202 . Those time slots 604 that do not contain input port identifiers 606 contain, instead, null identifiers 608 . Each time slot 604 contains either an input port identifier 606 or a null identifier 608 .
  • the presence, in a given time slot 604 , of a particular input port identifier 606 indicates to the master controller 210 that an input port 202 (an identifier of which is contained in a particular input port identifier 606 ) is available to transmit data (if it has waiting data) to the space switch 212 from the time corresponding to the given time slot 604 forward.
  • Each of the time slots 604 in the calendar 600 is representative of a short time period, say 100 nanoseconds.
  • time slot 604 in the calendar 600 the instant of time at which a given input port 202 is determined to be available is represented by a time slot 604 in the calendar 600 .
  • This will typically force a rounding up of the actual availability time to a nearest time slot 604 .
  • the duration of a time slot 604 in the calendar 600 therefore, should be small enough to permit an accurate representation of time and should be large enough to reduce the mean number of times a memory holding the calendar 600 has to be accessed before finding an indication of an input port 202 .
  • time slots 604 in the calendar 600 contain null identifiers 608 (i.e., all the time slots 604 that don not contain an input port identifier 606 ) and these must be read since the calendar 600 must be read sequentially.
  • the memory holding the calendar 600 must be a random-access memory however, since an address (index) at which an input port identifier 606 is written is arbitrary.
  • the number, K, of time slots 604 in the calendar 600 is significantly larger than the number of input ports 202 , N (each port of the space switch 212 has an entry in the calendar, even if the port is not active for an extended period of time).
  • K must be greater than N, where N time slots 604 contain input port identifiers 606 and (K-N) time slots 604 contain null identifiers 608 .
  • the duration of the calendar 600 must be larger than a maximum burst span. With a specified maximum burst span of 16 milliseconds, for example, an acceptable number (K) of time slots 604 in the calendar 600 is 250,000 with a slot time of 64 nanoseconds.
  • each time slot 604 in the calendar 600 has a duration equivalent to a single tick of the master time counter 314 .
  • each time slot 604 in the calendar 600 has a duration equivalent to an integer multiple of the duration of a single tick of the master time counter 314 .
  • Each port controller 206 has an awareness of time at the master time counter 314 , so that scheduling information received at the port controller 206 may be used to send a burst to the space switch 212 at the time indicated by scheduling information. This awareness may be derived from access to a clock bus or through a time locked local counter.
  • the calendar 600 may be implemented in multiple memory devices.
  • a calendar of 262,144 (2 18 ) time slots 604 can be implemented in 16 memory devices each having a capacity to store of 16,384 time slots 604 . Addressing a time slot 604 in a multiple-memory calendar is known in the art.
  • each element 704 corresponds to one of the output ports 208 .
  • Each element 704 in the M-element array 700 holds a state-transition-time indicator 706 .
  • a sixteen thousand slot calendar 600 may accommodate bursts having a length up to 1.6 milliseconds (i.e., 16 megabits at ten gigabits per second) without having to wrap around the current time when writing the availability of the input port 202 to the calendar 600 .
  • the burst-scheduling unit 408 scans the calendar 600 to detect a future time slot 604 containing an input port identifier 606 (step 802 ), resulting in a detected time slot 604 A.
  • the burst-scheduling unit 408 then communicates with the burst parameter queue 406 to acquire entries 506 (step 804 ) from the record 504 , in the data structure 500 ( FIG. 5 ), that corresponds to the input port 202 identified in the input port identifier 606 in the detected time slot 604 A. It is then determined whether there are entries 506 in the record 504 that corresponds to the identified input port 202 (step 805 ).
  • Each of the entries 506 identifies a destination and, from the destination, the burst-scheduling unit 408 may deduce an output port 208 . If there are entries to schedule (i.e., waiting burst requests), the burst-scheduling unit 408 extracts a state-transition-time indicator 706 (step 806 ) from each element 704 , in the M-element array 700 ( FIG. 7 ), that corresponds to an output port 208 deduced from destinations identified by the acquired entries 506 .
  • the burst-scheduling unit 408 determines a “gap” (step 808 ) by subtracting the index of the detected time slot 604 A from the index of the time slot found in each state-transition-time indicator 706 .
  • Each gap represents a time difference between a time at which the input port 202 is available and a time at which the respective output port 208 , requested in the respective burst transfer request, is available.
  • the burst-scheduling unit 408 does this for each of the acquired entries 506 for the input port 202 .
  • Each entry 506 identifies a single burst transfer request.
  • the burst-scheduling unit 408 selects the burst transfer request corresponding to the minimum gap (step 810 ).
  • the step of acquiring entries 506 from the record 504 may only require acquisition of a limited number of entries 506 .
  • the input port 202 is available before the output port 208 .
  • the time slot index identified in the state-transition-time indicator 706 corresponding to the availability of the output port 208 which was requested for the selected burst transfer request is then designated as a “scheduled time slot.” If the gap of the selected burst transfer request is negative, then the input port 202 is available after the output port 208 .
  • the time slot index in which the input port identifier 606 was detected in step 802 (corresponding to the time when the input port 202 is available) is then designated as the scheduled time slot.
  • the burst-scheduling unit 408 then transmits scheduling information (index of the scheduled time slot and identity of the burst transfer request) to the processor 302 (step 812 ) via the schedule transmitter 410 and the processor interface 402 .
  • scheduling information index of the scheduled time slot and identity of the burst transfer request
  • the processor 302 step 812
  • a negative gap is preferred to a positive gap because use of the input port 202 may begin at the time corresponding to the detected time slot 604 A, as the negative gap indicates that the requested output port 208 is already available.
  • the burst-scheduling unit 408 then updates the calendar 600 and the M-element array 700 (step 814 ).
  • FIG. 9 illustrates steps of the update method of step 814 .
  • the burst-scheduling unit 408 first sums the index of the scheduled time slot and the transfer-time determined from the size field 510 of the selected burst transfer request (step 902 ) and writes the input port identifier 606 of the selected burst transfer request in the time slot 604 indexed by the sum (step 904 ).
  • the writing of the input port identifier 606 effectively identifies, to the burst-scheduling unit 408 , the time at which the input port 202 will be available after transferring the burst corresponding to the selected burst transfer request.
  • the burst-scheduling unit 408 After writing the input port identifier 606 to the time slot 604 indexed by the sum, the burst-scheduling unit 408 writes a null identifier 608 in the scheduled time slot (step 906 ).
  • the burst-scheduling unit 408 writes a state-transition-time indicator 706 to the M-element array 700 (step 908 ) in the element 704 corresponding to the output port 208 of the selected burst transfer request.
  • the state-transition-time indicator 706 is an index of the time slot 604 indexed by the sum determined in step 902 .
  • pipelining techniques may also be used to reduce processing time.
  • the burst-scheduling unit 408 If, as determined in step 805 , there are no entries to schedule (i.e., waiting burst requests), the burst-scheduling unit 408 generates an artificial burst (step 816 ) where the size of the artificial burst is the “size of the selected burst” as far as step 902 is concerned.
  • the result of this generation of an artificial burst is that (in step 814 ) the input port identifier 606 is written to a deferred time slot 604 .
  • the processor 302 having received the scheduling information, transmits to the appropriate port controller 206 , via the input interface 306 , scheduling information to indicate a time at which to begin sending the burst corresponding to the selected burst transfer request to the space switch 212 .
  • the processor 302 also sends scheduling information (input-output configuration instructions) to the slave space switch controller 214 via the switch interface 312 .
  • the master controller 210 has already been operating, it is worth considering initial conditions, for the calendar 600 especially.
  • the first N time slots 604 may be occupied by the input port identifiers 606 that identify each of the N input ports 202 .
  • the data structure 500 is clear of burst transfer requests and the state-transition-time indicator 706 present in each element 704 of the M-element array 700 may be an index of the first time slot 604 in the calendar 600 .
  • the corresponding record 504 in the data structure 500 is accessed to acquire entries 506 . If the corresponding record 504 is found to be empty, the burst-scheduling unit 408 writes a null identifier 608 in the detected time slot 604 A and writes the input port identifier 606 at a deferred time slot.
  • the deferred time slot may be separated from the detected time slot 604 A by, for example, 128 time slots. At 100 nanoseconds per time slot 604 , this would be amount to a delay of about 13 microseconds.
  • the M-element array 700 ( FIG. 7 ) can only respond to a single read request at a time, the requests to read each state-transition-time indicator 706 from the elements 704 will be processed one after the other. To conserve time then, it may be desirable to maintain multiple identical copies of the M-element array 700 . Where multiple copies are maintained, extraction of a state-transition-time indicator 706 from elements 704 in step 806 may be performed simultaneously. It is preferable that the writing of a particular state-transition-time indicator 706 to a given element 704 of each copy of the M-element array 700 (step 908 ) be performed in a parallel manner.
  • the number of entries 506 acquired in step 804 should be limited to a value, J. If J entries 506 are acquired in step 804 , then there is only a requirement for J identical copies of the M-element array 700 . It is preferred that J not exceed four.
  • the master controller 210 may take advantage of a parallel processing strategy to further conserve processing time.
  • a parallel processing strategy may, for instance, involve considering a 64 by 64 space switch (64 input ports, 64 output ports) as comprising an arrangement of four 16 by 16 space switches. However, so that each input may be connected to any output, four arrangements must be considered.
  • An exemplary configuration 1000 for considering these arrangements is illustrated in FIG. 10 .
  • the exemplary configuration 1000 includes four input port groups (sub-sets) 1002 A, 1002 B, 1002 C, 1002 D (referred to individually or collectively as 1002 ) and four output port groups (sub-sets) 1004 A, 1004 B, 1004 C, 1004 D (referred to individually or collectively as 1004 ).
  • Each input port group includes 16 input ports and each output port group includes 16 output ports.
  • processors may perform scheduling for the 64 by 64 space switch, where each processor schedules on behalf of one input port group 1002 .
  • a scheduling session may be divided into as many scheduling time periods as there are processors. For each scheduling time period, a given processor (scheduling on behalf of one input port group 1002 ) will schedule only those connections destined for a particular output group 1004 .
  • the output group changes after every scheduling time period such that, by the end of the scheduling session, all four output port groups 1004 have been considered for connections from the input port group 1002 corresponding to the given processor.
  • the state of the exemplary configuration 1000 at a particular scheduling time period is illustrated in FIG. 10 .
  • the intersection of the output port group 1004 with the corresponding input port group 1002 for the particular scheduling time period is identified with a bold border.
  • a parallel processing data structure 1100 which is an alternative to the data structure 500 illustrated in FIG. 5A , is illustrated in FIG. 11 .
  • Each of the N records 1104 in the parallel processing data structure 1100 is divided into sub-records, where each sub-record in a given record 1104 corresponds to a single output port group 1004 .
  • Parameters of received burst transfer requests are stored as entries 506 in a record 1104 according to the input port 202 and in a sub-record according to the output port group 1004 .
  • the sub-records that correspond to the output port groups 1004 are illustrated in FIG. 11 as a number of rows 1102 A, 1102 B, 1102 C, 1102 D.
  • the input port identifier 606 must be from the input port group 1002 to which the given processor corresponds.
  • the given processor then communicates with the burst parameter queue 406 to acquire entries 506 (step 804 ) from the parallel processing data structure 1100 .
  • the entries 506 are acquired from the record 1104 that corresponds to the input port 202 identified in the input port identifier 606 in the detected time slot 604 and, furthermore, only from the sub-record corresponding to the output port group 1004 under consideration by the given processor in the current scheduling time period.
  • the row 1102 A of sub-records corresponding to the output port group 1004 under consideration by the given processor associated with a particular input port group 1002 A (which includes input ports N-3, N-2, N-1 and N) is identified with a bold border.
  • a hub and spoke data network 1200 is illustrated in FIG. 12 , including a bufferless core node 1210 X in place of the core node 102 .
  • a number of traffic sources 104 A, 104 B, 104 C, 104 N (referred to individually or collectively as 104 ) connect, via the edge nodes 108 and the bufferless core node 1210 X, to a number of traffic sinks 106 A, 106 B, 106 C, 106 M (referred to individually or collectively as 106 ).
  • the traffic sources 104 and the traffic sinks 106 are integrated, for instance, as a personal computer.
  • a space switch and space switch controller are maintained at the bufferless core node 1210 X.
  • An edge node 108 is illustrated in FIG. 13 .
  • Traffic is received from the traffic sources 104 or sent to the traffic sinks 106 at traffic interfaces 1302 A, 1302 B, 1302 C (referred to individually or collectively as 1302 ).
  • the traffic interfaces 1302 connect to buffers 1304 A, 1304 B, 1304 C (referred to individually or collectively as 1304 ).
  • the buffers 1304 are controlled by buffer controllers 1306 A, 1306 B, 1306 C (referred to individually or collectively as 1306 ) with regard to the timing of passing traffic to a core interface 1308 X that subsequently passes the traffic to the bufferless core node 1210 X.
  • the buffer controllers 1306 also connect to the core interface 1308 X for sending, to the bufferless core node 1210 X, burst transfer requests in a manner similar to the manner in which the port controllers 206 send burst transfer requests to the master controller 210 in FIG. 2 .
  • the core interface 1308 X maintains a connection to a slave time counter 1314 for time locking with a master time counter in a master controller.
  • a space switch 1412 connects N input ports 1402 A, 1402 B, 1402 C, . . . , 1402 N (referred to individually or collectively as 1402 ) to M output ports 1408 A, 1408 B, 1408 C, . . . , 1408 M (referred to individually or collectively as 1408 ) under control of a slave space switch controller 1414 .
  • Each of the N input ports 1402 is arranged to send burst transfer requests received from the edge nodes 108 to a master controller 1410 and to send burst traffic to the space switch 1412 .
  • a particular input port 1402 is arranged to receive a Wavelength Division Multiplexed (WDM) signal having 16 channels
  • WDM Wavelength Division Multiplexed
  • one channel i.e., one wavelength
  • the master controller 1410 passes scheduling information to the slave space switch controller 1414 .
  • the master controller 1410 may consult the edge nodes 108 , via the output ports 1408 , to perform conventional operational and maintenance functions. However, to avoid consulting the edge nodes 108 , edge-to-edge rate allocations may be introduced and updated as the need arises. The interval between successive updates may vary between 100 milliseconds and several hours, which is significantly larger than a mean burst duration.
  • a traffic interface 1302 A at a source edge node 108 A receives a burst from a subtending traffic source 104 A.
  • the burst is stored in the buffer 1304 A.
  • Parameters indicating the size and destination (e.g., a destination edge node 108 E) of the burst are communicated from the buffer controller 1306 A, via the core interface 1308 X, to the bufferless core node 1210 X in a burst transfer request.
  • the burst transfer request is received at one of the input ports 1402 and sent to the master controller 1410 .
  • the master controller 1410 executes a burst scheduling algorithm to generate scheduling information and communicates relevant parts of the generated scheduling information to the edge nodes 108 .
  • the master controller 1410 also communicates relevant parts of the generated scheduling information to the slave space switch controller 1414 .
  • the buffer 1304 A sends the burst to the bufferless core node 1210 X, via the core interface 1308 X, according to the scheduling information received at the buffer controller 1306 A.
  • a connection is established between the input port 1402 A and the output port 1408 B such that the burst is successfully transferred from the source edge node 108 A to the destination edge node 108 E.
  • the duty of routing of burst transfer requests to the master controller 1410 and bursts to the space switch 1412 may present a problem to the design of the input ports 1402 if the space switch 1412 is optical.
  • One solution to this problem is to relieve the input ports 1402 of this duty.
  • a bufferless core node 1210 Z is collocated with an edge node 108 J at a location 112 .
  • a stand-alone master controller 1610 Z exists separate from the bufferless core node 1210 Z.
  • the collocated edge node 108 J maintains a connection to the stand-alone master controller 1610 Z for transferring burst transfer requests, received from other edge nodes 108 (via the bufferless core node 1210 Z) and the subtending traffic sources 104 , to the space switch controller in the bufferless core node 1210 Z.
  • the collocated edge node 108 J is illustrated in detail. Like the typical edge node 108 of FIG. 13 , the collocated edge node 108 J includes traffic interfaces 1602 A, 1602 B, 1602 C, buffers 1604 A, 1604 B, 1604 C, buffer controllers 1606 A, 1606 B, 1606 C (referred to individually or collectively as 1606 ) and a core interface 1608 X.
  • the core interface 1608 X also maintains a connection to a slave time counter 1614 Z for time locking with a master time counter in the master controller 1610 Z.
  • the collocated edge node 108 J also includes a controller interface 1612 for sending burst transfer requests to the stand-alone master controller 1610 Z.
  • the buffer controllers 1606 communicate burst transfer requests to the controller interface 1612 rather than to the core interface 1608 X, as is the case in the typical edge node 108 in FIG. 13 .
  • the core interface 1608 X also communicates other burst transfer requests to the controller interface 1612 , in particular, burst transfer requests received from other edge nodes 108 .
  • the stand-alone master controller 1610 Z generates scheduling information based on the burst transfer requests and sends the scheduling information to the slave space switch controller in the bufferless core node 1210 Z.
  • the stand-alone master controller 1610 Z includes a processor 1702 .
  • the processor 1702 maintains connections to a memory 1704 , an edge node interface 1706 , a core node interface 1712 and a master time counter 1714 .
  • the master controller 210 receives burst transfer requests from the collocated edge node 108 J.
  • the processor 1702 is also connected to a burst-scheduling kernel 1710 . Based on the burst transfer requests received from the processor 1702 , the burst-scheduling kernel 1710 determines appropriate timing for switching at the space switch at the bufferless core node 1210 Z.
  • the processor 1702 passes scheduling information to the bufferless core node 1210 Z via the core node interface 1712 .
  • the processor 1702 also controls the timing of transmission of bursts, from the edge nodes 108 to the bufferless core node 1210 Z, by transmitting scheduling information to the edge nodes 108 via the edge node interface 1706 and the collocated edge node 108 J.
  • a space switch 1812 connects N input ports 1802 A, 1802 B, 1802 C, . . . , 1802 N (referred to individually or collectively as 1802 ) to M output ports 1808 A, 1808 B, 1808 C, . . . , 1808 M (referred to individually or collectively as 1808 ) under control of a slave space switch controller 1814 .
  • burst transfer requests pass through the bufferless core node 1210 Z and are sent to the collocated edge node 108 J.
  • the collocated edge node 108 J then forwards the burst transfer requests to the stand-alone master controller 1610 Z, where scheduling information is generated.
  • the scheduling information is received from the stand-alone master controller 1610 Z by a master controller interface 1816 .
  • the slave space switch controller 1814 then receives the scheduling information from the master controller interface 1816 .
  • the bufferless core node 1210 Z need not be limited to a single space switch 1812 .
  • the bufferless core node 1210 Z may include an assembly of multiple parallel space switches (not shown).
  • Each of the multiple space switches may require an associated burst-scheduling kernel, such as the burst-scheduling kernel 1710 in FIG. 17 , to be located at the master controller 1610 Z of the bufferless core node 1210 Z.
  • each of the multiple space switches may be associated with a unique burst scheduling unit (see 408 in FIG. 4 ).
  • the space switches in the assembly of multiple parallel space switches operate totally independently.
  • the traffic to a specific edge node 108 may, however, be carried by any of the channels of a multi-channel link (WDM fiber link) from a source edge node 108 to the bufferless core node 1210 .
  • a load-balancing algorithm (not described herein) is used to balance the traffic and thus increase throughput and/or decrease scheduling delay.
  • Successive bursts to the same sink edge node 108 may be transferred using different channels (different wavelengths) and, hence, may be switched in different space switches in the bufferless core node 1210 .
  • the transfer of successive bursts to the same sink edge node 108 using different channels should not be expanded to include the transfer of successive bursts to the same sink edge node 108 using different links where the delay differential between links (possibly several milliseconds) may complicate assembly of the bursts at the sink edge node 108 .
  • An advantage of burst switching is a freedom to select a space switch on a per-burst basis, as long as a predetermined time separation (a microsecond or so) is provided between successive bursts of a single data stream.
  • the time separation is required to offset the effect of propagation delay differentials present in different wavelengths of the same WDM signal.
  • propagation delay may be considered in view of the data network 1200 . If the edge node 108 A is one kilometer away from the bufferless core node 1210 X, scheduling information may take five microseconds to pass from the bufferless core node 1210 X to the edge node 108 A in an optical-fiber link. Similarly, a burst sent from the edge node 108 A would take five microseconds to travel to the bufferless core node 1210 X. A time period lasting five microseconds is represented in the calendar 600 by 500 time slots 604 of 100-nanoseconds each.
  • a burst may arrive at the bufferless core node 1210 X after the time at which the burst was scheduled to be passing through the space switch 1412 . Consequently, given knowledge, at the bufferless core node 1210 X, of an estimate of a maximum round trip propagation delay associated with the edge nodes 108 , scheduling can be arranged to take the propagation delay into account. For instance, the burst-scheduling kernel 1710 may schedule such that the earliest a burst may be scheduled, relative to a current time in the master time counter 1714 , is at least the estimated maximum round trip propagation delay time into the future.
  • propagation delay differential was not a problem in the core node 102 of FIG. 2 , which had input buffers.
  • the collocation of the collocated edge node 108 J with the bufferless core node 1210 Z in FIG. 15 removes concern of propagation delay differentials for traffic originating at the traffic sources 104 A, 104 B connected to the collocated edge node 108 J.
  • a time locking scheme is required so that bursts may be correctly scheduled.
  • the propagation delay between the time at which a burst leaves one of the other edge nodes 108 (i.e., the edge nodes 108 that are not collocated with the bufferless core node 1210 Z) and the time at which the burst arrives at the bufferless core node 1210 Z may be different for each of the other edge nodes 108 .
  • the other edge nodes 108 To switch these bursts, without contention or a requirement for burst storage at the bufferless core node 1210 Z, the other edge nodes 108 must be time locked to the bufferless core node 1210 Z.
  • a time locking technique also called time coordination, is described in the applicant's U.S. patent application Ser. No. 09/286,431, filed on Apr. 6, 1999, and entitled “Self-Configuring Distributed Switch,” the contents of which are incorporated herein by reference. With time locking, the scheduling method in accordance with the present invention guarantees that bursts arrive to available resources at the bufferless core node 1210 Z.
  • each other edge node 108 may “time lock,” with the collocated edge node 108 J.
  • each edge node 108 includes at least one local time counter (e.g., the slave time counter 1314 of FIG. 13 ) of equal width W.
  • a time locking request may be sent from a particular edge node 108 E ( FIG. 15 ), while noting the sending time (i.e., the value of the slave time counter at the particular edge node 108 E when the time locking request is sent), to the master controller 1610 Z.
  • the arrival time i.e., the value of the master time counter 1714 at the arrival time
  • a time locking response is generated, including an indication of the arrival time, and sent to the particular edge node 108 E.
  • a time difference between sending time and arrival time is determined at the particular edge node 108 E and used to adjust the slave time counter at the particular edge node 108 E.
  • scheduling information is received at the particular edge node 108 E from the stand-alone master controller 1610 Z, for instance, “start sending burst number 73 at a time counter state 3564 .” If the particular edge node 108 E starts sending burst number 73 at slave time counter state 3564 , the beginning of the burst will arrive at the bufferless core node 1210 Z at master time counter state 3564 .
  • the duration of each time counter cycle is equal and substantially larger than a maximum round-trip propagation delay from any edge node 108 to any core node 1210 in the data network 1200 .
  • the maximum round-trip propagation delay should be taken into account when performing scheduling at the stand-alone master controller 1610 Z.
  • the counters related to the time locking scheme are included in the controller interface 1612 of the collocated edge node 108 J of FIG. 16 and in the core interface of the generic edge node 108 of FIG. 13 .
  • FIG. 19 illustrates the data network 1200 supplemented with an additional bufferless core node 1210 Y.
  • a flow control process which operates at a higher level than the switch operations, may assign one of the bufferless core nodes 1210 Z, 1210 Y (referred to individually or collectively as 1210 ) to each traffic stream originating at a particular edge node 108 , where a traffic stream is an aggregation of traffic with identical source and destination edge nodes 108 .
  • the given edge node 108 may send a burst transfer request to the core node (say bufferless core node 1210 Z) assigned to the traffic stream of which the burst is part. Scheduling information is returned to the given edge node 108 . The given edge node 108 may then send the burst to the assigned bufferless core node 1210 Z according to the timing represented in the scheduling information.
  • the additional bufferless core node 1210 Y is illustrated as collated with an edge node 108 K at an additional location 114 .
  • An additional master controller 1610 Y corresponding to the additional bufferless core node 1210 Y, is also present at the additional location 114 .
  • An edge node 108 communicates with all core nodes 1210 in the sending and receiving modes. As such, the edge nodes 108 should be adapted to communicate with more than one bufferless core node 1210 .
  • This adaptation is shown for the collocated edge node 108 J in FIG. 20 .
  • Notably different from the collocated edge node 108 J as illustrated in FIG. 16 is the addition of a core interface 1608 Y corresponding to the bufferless core node 1210 Y.
  • the core interface 1608 Y corresponding to the bufferless core node 1210 Y requires a connection to a slave time counter 1614 Y.
  • a slave time counter at a given edge node 108 be time locked to the master time counter of the master controller 1610 of the bufferless core node 1210 .
  • the scheduling information transmitted by the master controller 1610 to the edge nodes 108 is based on the time indication of the master time counter 1714 as it corresponds to the scheduled time slot in the calendar 600 .
  • the time slots 604 in the calendar 600 must, therefore, also be time locked to the master time counter 1714 .
  • the selection of the time counter cycle in use at the master time counter 1714 and the calendar cycle are important design choices. Where a master time counter 1714 counts using W bits, the duration of the master time counter cycle is 2 W multiplied by the duration of a period of a clock used to drive the master time counter.
  • the number of counter states is about 4.29 ⁇ 10 9 and duration of the master time counter is more than 68 seconds. This is orders of magnitude higher than the round-trip propagation delay between any two points on Earth (assuming optical transmission).
  • Increasing the duration of the master time counter 1714 involves adding a few bits, resulting in a very small increase in hardware cost and transport of time locking signals across the network.
  • increasing the duration of the calendar 600 requires increasing the depth of a memory used to maintain the calendar 600 and/or increasing the duration of each time slot 604 in the calendar. The latter results in decreasing the accuracy of time representation, and hence in wasted time, as will be explained below.
  • each time slot 604 has a duration of eight microseconds and the number of calendar time slots 604 is 65,536, the duration of the calendar 600 is more than 500 milliseconds.
  • a time slot 604 of eight microseconds is, however, comparable with the duration of a typical burst. At 10 Gb/s, an eight microsecond bursts is about ten kilobyte long. It is desirable that the duration of each time slot 604 be a small fraction of the mean burst duration. A reasonable duration for a time slot 604 is 64 nanoseconds. However, if the duration of the calendar 600 is to be maintained at 500 milliseconds, the calendar 600 requires eight million slots.
  • a compromise is to select a duration of the calendar 600 that is just sufficient to handle the largest possible burst and use an associated adder or cycle counter to be cognizant of the calendar time relationship to the master time counter time.
  • the largest burst duration would be imposed by a standardization process. In a channel of 10 Gb/s, a burst of one megabyte has a duration of less than one millisecond. A standardized upper-bound of the burst length is likely to be even less than one megabyte in order to avoid delay jitter.
  • the duration of the calendar 600 can be selected to be less than 16 milliseconds. With a duration of each time slot 604 set to 64 nanoseconds, the number of required time slots 604 would be about 262,144. This can be placed in four memory devices of 65,536 words each, a word corresponding to a time slot 604 .
  • Relating a time slot 604 in the calendar 600 to the state of the master time counter 1714 is greatly simplified if the ratio of the number of master time counter states to the number of time slots 604 is a power of two, and the ratio of the duration of a time slot 604 to the duration of the clock used to drive the master time counter is also a power of two.
  • the number of master time counter states exceeds or equals the number of calendar slots and the duration of a calendar slots exceeds or equals the clock period.
  • the width of the master time counter is 32 bits
  • the width of a calendar address is 18 bits (2 18 , i.e., 262,144 time slots 604 )
  • the duration of a time slot 604 is four times the period of the clock used to drive the master time counter
  • duration of the master time counter is 4,096 times the duration of the calendar 600 .
  • the width of the master time counter is 24 bits, with 262,144 calendar slots, a clock period of 16 nanoseconds and a duration of each time slot 604 of 64 nanoseconds
  • the duration of the master time counter 1714 becomes about 268.72 milliseconds, which is 16 times the calendar period of about 16.77 milliseconds.
  • the master clock period is selected to be reasonably short to ensure accurate time representation for time locking purposes.
  • FIG. 21 depicts a master time counter cycle 2102 and a calendar cycle 2104 for an exemplary case wherein a duration 2112 of the master time counter cycle 2102 is exactly four times a duration 2114 of the calendar cycle 2104 .
  • Time locking of the calendar to the master time counter is essential as indicated in FIG. 21 .
  • the scheduling of future burst transfers based on burst transfer requests received from a specific edge node 108 , associated with a specific input port of a bufferless core node 1210 , is illustrated in FIG. 22 .
  • Changes in the state of a calendar 2200 are illustrated as they correspond to the specific input port.
  • the calendar 2200 has 32 time slots and is shown as four calendar cycles 2200 S, 2200 T, 2200 U, 2200 V.
  • the duration of the master time counter is four times the duration of the calendar 2200 .
  • a time slot 2202 A contains an identifier of the input port, typically an input port number.
  • Each other time slot in the calendar 2200 contains either an input port identifier or a null identifier, although, for simplicity, these identifiers are not shown.
  • the time slot 2202 A is encountered and an input port identifier 2206 is recognized.
  • the burst scheduling method of FIG. 8 is then performed, along with the map maintenance method of FIG. 9 .
  • These methods result in the input port identifier 2206 being replaced with a null identifier in time slot 2202 A and the input port identifier 2206 being written in time slot 2202 B.
  • the index of time slot 2202 D is smaller than the index of time slot 2202 C because the adder determining the index of the time slot in which to write the input port identifier 2206 (step 902 ) has a word length that exactly corresponds to the number of time slots in the calendar 2200 (note that calendar length is a power of 2).
  • calendar length is a power of 2.
  • the scheduling procedure described above enables scheduling bursts for a look-ahead period as large as the duration of the master time counter.
  • the duration of the master time counter is 268 milliseconds ( 224 master time counter slots, 16 nanosecond clock period), for example, at 10 GHz, bursts of cumulative length as high as 200 megabytes can be scheduled.
  • a scheduled time indication i.e., the master time counter state corresponding to the scheduled time slot, to be reported to a respective edge node
  • an indication of the relative calendar cycle number with respect to the master time counter cycle must be provided along with the scheduled time slot.
  • this indication is 0, 1, 2 or 3.
  • the scheduled time indication that is communicated to the requesting edge node is 84.
  • a portion of the network capacity in the data network 1200 may be dedicated to relatively well-behaved traffic. That is, non-bursty traffic.
  • a master controller may include a second scheduler dedicated to more traditional circuit switching.
  • the master controller 1610 Y illustrated in FIG. 23 includes a processor 2302 .
  • the processor 2302 maintains connections to a memory 2304 , an edge node interface 2306 , a core node interface 2312 and a master time counter 2314 .
  • the master controller 1610 Y illustrated in FIG. 23 also includes a circuit-scheduling kernel 2316 for scheduling transfers between edge nodes 108 on a longer term basis.
  • the edge nodes 108 may perform some processing of bursts. This processing may include expansion of bursts to have a length that is a discrete number of segments or aggregation of small bursts.
  • the present invention is applicable without dependence on whether switching in the data network 1200 is electrical or optical and without dependence on whether transmission in the data network 1200 is wireline or wireless.
  • the optical switching example is particularly instructive, however, in that, given recent developments in Dense Wavelength Division Multiplexing, a link between an edge node 108 and a bufferless core node 1210 may include multiple (e.g., 32) channels. If the data network 1200 is to work as described in conjunction with FIG. 12 , one of the multiple channels may be completely dedicated to the transfer of burst transfer requests. However, the transfer of burst transfer requests represent a very small percentage of the available capacity of such a channel and the unused capacity of the dedicated channel is wasted.
  • the present invention allows bursts that switch through the core nodes to employ the core nodes, and associated space switches, nearly constantly, that is, with virtually no data loss. As such, the network resources are used more efficiently.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Small-Scale Networks (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

At a master controller of a space switch in a node in a data network, a request is received from a source node that requests a connection to be established through the space switch. This request is compared to other such requests so that a schedule may be established for access to the space switch. The schedule is then sent to the source nodes as well as to a slave controller of the space switch. The source nodes send data bursts which are received at the space switch during a short guard time between successive reconfigurations of the space switch. Data bursts are received at the space switch at a precisely determined instant of time that ensures that the space switch has already reconfigured to provide requested paths for the individual bursts. The scheduling is pipelined and performed in a manner that attempts to reduce mismatch intervals of the occupancy states of input and output ports of the space switch. The method thus allows efficient utilization of the data network resources while ensuring virtually no data loss.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 09,750,071 filed Dec. 29, 2000.
  • STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH
  • This work was supported by the United States Government under Technology Investment Agreement TIA F30602-98-2-0194. The Government has certain rights in this invention.
  • FIELD OF THE INVENTION
  • The present invention relates to data communication networks and, in particular, to burst switching in a high capacity network.
  • BACKGROUND OF THE INVENTION
  • In burst switching, a source node sends a burst transfer request to a core node to indicate that a burst of data is coming, the size of the burst and the destination of the burst. Responsive to this burst transfer request, the core node configures a space switch to connect a link on which the burst will be received to a link to the requested burst destination. In a first scheme, the burst follows the burst transfer request after a predetermined time period (a scheduling time) and it is expected that, when the burst arrives at the core node, the space switch will have been properly configured by the core node. In a second scheme, the source node waits for a message from the core node, where the message acknowledges that the space switch in the core node is properly configured, before sending the burst.
  • Often core nodes are used that do not have buffers to buffer incoming data. Core nodes without buffers are desirable because: it may not be possible to provide buffers without an expensive optical-electrical conversion at input and electrical-optical conversion at output of an optical space switch; and the core node may be distant from the source and sink (edge) nodes, therefore requiring remote buffer management in an edge-controlled network.
  • In the first scheme, a burst may arrive at a core node before the space switch is properly configured and, if the core node does not include a buffer, the burst may be lost. Furthermore, until the source node fails to receive an acknowledgement of receipt of the burst from the burst destination, the fact that the burst has been lost at the core node is unknown to the source node. Having not received acknowledgement of receipt of the burst, the source node may then retransmit the burst. In the second scheme, the time delay involved in sending a burst transfer request and receiving an acceptance before sending a burst may be unacceptably high, leading to low network utilization. Despite these shortcomings, burst switching is gaining popularity as a technique to transfer data in high-speed networks since it simplifies many of the control functions and does not require capacity to be reserved when it may not always be in use. Furthermore, burst switching reduces a need for characterizing the traffic. Clearly, a burst switching technique that allows for greater network utilization is desirable.
  • SUMMARY OF THE INVENTION
  • At a controller of a space switch, a novel burst scheduling technique allows efficient utilization of network resources. Burst transfer requests are received at the space switch controller and pipelined such that the controller may determine a schedule for allowing the bursts, represented by the burst transfer requests, access to the space switch. According to the schedule, scheduling information is distributed to the sources of the burst transfer requests and to a controller of the space switch.
  • Advantageously, the novel burst scheduling technique allows for utilization of network resources that is more efficient than typical burst switching techniques, especially when the novel burst scheduling technique is used in combination with known time locking methods. The novel burst scheduling technique enables the application of burst switching to wide coverage networks. Instead of handling burst requests one-by-one, burst requests are pipelined and the handling of the bursts is scheduled over a long future period.
  • In accordance with an aspect of the present invention there is provided a method of controlling a space switch to establish time-varying connections, the method includes receiving a stream of burst transfer requests from a source node, each of the burst transfer requests including parameters specifying a requested connection and a duration for the requested connection, generating scheduling information for each of the burst transfer requests based on the parameters, transmitting the scheduling information to the source node and transmitting instructions to a slave controller for the space switch, where the instructions are based on the scheduling information and instruct the space switch to establish the requested connection. In another aspect of the invention a space switch master controller is provided for performing this method. In a further aspect of the present invention, there is provided a software medium that permits a general purpose computer to carry out this method.
  • In accordance with another aspect of the present invention there is provided a method of generating scheduling information. The method includes determining a next-available input port among a plurality of input ports and a time index at which the next-available input port will become available and, for each burst transfer request of a plurality of burst transfer requests received in relation to the next-available input port, and where each the each burst transfer request includes an identity of a burst and a destination for the burst: determining, from the destination for the burst, a corresponding output port among a plurality of output ports; determining a time gap, where the time gap is a difference between: the time index at which the next-available input port will become available; and a time index at which the corresponding output port will become available. The method further includes selecting one of the plurality of burst transfer requests as a selected burst transfer request, where the selected burst transfer request has a minimum time gap of the plurality of burst transfer requests, selecting a scheduled time index, where the scheduled time index is one of the time index at which the next-available input port is available and the time index at which the corresponding output port is available and transmitting scheduling information for a burst identified by the selected burst transfer request, the scheduling information based on the scheduled time index. In another aspect of the invention a burst scheduler is provided for performing this method. In a further aspect of the present invention, there is provided a software medium that permits a general purpose computer to carry out this method.
  • In accordance with a further aspect of the present invention there is provided a core node in a data network. The core node includes a space switch, a plurality of input ports, a plurality of output ports and a slave controller for the space switch for receiving instructions from a master controller of the space switch, the instructions including specifications of temporary connections to establish between the plurality of input ports and the plurality of output ports and indications of timing with which to establish the connections.
  • In accordance with a still further aspect of the present invention there is provided a data network including a plurality of edge nodes, a plurality of core nodes, each core node of the plurality of core nodes including a space switch and a master controller for one the space switch in one the core node for: receiving a stream of burst transfer requests from one of the plurality of edge nodes, each of the burst transfer requests including parameters specifying a requested connection and a duration for the requested connection; generating scheduling information for each of the burst transfer requests based on the parameters; transmitting the scheduling information to the one of the plurality of edge nodes; and transmitting the instructions to a slave controller for the one the space switch, where the instructions are based on the scheduling information.
  • Other aspects and features of the present invention will become apparent to those of ordinary skill in the art upon review of the following description of specific embodiments of the invention in conjunction with the accompanying figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • In the figures which illustrate example embodiments of this invention:
  • FIG. 1 schematically illustrates a hub and spoke network including a core node that may employ embodiments of the present invention;
  • FIG. 2 illustrates the core node of FIG. 1;
  • FIG. 3 illustrates a master controller for use in the core node of FIG. 2;
  • FIG. 4 illustrates a burst scheduler for use in the space switch controller of FIG. 3;
  • FIG. 5A illustrates a data structure for use in an embodiment of the present invention;
  • FIG. 5B illustrates an entry in the data structure of FIG. 5A;
  • FIG. 6 illustrates a time-space map for use in an embodiment of the present invention;
  • FIG. 7 illustrates an M-entry Map for use in an embodiment of the present invention;
  • FIG. 8 illustrates steps of a burst scheduling method for use in an embodiment of the present invention;
  • FIG. 9 illustrates steps of a map maintenance method for use in an embodiment of the present invention;
  • FIG. 10 illustrates an exemplary configuration of groups of ports of a space switch for parallel processing in an embodiment of the present invention;
  • FIG. 11 illustrates a data structure adapted from the data structure in FIG. 5A for use in a parallel processing embodiment of the present invention;
  • FIG. 12 illustrates a data network for use with an embodiment of the present invention;
  • FIG. 13 illustrates an edge node for use in the data network of FIG. 12;
  • FIG. 14 illustrates an electronic core node for use in the data network of FIG. 12;
  • FIG. 15 illustrates a data network that is an adaptation of the data network of FIG. 12 wherein a core node and an edge node have been collocated;
  • FIG. 16 illustrates an edge node for use in the data network of FIG. 15;
  • FIG. 17 illustrates a master controller including a burst scheduler for use in the data network of FIG. 15;
  • FIG. 18 illustrates a core node for use in the data network of FIG. 15;
  • FIG. 19 illustrates a data network that is an adaptation of the data network of FIG. 15 wherein a second core node and a second edge node have been collocated;
  • FIG. 20 illustrates an edge node for use in the data network of FIG. 19;
  • FIG. 21 depicts a master time counter cycle and a calendar cycle for a master controller for use in an embodiment of the present invention;
  • FIG. 22 illustrates scheduling of burst transfers and resultant changes in the state of a calendar in an embodiment of the present invention; and
  • FIG. 23 illustrates a master controller including a burst scheduler and a circuit scheduler for use in the data network of FIG. 19.
  • DETAILED DESCRIPTION
  • FIG. 1 illustrates a rudimentary “hub and spoke” data network 100 wherein a number of edge nodes 108A, 108B, 108C, 108D, 108E, 108F, 108G, 108H (referred to individually or collectively as 108) connect to each other via a core node 102. An edge node 108 includes a source node that supports traffic sources and a sink node that supports traffic sinks. Traffic sources and traffic sinks (not shown) are usually paired and each source node is usually integrated with a sink node with which it shares memory and control.
  • The core node 102 may be considered in greater detail in view of FIG. 2, which illustrates an electronic core node. The core node 102 includes N input ports 202A, 202B, 202C, . . . , 202N (referred to individually or collectively as 202) for receiving data from the edge nodes 108 of FIG. 1. Each of the N input ports 202 is connected to a corresponding buffer 204A, 204B, 204C, . . . , 204N (referred to individually or collectively as 204) that is connected to a corresponding port controller 206A, 206B, 206C, . . . , 206N (referred to individually or collectively as 206). A space switch 212 directs input received from each of the buffers 204 to an appropriate one of M output ports 208A, 208B, 208C, . . . , 208M (referred to individually or collectively as 208) under control of a slave space switch controller 214. Notably, although the core node 102 and the space switch 212 are described as having a number of inputs, N, that is different from the number, M, of outputs, quite often the number of inputs and outputs is equal, i.e., N=M. A master controller 210 is communicatively coupled to the port controllers 206 and the output ports 208 as well as to the slave space switch controller 214. Each of the control functions of the master controller 210 can be implemented in application-specific hardware, which is the preferred implementation when high speed is a requirement. In an alternative implementation, the master controller 210 may be loaded with burst scheduling and time locking software for executing methods exemplary of this invention from a software medium 224 which could be a disk, a tape, a chip or a random access memory containing a file downloaded from a remote source.
  • As illustrated in detail in FIG. 3, the master controller 210 includes a processor 302. The processor 302 maintains connections to a memory 304, an input interface 306, an output interface 308, a switch interface 312 and a master time counter 314. At the input interface 306, the master controller 210 receives burst transfer requests from the port controllers 206. At the output interface, the master controller 210 may communicate with the output ports 208 to perform conventional operational and maintenance functions. The processor 302 is also connected to a burst-scheduling kernel 310. Based on the burst transfer requests received from the processor 302, the burst-scheduling kernel 310 determines appropriate timing for switching at the space switch 212. According to the determined timing received from the burst-scheduling kernel 310, the processor 302 passes scheduling information to the slave space switch controller 214 via the switch interface 312. The processor 302 also controls the timing of transmission of bursts, from the buffers 204 to the space switch 212, by transmitting scheduling information to the port controllers 206 via the input interface 306.
  • The burst-scheduling kernel 310 may now be described in view of FIG. 4. The burst-scheduling kernel 310 receives burst transfer requests from the processor 302 via a processor interface 402 and a burst parameter receiver 404. The burst parameter receiver 404 may, for instance, be implemented as a time slotted bus. The parameters of these bursts are queued at a burst parameter queue 406 before being accessed by a burst-scheduling unit 408. Included in the burst-scheduling unit 408 may be a time-space map and a space-time map as well as comparators and selectors for generating scheduling information (co-ordination between these maps). The maps are implemented in partitioned random-access memories. After generating scheduling information for a burst, the scheduling information is transferred to the processor 302 via a schedule transmitter 410 and the processor interface 402.
  • In overview, an input port 202A of core node 102 receives a burst from a subtending edge node 108. The burst is stored in the buffer 204A. Parameters indicating the size (e.g., two megabits) and destination (e.g., a particular edge node 108B) of the burst are communicated from the port controller 206A to the master controller 210 as a burst transfer request. The burst-scheduling unit 408 of the master controller 210 executes a burst scheduling algorithm to generate scheduling information and communicates relevant parts of the generated scheduling information to the port controllers 206. The master controller 210 also communicates relevant parts of the generated scheduling information to the slave space switch controller 214. According to the scheduling information received at the port controller 206A, the buffer 204A sends bursts to the space switch 212. At the space switch 212, a connection is established between the buffer 204A and the output port 208B, according to instructions received from the slave space switch controller 214, such that the burst is successfully transferred from an edge node 108 associated with the traffic source to the edge node 108 associated with the traffic sink.
  • At the master controller 210 (see FIG. 3), the burst transfer request is received by the input interface 306 and passed to the processor 302. The processor 302 then sends the burst transfer request to the burst-scheduling kernel 310. At the burst-scheduling kernel 310 in FIG. 4, the burst transfer request is received at the processor interface 402 and the included burst parameters are extracted at the burst parameter receiver 404. The parameters are queued at the burst parameter queue 406 and subsequently stored at the burst-scheduling unit 408 in a data structure 500 (FIG. 5A). The parameters are stored as an entry 506 in a record 504, where the entry 506 is associated with the burst described by the received parameters. Each record 504 has a plurality of entries 506, and each entry 506 is associated with a burst waiting in a buffer 204. As the number of bursts waiting in each buffer 204 may be different, the records 504 may be of varying sizes. As well, the plurality of entries 506 in each record 504 may be a linked list as will be described hereinafter. Furthermore, the data structure 500 is made up of N records 504, where each record 504 corresponds to one of the N input ports 202 (FIG. 2). As illustrated in FIG. 5B, each entry 506 includes a destination field 508 for storing the destination parameter of the burst and a size field 510 for storing the transfer-time (size) parameter of the burst.
  • A generic memory device storing an array that has a time-varying number of data units must have a sufficient capacity to store the expected maximum number of data units. If several arrays, each having a time-varying number of data units, share the generic memory device, then the allocation of the expected maximum number of data units for each array may be considered wasteful. The data structure 500 stores entries 506 containing parameters of burst transfer requests received from each of the input ports 202. The number of entries 506 for any particular input port 202 may vary violently with time, i.e., number of entries 506 for the particular input port 202 may have a high coefficient of variation. However, the total number of entries 506 waiting in the data structure 500 and corresponding to the N input ports 202 would have a much smaller coefficient of variation when N is large, as would be expected in this case. The size of memory required for the data structure 500 can then be significantly reduced if the entries 506 are stored as N interleaved linked lists. Interleaved linked lists are well known in the art and are not described here. Essentially, interleaved linked lists allow dynamic sharing of a memory by X (where X>1) data groupings using X insertion pointers and X removal pointers. Thus, the interleaved linked lists are addressed independently but they share the same memory device.
  • The number, X, of data groupings in the data structure 500 is at least equal to the number of input ports, N, though X may be higher than N if traffic classes are introduced. X may also be higher than N if data from a source node to a sink node uses multiple paths through different core nodes (as will be described hereinafter), since the data of each path must be identified. Thus, the use of an interleaved linked list is preferred to the use of a memory structured to provide a fixed memory partition per traffic stream. A traffic stream is an aggregation of traffic from a particular source edge node 108 to a particular destination edge node 108, often resulting in a succession of bursts.
  • The burst-scheduling unit 408 maintains two other data structures, namely a calendar (i.e., a time-space map) 600 (see FIG. 6) and an M-element array (i.e., a space-time map) 700 (see FIG. 7).
  • The calendar 600 is divided into K time slots 604; indexed from 1 to K. Some of the time slots 604 in the calendar 600 contain identifiers 606 of input ports 202. Those time slots 604 that do not contain input port identifiers 606 contain, instead, null identifiers 608. Each time slot 604 contains either an input port identifier 606 or a null identifier 608. The presence, in a given time slot 604, of a particular input port identifier 606 indicates to the master controller 210 that an input port 202 (an identifier of which is contained in a particular input port identifier 606) is available to transmit data (if it has waiting data) to the space switch 212 from the time corresponding to the given time slot 604 forward. Each of the time slots 604 in the calendar 600 is representative of a short time period, say 100 nanoseconds.
  • Thus, the instant of time at which a given input port 202 is determined to be available is represented by a time slot 604 in the calendar 600. This will typically force a rounding up of the actual availability time to a nearest time slot 604. The duration of a time slot 604 in the calendar 600, therefore, should be small enough to permit an accurate representation of time and should be large enough to reduce the mean number of times a memory holding the calendar 600 has to be accessed before finding an indication of an input port 202. Several time slots 604 in the calendar 600 contain null identifiers 608 (i.e., all the time slots 604 that don not contain an input port identifier 606) and these must be read since the calendar 600 must be read sequentially. The memory holding the calendar 600 must be a random-access memory however, since an address (index) at which an input port identifier 606 is written is arbitrary.
  • Preferably, the number, K, of time slots 604 in the calendar 600, is significantly larger than the number of input ports 202, N (each port of the space switch 212 has an entry in the calendar, even if the port is not active for an extended period of time). In general, K must be greater than N, where N time slots 604 contain input port identifiers 606 and (K-N) time slots 604 contain null identifiers 608. Further, the duration of the calendar 600 must be larger than a maximum burst span. With a specified maximum burst span of 16 milliseconds, for example, an acceptable number (K) of time slots 604 in the calendar 600 is 250,000 with a slot time of 64 nanoseconds.
  • There is a requirement that the calendar 600 be time locked to the master time counter 314 as will be described hereinafter. In one embodiment of the present invention, each time slot 604 in the calendar 600 has a duration equivalent to a single tick of the master time counter 314. In other embodiments, each time slot 604 in the calendar 600 has a duration equivalent to an integer multiple of the duration of a single tick of the master time counter 314. Each port controller 206 has an awareness of time at the master time counter 314, so that scheduling information received at the port controller 206 may be used to send a burst to the space switch 212 at the time indicated by scheduling information. This awareness may be derived from access to a clock bus or through a time locked local counter.
  • In order to speed up the process, the calendar 600 may be implemented in multiple memory devices. For example, a calendar of 262,144 (218) time slots 604, can be implemented in 16 memory devices each having a capacity to store of 16,384 time slots 604. Addressing a time slot 604 in a multiple-memory calendar is known in the art.
  • In the M-element array 700, each element 704 corresponds to one of the output ports 208. Each element 704 in the M-element array 700 holds a state-transition-time indicator 706. The state-transition-time indicator 706 is an index of a time slot 604 in the calendar 600 representative of a point in time at which the respective output port 208 will be available to transmit data. If, for instance, the calendar 600 has sixteen thousand time slots 604 (i.e., K=16,000), each element 704 in the M-element array 700 may be two bytes long (i.e., capable of holding a binary representation of a time slot index as high as 65,536). Where each of the time slots 604 is 100 nanoseconds long, a sixteen thousand slot calendar 600 may accommodate bursts having a length up to 1.6 milliseconds (i.e., 16 megabits at ten gigabits per second) without having to wrap around the current time when writing the availability of the input port 202 to the calendar 600.
  • To examine scheduling in detail, we may first assume that the master controller 210 has already been operating, that is, assume that burst transfer requests have been satisfied and bursts are therefore flowing from the input ports 202 to the output ports 208 of the core node 102.
  • The burst-scheduling unit 408 scans the calendar 600 to detect a future time slot 604 containing an input port identifier 606 (step 802), resulting in a detected time slot 604A. The burst-scheduling unit 408 then communicates with the burst parameter queue 406 to acquire entries 506 (step 804) from the record 504, in the data structure 500 (FIG. 5), that corresponds to the input port 202 identified in the input port identifier 606 in the detected time slot 604A. It is then determined whether there are entries 506 in the record 504 that corresponds to the identified input port 202 (step 805). Each of the entries 506 identifies a destination and, from the destination, the burst-scheduling unit 408 may deduce an output port 208. If there are entries to schedule (i.e., waiting burst requests), the burst-scheduling unit 408 extracts a state-transition-time indicator 706 (step 806) from each element 704, in the M-element array 700 (FIG. 7), that corresponds to an output port 208 deduced from destinations identified by the acquired entries 506. The burst-scheduling unit 408 then determines a “gap” (step 808) by subtracting the index of the detected time slot 604A from the index of the time slot found in each state-transition-time indicator 706. Each gap represents a time difference between a time at which the input port 202 is available and a time at which the respective output port 208, requested in the respective burst transfer request, is available. The burst-scheduling unit 408 does this for each of the acquired entries 506 for the input port 202. Each entry 506 identifies a single burst transfer request. The burst-scheduling unit 408 then selects the burst transfer request corresponding to the minimum gap (step 810). As will be mentioned hereinafter, to simplify circuitry, the step of acquiring entries 506 from the record 504 (step 804) may only require acquisition of a limited number of entries 506.
  • If the gap of the selected burst transfer request is positive, then the input port 202 is available before the output port 208. The time slot index identified in the state-transition-time indicator 706 corresponding to the availability of the output port 208 which was requested for the selected burst transfer request is then designated as a “scheduled time slot.” If the gap of the selected burst transfer request is negative, then the input port 202 is available after the output port 208. The time slot index in which the input port identifier 606 was detected in step 802 (corresponding to the time when the input port 202 is available) is then designated as the scheduled time slot. The burst-scheduling unit 408 then transmits scheduling information (index of the scheduled time slot and identity of the burst transfer request) to the processor 302 (step 812) via the schedule transmitter 410 and the processor interface 402. When determining a minimum gap in step 810, a negative gap is preferred to a positive gap because use of the input port 202 may begin at the time corresponding to the detected time slot 604A, as the negative gap indicates that the requested output port 208 is already available.
  • The burst-scheduling unit 408 then updates the calendar 600 and the M-element array 700 (step 814). FIG. 9 illustrates steps of the update method of step 814. The burst-scheduling unit 408 first sums the index of the scheduled time slot and the transfer-time determined from the size field 510 of the selected burst transfer request (step 902) and writes the input port identifier 606 of the selected burst transfer request in the time slot 604 indexed by the sum (step 904). The writing of the input port identifier 606 effectively identifies, to the burst-scheduling unit 408, the time at which the input port 202 will be available after transferring the burst corresponding to the selected burst transfer request. Notably, only one input port identifier 606 may occupy a single time slot 604. Consequently, if another input port identifier 606 is already present in the time slot 604 indexed by the sum, the burst-scheduling unit 408 will write to the next available time slot 604. After writing the input port identifier 606 to the time slot 604 indexed by the sum, the burst-scheduling unit 408 writes a null identifier 608 in the scheduled time slot (step 906).
  • Subsequently, or concurrently, the burst-scheduling unit 408 writes a state-transition-time indicator 706 to the M-element array 700 (step 908) in the element 704 corresponding to the output port 208 of the selected burst transfer request. The state-transition-time indicator 706 is an index of the time slot 604 indexed by the sum determined in step 902. As will be apparent to a person skilled in the art, pipelining techniques may also be used to reduce processing time.
  • If, as determined in step 805, there are no entries to schedule (i.e., waiting burst requests), the burst-scheduling unit 408 generates an artificial burst (step 816) where the size of the artificial burst is the “size of the selected burst” as far as step 902 is concerned. The result of this generation of an artificial burst is that (in step 814) the input port identifier 606 is written to a deferred time slot 604.
  • The processor 302, having received the scheduling information, transmits to the appropriate port controller 206, via the input interface 306, scheduling information to indicate a time at which to begin sending the burst corresponding to the selected burst transfer request to the space switch 212. The processor 302 also sends scheduling information (input-output configuration instructions) to the slave space switch controller 214 via the switch interface 312.
  • As the above assumes that the master controller 210 has already been operating, it is worth considering initial conditions, for the calendar 600 especially. As all of the input ports 202 are available initially, yet only one input port identifier 606 may occupy each time slot 604, the first N time slots 604 may be occupied by the input port identifiers 606 that identify each of the N input ports 202. Initially, the data structure 500 is clear of burst transfer requests and the state-transition-time indicator 706 present in each element 704 of the M-element array 700 may be an index of the first time slot 604 in the calendar 600.
  • When an input port 202 is determined to be available, i.e., when the input port identifier 606 is read from a detected time slot 604A (step 802), the corresponding record 504 in the data structure 500 is accessed to acquire entries 506. If the corresponding record 504 is found to be empty, the burst-scheduling unit 408 writes a null identifier 608 in the detected time slot 604A and writes the input port identifier 606 at a deferred time slot. The deferred time slot may be separated from the detected time slot 604A by, for example, 128 time slots. At 100 nanoseconds per time slot 604, this would be amount to a delay of about 13 microseconds.
  • If the M-element array 700 (FIG. 7) can only respond to a single read request at a time, the requests to read each state-transition-time indicator 706 from the elements 704 will be processed one after the other. To conserve time then, it may be desirable to maintain multiple identical copies of the M-element array 700. Where multiple copies are maintained, extraction of a state-transition-time indicator 706 from elements 704 in step 806 may be performed simultaneously. It is preferable that the writing of a particular state-transition-time indicator 706 to a given element 704 of each copy of the M-element array 700 (step 908) be performed in a parallel manner.
  • Where maintaining multiple identical copies of the M-element array 700 conserves time, this is done at the cost of memory. Thus, the number of entries 506 acquired in step 804 should be limited to a value, J. If J entries 506 are acquired in step 804, then there is only a requirement for J identical copies of the M-element array 700. It is preferred that J not exceed four.
  • When the space switch 212 has a relatively high number of ports (input and output) the master controller 210, and in particular the burst-scheduling kernel 310, may take advantage of a parallel processing strategy to further conserve processing time. Such a parallel processing strategy may, for instance, involve considering a 64 by 64 space switch (64 input ports, 64 output ports) as comprising an arrangement of four 16 by 16 space switches. However, so that each input may be connected to any output, four arrangements must be considered. An exemplary configuration 1000 for considering these arrangements is illustrated in FIG. 10. The exemplary configuration 1000 includes four input port groups (sub-sets) 1002A, 1002B, 1002C, 1002D (referred to individually or collectively as 1002) and four output port groups (sub-sets) 1004A, 1004B, 1004C, 1004D (referred to individually or collectively as 1004). Each input port group includes 16 input ports and each output port group includes 16 output ports.
  • Four processors may perform scheduling for the 64 by 64 space switch, where each processor schedules on behalf of one input port group 1002. A scheduling session may be divided into as many scheduling time periods as there are processors. For each scheduling time period, a given processor (scheduling on behalf of one input port group 1002) will schedule only those connections destined for a particular output group 1004. The output group changes after every scheduling time period such that, by the end of the scheduling session, all four output port groups 1004 have been considered for connections from the input port group 1002 corresponding to the given processor. The state of the exemplary configuration 1000 at a particular scheduling time period is illustrated in FIG. 10. The intersection of the output port group 1004 with the corresponding input port group 1002 for the particular scheduling time period is identified with a bold border.
  • A parallel processing data structure 1100, which is an alternative to the data structure 500 illustrated in FIG. 5A, is illustrated in FIG. 11. Each of the N records 1104 in the parallel processing data structure 1100 is divided into sub-records, where each sub-record in a given record 1104 corresponds to a single output port group 1004. Parameters of received burst transfer requests are stored as entries 506 in a record 1104 according to the input port 202 and in a sub-record according to the output port group 1004. The sub-records that correspond to the output port groups 1004 are illustrated in FIG. 11 as a number of rows 1102A, 1102B, 1102C, 1102D.
  • When a given processor of the parallel processors in the burst-scheduling unit 408 scans the calendar 600 to detect a future time slot 604 containing an input port identifier 606 (step 802), the input port identifier 606 must be from the input port group 1002 to which the given processor corresponds. The given processor then communicates with the burst parameter queue 406 to acquire entries 506 (step 804) from the parallel processing data structure 1100. The entries 506 are acquired from the record 1104 that corresponds to the input port 202 identified in the input port identifier 606 in the detected time slot 604 and, furthermore, only from the sub-record corresponding to the output port group 1004 under consideration by the given processor in the current scheduling time period. In FIG. 11, the row 1102A of sub-records corresponding to the output port group 1004 under consideration by the given processor associated with a particular input port group 1002A (which includes input ports N-3, N-2, N-1 and N) is identified with a bold border.
  • A hub and spoke data network 1200 is illustrated in FIG. 12, including a bufferless core node 1210X in place of the core node 102. In the data network 1200, a number of traffic sources 104A, 104B, 104C, 104N (referred to individually or collectively as 104) connect, via the edge nodes 108 and the bufferless core node 1210X, to a number of traffic sinks 106A, 106B, 106C, 106M (referred to individually or collectively as 106). In practice, the traffic sources 104 and the traffic sinks 106 are integrated, for instance, as a personal computer. A space switch and space switch controller are maintained at the bufferless core node 1210X.
  • An edge node 108, typical of the edge nodes 108 in FIG. 12, is illustrated in FIG. 13. Traffic is received from the traffic sources 104 or sent to the traffic sinks 106 at traffic interfaces 1302A, 1302B, 1302C (referred to individually or collectively as 1302). The traffic interfaces 1302 connect to buffers 1304A, 1304B, 1304C (referred to individually or collectively as 1304). The buffers 1304 are controlled by buffer controllers 1306A, 1306B, 1306C (referred to individually or collectively as 1306) with regard to the timing of passing traffic to a core interface 1308X that subsequently passes the traffic to the bufferless core node 1210X. The buffer controllers 1306 also connect to the core interface 1308X for sending, to the bufferless core node 1210X, burst transfer requests in a manner similar to the manner in which the port controllers 206 send burst transfer requests to the master controller 210 in FIG. 2. The core interface 1308X maintains a connection to a slave time counter 1314 for time locking with a master time counter in a master controller.
  • At the bufferless core node 1210X, illustrated in detail in FIG. 14, a space switch 1412 connects N input ports 1402A, 1402B, 1402C, . . . , 1402N (referred to individually or collectively as 1402) to M output ports 1408A, 1408B, 1408C, . . . , 1408M (referred to individually or collectively as 1408) under control of a slave space switch controller 1414. Each of the N input ports 1402 is arranged to send burst transfer requests received from the edge nodes 108 to a master controller 1410 and to send burst traffic to the space switch 1412. If, for instance, a particular input port 1402 is arranged to receive a Wavelength Division Multiplexed (WDM) signal having 16 channels, one channel (i.e., one wavelength) may be devoted to the transfer of burst transfer requests from the subtending edge node 108 to the master controller 1410. As in the core node 102 of FIG. 2, the master controller 1410 passes scheduling information to the slave space switch controller 1414.
  • The master controller 1410 may consult the edge nodes 108, via the output ports 1408, to perform conventional operational and maintenance functions. However, to avoid consulting the edge nodes 108, edge-to-edge rate allocations may be introduced and updated as the need arises. The interval between successive updates may vary between 100 milliseconds and several hours, which is significantly larger than a mean burst duration.
  • In overview, a traffic interface 1302A at a source edge node 108A receives a burst from a subtending traffic source 104A. The burst is stored in the buffer 1304A. Parameters indicating the size and destination (e.g., a destination edge node 108E) of the burst are communicated from the buffer controller 1306A, via the core interface 1308X, to the bufferless core node 1210X in a burst transfer request. At the bufferless core node 1210X, the burst transfer request is received at one of the input ports 1402 and sent to the master controller 1410. The master controller 1410 executes a burst scheduling algorithm to generate scheduling information and communicates relevant parts of the generated scheduling information to the edge nodes 108. The master controller 1410 also communicates relevant parts of the generated scheduling information to the slave space switch controller 1414. At the edge node 108A, the buffer 1304A sends the burst to the bufferless core node 1210X, via the core interface 1308X, according to the scheduling information received at the buffer controller 1306A. At the space switch 1412 of the bufferless core node 1210X, a connection is established between the input port 1402A and the output port 1408B such that the burst is successfully transferred from the source edge node 108A to the destination edge node 108E.
  • As will be apparent to a person skilled in the art, the duty of routing of burst transfer requests to the master controller 1410 and bursts to the space switch 1412 may present a problem to the design of the input ports 1402 if the space switch 1412 is optical. One solution to this problem is to relieve the input ports 1402 of this duty. In a version of the data network 1200 of FIG. 12, which is altered to suit an optical space switch and illustrated in FIG. 15, a bufferless core node 1210Z is collocated with an edge node 108J at a location 112. Additionally, a stand-alone master controller 1610Z exists separate from the bufferless core node 1210Z. The collocated edge node 108J maintains a connection to the stand-alone master controller 1610Z for transferring burst transfer requests, received from other edge nodes 108 (via the bufferless core node 1210Z) and the subtending traffic sources 104, to the space switch controller in the bufferless core node 1210Z. In this solution, it is necessary that the edge nodes 108 be aware that burst transfer requests are to be sent to the collocated edge node 108J. This solution avoids dedication of an entire wavelength to signaling, which typically has a low bit rate.
  • In FIG. 16, the collocated edge node 108J is illustrated in detail. Like the typical edge node 108 of FIG. 13, the collocated edge node 108J includes traffic interfaces 1602A, 1602B, 1602C, buffers 1604A, 1604B, 1604C, buffer controllers 1606A, 1606B, 1606C (referred to individually or collectively as 1606) and a core interface 1608X. The core interface 1608X also maintains a connection to a slave time counter 1614Z for time locking with a master time counter in the master controller 1610Z. However, in addition to the typical edge node 108 in FIG. 13, the collocated edge node 108J also includes a controller interface 1612 for sending burst transfer requests to the stand-alone master controller 1610Z. The buffer controllers 1606 communicate burst transfer requests to the controller interface 1612 rather than to the core interface 1608X, as is the case in the typical edge node 108 in FIG. 13. The core interface 1608X also communicates other burst transfer requests to the controller interface 1612, in particular, burst transfer requests received from other edge nodes 108. The stand-alone master controller 1610Z generates scheduling information based on the burst transfer requests and sends the scheduling information to the slave space switch controller in the bufferless core node 1210Z.
  • As illustrated in detail in FIG. 17, the stand-alone master controller 1610Z includes a processor 1702. The processor 1702 maintains connections to a memory 1704, an edge node interface 1706, a core node interface 1712 and a master time counter 1714. At the edge node interface 1706, the master controller 210 receives burst transfer requests from the collocated edge node 108J. The processor 1702 is also connected to a burst-scheduling kernel 1710. Based on the burst transfer requests received from the processor 1702, the burst-scheduling kernel 1710 determines appropriate timing for switching at the space switch at the bufferless core node 1210Z. According to the determined timing received from the burst-scheduling kernel 1710, the processor 1702 passes scheduling information to the bufferless core node 1210Z via the core node interface 1712. The processor 1702 also controls the timing of transmission of bursts, from the edge nodes 108 to the bufferless core node 1210Z, by transmitting scheduling information to the edge nodes 108 via the edge node interface 1706 and the collocated edge node 108J.
  • At the bufferless core node 1210Z, illustrated in detail in FIG. 18, a space switch 1812 connects N input ports 1802A, 1802B, 1802C, . . . , 1802N (referred to individually or collectively as 1802) to M output ports 1808A, 1808B, 1808C, . . . , 1808M (referred to individually or collectively as 1808) under control of a slave space switch controller 1814. Instead of requiring that the N input ports 1802 be arranged to send burst transfer requests from the edge nodes 108 to a master controller and send bursts to the space switch 1812, burst transfer requests pass through the bufferless core node 1210Z and are sent to the collocated edge node 108J. The collocated edge node 108J then forwards the burst transfer requests to the stand-alone master controller 1610Z, where scheduling information is generated. The scheduling information is received from the stand-alone master controller 1610Z by a master controller interface 1816. The slave space switch controller 1814 then receives the scheduling information from the master controller interface 1816.
  • The bufferless core node 1210Z need not be limited to a single space switch 1812. Especially where each input port 1802 and output port 1808 supports multiple channels over respective links to or from respective edge nodes 108, as is the case in WDM, the bufferless core node 1210Z may include an assembly of multiple parallel space switches (not shown). Each of the multiple space switches may require an associated burst-scheduling kernel, such as the burst-scheduling kernel 1710 in FIG. 17, to be located at the master controller 1610Z of the bufferless core node 1210Z. Alternatively, each of the multiple space switches may be associated with a unique burst scheduling unit (see 408 in FIG. 4).
  • The space switches in the assembly of multiple parallel space switches operate totally independently. The traffic to a specific edge node 108 may, however, be carried by any of the channels of a multi-channel link (WDM fiber link) from a source edge node 108 to the bufferless core node 1210. Preferably, a load-balancing algorithm (not described herein) is used to balance the traffic and thus increase throughput and/or decrease scheduling delay.
  • Successive bursts to the same sink edge node 108 may be transferred using different channels (different wavelengths) and, hence, may be switched in different space switches in the bufferless core node 1210. However, the transfer of successive bursts to the same sink edge node 108 using different channels should not be expanded to include the transfer of successive bursts to the same sink edge node 108 using different links where the delay differential between links (possibly several milliseconds) may complicate assembly of the bursts at the sink edge node 108.
  • Note that conventional WDM demultiplexers and WDM multiplexers are required at the input ports 1802 and output ports 1808 of a bufferless core node 1210 employing multiple parallel space switches. They are not illustrated in the figures, however, their use being well known in the art.
  • An advantage of burst switching is a freedom to select a space switch on a per-burst basis, as long as a predetermined time separation (a microsecond or so) is provided between successive bursts of a single data stream. The time separation is required to offset the effect of propagation delay differentials present in different wavelengths of the same WDM signal.
  • Returning to FIG. 12, propagation delay may be considered in view of the data network 1200. If the edge node 108A is one kilometer away from the bufferless core node 1210X, scheduling information may take five microseconds to pass from the bufferless core node 1210X to the edge node 108A in an optical-fiber link. Similarly, a burst sent from the edge node 108A would take five microseconds to travel to the bufferless core node 1210X. A time period lasting five microseconds is represented in the calendar 600 by 500 time slots 604 of 100-nanoseconds each. It may be that, as a consequence of propagation delay, a burst may arrive at the bufferless core node 1210X after the time at which the burst was scheduled to be passing through the space switch 1412. Consequently, given knowledge, at the bufferless core node 1210X, of an estimate of a maximum round trip propagation delay associated with the edge nodes 108, scheduling can be arranged to take the propagation delay into account. For instance, the burst-scheduling kernel 1710 may schedule such that the earliest a burst may be scheduled, relative to a current time in the master time counter 1714, is at least the estimated maximum round trip propagation delay time into the future.
  • Notably, propagation delay differential was not a problem in the core node 102 of FIG. 2, which had input buffers. The collocation of the collocated edge node 108J with the bufferless core node 1210Z in FIG. 15 removes concern of propagation delay differentials for traffic originating at the traffic sources 104A, 104B connected to the collocated edge node 108J. However, for the other edge nodes 108, a time locking scheme is required so that bursts may be correctly scheduled.
  • The propagation delay between the time at which a burst leaves one of the other edge nodes 108 (i.e., the edge nodes 108 that are not collocated with the bufferless core node 1210Z) and the time at which the burst arrives at the bufferless core node 1210Z may be different for each of the other edge nodes 108. To switch these bursts, without contention or a requirement for burst storage at the bufferless core node 1210Z, the other edge nodes 108 must be time locked to the bufferless core node 1210Z. A time locking technique, also called time coordination, is described in the applicant's U.S. patent application Ser. No. 09/286,431, filed on Apr. 6, 1999, and entitled “Self-Configuring Distributed Switch,” the contents of which are incorporated herein by reference. With time locking, the scheduling method in accordance with the present invention guarantees that bursts arrive to available resources at the bufferless core node 1210Z.
  • Given the collocation of the collocated edge node 108J with the bufferless core node 1210Z and the corresponding fact that all burst transfer requests of the bufferless core node 1210Z pass though the collocated edge node 108J, each other edge node 108 may “time lock,” with the collocated edge node 108J.
  • The time locking may be performed using any one of a number of time locking schemes. In one such scheme, each edge node 108 includes at least one local time counter (e.g., the slave time counter 1314 of FIG. 13) of equal width W. In one embodiment of the present invention, a time locking request may be sent from a particular edge node 108E (FIG. 15), while noting the sending time (i.e., the value of the slave time counter at the particular edge node 108E when the time locking request is sent), to the master controller 1610Z. When the time locking request is received at the master controller 1610Z, the arrival time (i.e., the value of the master time counter 1714 at the arrival time) is noted. A time locking response is generated, including an indication of the arrival time, and sent to the particular edge node 108E. A time difference between sending time and arrival time is determined at the particular edge node 108E and used to adjust the slave time counter at the particular edge node 108E. In future, scheduling information is received at the particular edge node 108E from the stand-alone master controller 1610Z, for instance, “start sending burst number 73 at a time counter state 3564.” If the particular edge node 108E starts sending burst number 73 at slave time counter state 3564, the beginning of the burst will arrive at the bufferless core node 1210Z at master time counter state 3564. Preferably, the duration of each time counter cycle is equal and substantially larger than a maximum round-trip propagation delay from any edge node 108 to any core node 1210 in the data network 1200. Furthermore, the maximum round-trip propagation delay should be taken into account when performing scheduling at the stand-alone master controller 1610Z. Preferably, the counters related to the time locking scheme are included in the controller interface 1612 of the collocated edge node 108J of FIG. 16 and in the core interface of the generic edge node 108 of FIG. 13.
  • FIG. 19 illustrates the data network 1200 supplemented with an additional bufferless core node 1210Y. With the additional bufferless core node 1210Y, a flow control process, which operates at a higher level than the switch operations, may assign one of the bufferless core nodes 1210Z, 1210Y (referred to individually or collectively as 1210) to each traffic stream originating at a particular edge node 108, where a traffic stream is an aggregation of traffic with identical source and destination edge nodes 108. When a burst arrives at a given edge node 108, the given edge node 108 may send a burst transfer request to the core node (say bufferless core node 1210Z) assigned to the traffic stream of which the burst is part. Scheduling information is returned to the given edge node 108. The given edge node 108 may then send the burst to the assigned bufferless core node 1210Z according to the timing represented in the scheduling information. The additional bufferless core node 1210Y is illustrated as collated with an edge node 108K at an additional location 114. An additional master controller 1610Y, corresponding to the additional bufferless core node 1210Y, is also present at the additional location 114.
  • An edge node 108 communicates with all core nodes 1210 in the sending and receiving modes. As such, the edge nodes 108 should be adapted to communicate with more than one bufferless core node 1210. This adaptation is shown for the collocated edge node 108J in FIG. 20. Notably different from the collocated edge node 108J as illustrated in FIG. 16 is the addition of a core interface 1608Y corresponding to the bufferless core node 1210Y. The core interface 1608Y corresponding to the bufferless core node 1210Y requires a connection to a slave time counter 1614Y. As will be apparent to a person skilled in the art, there may be many more than two bufferless core nodes 1210 in a data network and many more than eight edge nodes 108.
  • As stated above, there is a requirement that a slave time counter at a given edge node 108 be time locked to the master time counter of the master controller 1610 of the bufferless core node 1210. The scheduling information transmitted by the master controller 1610 to the edge nodes 108 is based on the time indication of the master time counter 1714 as it corresponds to the scheduled time slot in the calendar 600. The time slots 604 in the calendar 600 must, therefore, also be time locked to the master time counter 1714. The selection of the time counter cycle in use at the master time counter 1714 and the calendar cycle are important design choices. Where a master time counter 1714 counts using W bits, the duration of the master time counter cycle is 2W multiplied by the duration of a period of a clock used to drive the master time counter. With W=32, and a clock period of 16 nanoseconds, for example, the number of counter states is about 4.29×109 and duration of the master time counter is more than 68 seconds. This is orders of magnitude higher than the round-trip propagation delay between any two points on Earth (assuming optical transmission).
  • Increasing the duration of the master time counter 1714 involves adding a few bits, resulting in a very small increase in hardware cost and transport of time locking signals across the network. By contrast, increasing the duration of the calendar 600 requires increasing the depth of a memory used to maintain the calendar 600 and/or increasing the duration of each time slot 604 in the calendar. The latter results in decreasing the accuracy of time representation, and hence in wasted time, as will be explained below.
  • If, for example, each time slot 604 has a duration of eight microseconds and the number of calendar time slots 604 is 65,536, the duration of the calendar 600 is more than 500 milliseconds. A time slot 604 of eight microseconds is, however, comparable with the duration of a typical burst. At 10 Gb/s, an eight microsecond bursts is about ten kilobyte long. It is desirable that the duration of each time slot 604 be a small fraction of the mean burst duration. A reasonable duration for a time slot 604 is 64 nanoseconds. However, if the duration of the calendar 600 is to be maintained at 500 milliseconds, the calendar 600 requires eight million slots. A compromise is to select a duration of the calendar 600 that is just sufficient to handle the largest possible burst and use an associated adder or cycle counter to be cognizant of the calendar time relationship to the master time counter time. The largest burst duration would be imposed by a standardization process. In a channel of 10 Gb/s, a burst of one megabyte has a duration of less than one millisecond. A standardized upper-bound of the burst length is likely to be even less than one megabyte in order to avoid delay jitter. Thus, the duration of the calendar 600 can be selected to be less than 16 milliseconds. With a duration of each time slot 604 set to 64 nanoseconds, the number of required time slots 604 would be about 262,144. This can be placed in four memory devices of 65,536 words each, a word corresponding to a time slot 604.
  • Relating a time slot 604 in the calendar 600 to the state of the master time counter 1714 is greatly simplified if the ratio of the number of master time counter states to the number of time slots 604 is a power of two, and the ratio of the duration of a time slot 604 to the duration of the clock used to drive the master time counter is also a power of two. Notably, the number of master time counter states exceeds or equals the number of calendar slots and the duration of a calendar slots exceeds or equals the clock period.
  • If the width of the master time counter is 32 bits, the width of a calendar address is 18 bits (218, i.e., 262,144 time slots 604), and the duration of a time slot 604 is four times the period of the clock used to drive the master time counter, then duration of the master time counter is 4,096 times the duration of the calendar 600. Reducing the width of the master time counter to 24 bits, with 262,144 calendar slots, a clock period of 16 nanoseconds and a duration of each time slot 604 of 64 nanoseconds, the duration of the master time counter 1714 becomes about 268.72 milliseconds, which is 16 times the calendar period of about 16.77 milliseconds. The master clock period is selected to be reasonably short to ensure accurate time representation for time locking purposes.
  • FIG. 21 depicts a master time counter cycle 2102 and a calendar cycle 2104 for an exemplary case wherein a duration 2112 of the master time counter cycle 2102 is exactly four times a duration 2114 of the calendar cycle 2104. Time locking of the calendar to the master time counter is essential as indicated in FIG. 21.
  • The scheduling of future burst transfers based on burst transfer requests received from a specific edge node 108, associated with a specific input port of a bufferless core node 1210, is illustrated in FIG. 22. Changes in the state of a calendar 2200 are illustrated as they correspond to the specific input port. In particular, the calendar 2200 has 32 time slots and is shown as four calendar cycles 2200S, 2200T, 2200U, 2200V. In this example, the duration of the master time counter is four times the duration of the calendar 2200. A time slot 2202A contains an identifier of the input port, typically an input port number. Each other time slot in the calendar 2200 contains either an input port identifier or a null identifier, although, for simplicity, these identifiers are not shown. As the calendar 2200 is scanned, the time slot 2202A is encountered and an input port identifier 2206 is recognized. The burst scheduling method of FIG. 8 is then performed, along with the map maintenance method of FIG. 9. These methods result in the input port identifier 2206 being replaced with a null identifier in time slot 2202A and the input port identifier 2206 being written in time slot 2202B. The methods of FIGS. 8 and 9 are repeated when the input port identifier 2206 is encountered in time slot 2202B, resulting in a null identifier in time slot 2202B and the input port identifier 2206 being written in time slot 2202C. When the input port identifier 2206 is encountered in time slot 2202C, the input port identifier 2206 being written in time slot 2202D which is in the second calendar cycle 2200T and has a numerically smaller index in the calendar 2200. The index of time slot 2202D is smaller than the index of time slot 2202C because the adder determining the index of the time slot in which to write the input port identifier 2206 (step 902) has a word length that exactly corresponds to the number of time slots in the calendar 2200 (note that calendar length is a power of 2). When the input port identifier 2206 is encountered in time slot 22021 in the fourth calendar cycle 2200V, the input port identifier 2206 is written to time slot 2202X in the first calendar cycle 2200S. Scheduling availability of the input port in the first calendar cycle 2200S means that the input port will not be available until the master clock cycle subsequent to the master clock cycle in which time slot 22021 was encountered.
  • It is emphasized that the scheduling procedure described above enables scheduling bursts for a look-ahead period as large as the duration of the master time counter. Where the duration of the master time counter is 268 milliseconds (224 master time counter slots, 16 nanosecond clock period), for example, at 10 GHz, bursts of cumulative length as high as 200 megabytes can be scheduled.
  • To compute a scheduled time indication, i.e., the master time counter state corresponding to the scheduled time slot, to be reported to a respective edge node, an indication of the relative calendar cycle number with respect to the master time counter cycle must be provided along with the scheduled time slot. In the example of FIG. 22, this indication is 0, 1, 2 or 3. The scheduled time indication is then the cycle indication, left shifted by 5 bits (log2 32) added to the scheduled time slot from the calendar. For example, if time slot 2202G, which is at time index 20 in the third calendar cycle (relative calendar indication 2), is the scheduled time slot, the scheduled time indication is 2×32+20=84. The scheduled time indication that is communicated to the requesting edge node is 84.
  • A portion of the network capacity in the data network 1200 may be dedicated to relatively well-behaved traffic. That is, non-bursty traffic. To this end, a master controller may include a second scheduler dedicated to more traditional circuit switching. Like the master controller 1610Z illustrated in FIG. 17, the master controller 1610Y illustrated in FIG. 23 includes a processor 2302. The processor 2302 maintains connections to a memory 2304, an edge node interface 2306, a core node interface 2312 and a master time counter 2314. The master controller 1610Y illustrated in FIG. 23 also includes a circuit-scheduling kernel 2316 for scheduling transfers between edge nodes 108 on a longer term basis.
  • In one embodiment of the present invention, the edge nodes 108 (or the port controllers 206) may perform some processing of bursts. This processing may include expansion of bursts to have a length that is a discrete number of segments or aggregation of small bursts.
  • Notably, the present invention is applicable without dependence on whether switching in the data network 1200 is electrical or optical and without dependence on whether transmission in the data network 1200 is wireline or wireless. The optical switching example is particularly instructive, however, in that, given recent developments in Dense Wavelength Division Multiplexing, a link between an edge node 108 and a bufferless core node 1210 may include multiple (e.g., 32) channels. If the data network 1200 is to work as described in conjunction with FIG. 12, one of the multiple channels may be completely dedicated to the transfer of burst transfer requests. However, the transfer of burst transfer requests represent a very small percentage of the available capacity of such a channel and the unused capacity of the dedicated channel is wasted. This is why co-location of an edge node 108 with a bufferless core node 1210 is used. The optical example is also well suited to the consideration herein that the core node 1210X (FIG. 14) is bufferless, as an efficient method for buffering optically received data has not yet been devised.
  • Advantageously, the present invention allows bursts that switch through the core nodes to employ the core nodes, and associated space switches, nearly constantly, that is, with virtually no data loss. As such, the network resources are used more efficiently.
  • Other modifications will be apparent to those skilled in the art and, therefore, the invention is defined in the claims.

Claims (20)

1. An edge node in a network, said edge node comprising:
an edge controller including a slave time counter;
means for time-locking said slave time counter to a master time counter associated with a core node in said network;
means for forming data bursts;
means for associating, with each of said data bursts, indicators of destination and burst duration;
means for communicating said burst-indicators to said core node;
means for receiving burst-transfer schedules from said core node; and
means for transmitting data bursts according to said schedules.
2. The edge node of claim 1 wherein said communicating and said transmitting are time interleaved.
3. The edge node of claim 1 wherein said communicating and said transmitting are concurrent.
4. The edge node of claim 1 wherein each of said data bursts has a duration not exceeding a predetermined limit.
5. The edge node of claim 4 wherein said schedules are determined with reference to a calendar of fixed calendar duration.
6. The edge node of claim 5 wherein said predetermined limit does not exceed said calendar duration.
7. A core node in a burst-switching network, said core node comprising:
at least one space switch each having a plurality of input ports, a plurality of output ports, and a burst scheduler for scheduling transfer of bursts of arbitrary sizes from said plurality of input ports to said plurality of output ports; and
a master controller operable to:
exchange time-locking signals with each of a plurality of edge nodes;
receive, from at least one edge node of said plurality of edge nodes, a stream of burst-transfer requests for bursts of arbitrary sizes; and
allocate each of said burst-transfer requests to one of said at least one space switch.
8. The core node of claim 7 wherein said master controller is further operable to communicate an indication of a burst-transfer time corresponding to said each of said burst-transfer requests to an edge node from which said each of said burst-transfer requests originated.
9. The core node of claim 8 wherein said burst scheduler comprises:
a first memory device for storing a calendar divided into a number of divisions where the presence in a division of an identifier of a particular input port indicates that said input port is available to transmit a new burst;
a second memory device for storing the availability time of each of said output ports;
a third memory device for storing burst-transfer requests for transfer of bursts from said input ports, each of said burst-transfer requests specifying an input port, an output port, and duration of a corresponding burst; and
a burst-scheduling kernel operable to:
select at least two burst-transfer requests;
determine a time gap between an availability time of a particular input port and an availability time of a candidate output port corresponding to each of said at least two burst-transfer requests; and
schedule a preferred burst-transfer request having the least time gap.
10. A scheduler for scheduling transfer of bursts from a plurality of N input ports to a plurality of M output ports of a space switch, the scheduler comprising:
first means for determining a first time index at which each input port becomes unoccupied;
second means for determining a second time index at which each output port becomes unoccupied;
third means for receiving burst-transfer requests, each of said burst-transfer requests indicating an input port, an output port, and a burst duration;
fourth means for selecting a particular input port having the least first time index; and
fifth means for updating said least first time index corresponding to said particular input port.
11. The scheduler of claim 10 wherein said first means includes a first memory device for storing a calendar having a predefined calendar period and divided into a number K of divisions, each division corresponding to a time slot in said calendar period and containing an indication of a state transition in said plurality of input ports so that the presence in a division of an identifier of a particular input port indicates that said particular input port is available to transmit a new burst and the presence of a null value indicates that non of said input ports changes occupancy state during said each division.
12. The scheduler of claim 11 wherein said indication is an identifier of a specific input port belonging to said plurality of input ports, where said specific input port is scheduled to become unoccupied and available for transmitting a burst during the time slot corresponding to said indication.
13. The scheduler of claim 11 wherein said number of divisions K at least equals the number N of input ports.
14. The scheduler of claim 11 wherein said number K is substantially larger than the number N of input ports.
15. The scheduler of claim 11 wherein said second means includes a second memory device for storing an availability time of each of said M output ports.
16. The scheduler of claim 10 wherein said third means includes a third memory device for storing burst-transfer requests for transfer of bursts from said plurality of input ports, each of said burst-transfer requests specifying an input port, an output port, and a burst duration of a corresponding burst.
17. The scheduler of claim 15 wherein said fourth means includes means for sequentially reading the contents of successive divisions of said calendar until a particular index of a calendar division containing an indication of an identifier of a next-encountered input port is found, said next-encountered input port becoming a next-available input port.
18. The scheduler of claim 17 wherein, when there are no burst-transfer requests corresponding to said next-available input port, said fifth means:
selects an artificial burst having a duration of a pre-selected number of time slots;
adds said pre-selected number of time slots to said particular index to determine a new time index; and
writes an identifier of said next-available input port in a calendar division containing a null value and corresponding to the nearest time slot following said new time index.
19. The scheduler of claim 17 wherein, when there is one burst-transfer request corresponding to said next-available input port, said fifth means:
adds a burst-duration indicated in said one burst transfer request to said particular index to determine a new time index;
writes an identifier of said next-available input port in a calendar division containing a null value and corresponding to the nearest time slot following said new time index; and
writes said new time index in an entry in said second memory corresponding to an output port indicated in said one burst-transfer request.
20. The scheduler of claim 17 wherein, when there are at least two burst-transfer requests corresponding to said next-available input port, said fifth means:
identifies candidate output ports each corresponding to one of said at least two burst-transfer requests;
reads an availability time of each of said candidate output ports from said second memory device;
determines a time gap between the availability time of said next-available input port and said availability time of each of said candidate output ports;
selects a preferred candidate output port having the least time gap;
adds a burst-duration indicated in the burst transfer request corresponding to said preferred candidate output port to determine a new time index;
writes an identifier of said next-available input port in a calendar division containing a null value and corresponding to the nearest time slot following said new time index; and
writes said new time index in an entry in said second memory corresponding to said preferred output port.
US11/124,656 2000-12-29 2005-05-09 Burst switching in a high capacity network Abandoned US20050207339A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/124,656 US20050207339A1 (en) 2000-12-29 2005-05-09 Burst switching in a high capacity network

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/750,071 US6907002B2 (en) 2000-12-29 2000-12-29 Burst switching in a high capacity network
US11/124,656 US20050207339A1 (en) 2000-12-29 2005-05-09 Burst switching in a high capacity network

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09/750,071 Continuation US6907002B2 (en) 2000-12-29 2000-12-29 Burst switching in a high capacity network

Publications (1)

Publication Number Publication Date
US20050207339A1 true US20050207339A1 (en) 2005-09-22

Family

ID=25016364

Family Applications (2)

Application Number Title Priority Date Filing Date
US09/750,071 Expired - Fee Related US6907002B2 (en) 2000-12-29 2000-12-29 Burst switching in a high capacity network
US11/124,656 Abandoned US20050207339A1 (en) 2000-12-29 2005-05-09 Burst switching in a high capacity network

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/750,071 Expired - Fee Related US6907002B2 (en) 2000-12-29 2000-12-29 Burst switching in a high capacity network

Country Status (2)

Country Link
US (2) US6907002B2 (en)
EP (1) EP1220567A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070217405A1 (en) * 2006-03-16 2007-09-20 Nortel Networks Limited Scalable balanced switches
US20080013543A1 (en) * 2001-09-27 2008-01-17 International Business Machines Corporation Apparatus and method to coordinate calendar searches in a network scheduler
US20080031256A1 (en) * 2003-07-10 2008-02-07 International Business Machines Corporation Apparatus and method to coordinate calendar searches in a network scheduler given limited resources
US20080165688A1 (en) * 2003-05-14 2008-07-10 Beshai Maged E Regulating Data-Burst Transfer
US20090019183A1 (en) * 2007-07-10 2009-01-15 Qualcomm Incorporated Methods and apparatus for data exchange in peer to peer communications
US20160241381A1 (en) * 2011-07-20 2016-08-18 Aviat U.S., Inc. Systems and methods of clock synchronization between devices on a network
US10349459B2 (en) * 2014-03-07 2019-07-09 Huawei Technologies Co., Ltd. Relay node RN, donor eNodeB DeNB and communication method

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7403578B2 (en) * 2001-06-08 2008-07-22 Broadcom Corporation Robust burst detection and acquisition system and method
US7233590B2 (en) * 2001-07-06 2007-06-19 Nortel Networks Limited Switched channel-band network
US7190900B1 (en) * 2001-07-20 2007-03-13 Lighthouse Capital Partners Iv, Lp System and method for implementing dynamic scheduling of data in a non-blocking all-optical switching network
CA2410143C (en) * 2001-11-02 2010-02-02 Nippon Telegraph And Telephone Corporation Optical dynamic burst switch
CA2410064C (en) * 2001-11-02 2007-12-04 Nippon Telegraph And Telephone Corporation Optical dynamic burst switch
US7187654B1 (en) 2001-11-13 2007-03-06 Nortel Networks Limited Rate-controlled optical burst switching
US7215666B1 (en) * 2001-11-13 2007-05-08 Nortel Networks Limited Data burst scheduling
US7085849B1 (en) * 2002-03-08 2006-08-01 Juniper Networks, Inc. Scheduler systems and methods for transmit system interfaces
US7117257B2 (en) * 2002-03-28 2006-10-03 Nortel Networks Ltd Multi-phase adaptive network configuration
US20040037558A1 (en) * 2002-08-20 2004-02-26 Nortel Networks Limited Modular high-capacity switch
US7535841B1 (en) * 2003-05-14 2009-05-19 Nortel Networks Limited Flow-rate-regulated burst switches
US7127547B2 (en) * 2003-09-30 2006-10-24 Agere Systems Inc. Processor with multiple linked list storage feature
US7397792B1 (en) * 2003-10-09 2008-07-08 Nortel Networks Limited Virtual burst-switching networks
US8064341B2 (en) * 2003-10-10 2011-11-22 Nortel Networks Limited Temporal-spatial burst switching
US7539181B2 (en) * 2004-12-13 2009-05-26 Nortel Networks Limited Balanced bufferless switch
JP3998691B2 (en) * 2005-05-26 2007-10-31 沖電気工業株式会社 Data transfer network
US8804751B1 (en) 2005-10-04 2014-08-12 Force10 Networks, Inc. FIFO buffer with multiple stream packet segmentation
KR100921458B1 (en) * 2005-10-31 2009-10-13 엘지전자 주식회사 Method of transmitting and receiving control information in wireless mobile communications system
US9049205B2 (en) * 2005-12-22 2015-06-02 Genesys Telecommunications Laboratories, Inc. System and methods for locating and acquisitioning a service connection via request broadcasting over a data packet network
US8675743B2 (en) 2007-08-03 2014-03-18 Apple Inc. Feedback scheduling to reduce feedback rates in MIMO systems
JP5341503B2 (en) 2008-12-26 2013-11-13 株式会社東芝 Memory device, host device, and sampling clock adjustment method
US8295698B2 (en) * 2009-08-27 2012-10-23 Maged E Beshai Time-coherent global network
EP2337372B1 (en) * 2009-12-18 2012-02-08 Alcatel Lucent High capacity switching system
US8582437B2 (en) * 2011-06-21 2013-11-12 Broadcom Corporation System and method for increasing input/output speeds in a network switch
US9348775B2 (en) * 2012-03-16 2016-05-24 Analog Devices, Inc. Out-of-order execution of bus transactions
US10834056B2 (en) * 2018-07-31 2020-11-10 Ca, Inc. Dynamically controlling firewall ports based on server transactions to reduce risks
US11356240B2 (en) 2020-05-05 2022-06-07 Maged E. Beshai Time alignment of access nodes to optical distributors of a global network
TWI819635B (en) * 2022-06-01 2023-10-21 瑞昱半導體股份有限公司 Memory control system and memory control method

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5377327A (en) * 1988-04-22 1994-12-27 Digital Equipment Corporation Congestion avoidance scheme for computer networks
US5231631A (en) * 1989-08-15 1993-07-27 At&T Bell Laboratories Arrangement for regulating traffic in a high speed data network
US5107489A (en) * 1989-10-30 1992-04-21 Brown Paul J Switch and its protocol for making dynamic connections
US5241536A (en) 1991-10-03 1993-08-31 Northern Telecom Limited Broadband input buffered atm switch
JPH05207023A (en) * 1992-01-24 1993-08-13 Hitachi Ltd Mass data transmitting method
CA2095755C (en) * 1992-08-17 1999-01-26 Mark J. Baugher Network priority management
CA2112756C (en) * 1993-01-06 1999-12-14 Chinatsu Ikeda Burst band-width reservation method in asynchronous transfer mode (atm) network
JPH09121217A (en) * 1995-08-23 1997-05-06 Fujitsu Ltd Method for burst transfer
US5745486A (en) 1995-10-26 1998-04-28 Northern Telecom Limited High capacity ATM switch
JP3742481B2 (en) * 1996-11-18 2006-02-01 富士通株式会社 Fixed-length cell handling type exchange and fixed-length cell readout speed control method
US5953318A (en) * 1996-12-04 1999-09-14 Alcatel Usa Sourcing, L.P. Distributed telecommunications switching system and method
JPH1117685A (en) * 1997-06-20 1999-01-22 Oki Electric Ind Co Ltd Band management circuit, transmitter and transmission system
US6405257B1 (en) * 1998-06-26 2002-06-11 Verizon Laboratories Inc. Method and system for burst congestion control in an internet protocol network
US6317415B1 (en) * 1998-09-28 2001-11-13 Raytheon Company Method and system for communicating information in a network
US6560196B1 (en) * 1998-11-19 2003-05-06 Cisco Technology, Inc. Method and apparatus for controlling the transmission of cells across a network

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7733873B2 (en) * 2001-09-27 2010-06-08 International Business Machines Corporation Coordination of calendar searches in a network scheduler
US20080013543A1 (en) * 2001-09-27 2008-01-17 International Business Machines Corporation Apparatus and method to coordinate calendar searches in a network scheduler
US7817543B2 (en) * 2003-05-14 2010-10-19 Nortel Networks Limited Regulating data-burst transfer
US20080165688A1 (en) * 2003-05-14 2008-07-10 Beshai Maged E Regulating Data-Burst Transfer
US20080031256A1 (en) * 2003-07-10 2008-02-07 International Business Machines Corporation Apparatus and method to coordinate calendar searches in a network scheduler given limited resources
US8139594B2 (en) * 2003-07-10 2012-03-20 International Business Machines Corporation Apparatus and method to coordinate calendar searches in a network scheduler given limited resources
US20070217405A1 (en) * 2006-03-16 2007-09-20 Nortel Networks Limited Scalable balanced switches
US8687628B2 (en) 2006-03-16 2014-04-01 Rockstar Consortium USLP Scalable balanced switches
US20090019183A1 (en) * 2007-07-10 2009-01-15 Qualcomm Incorporated Methods and apparatus for data exchange in peer to peer communications
US9037750B2 (en) * 2007-07-10 2015-05-19 Qualcomm Incorporated Methods and apparatus for data exchange in peer to peer communications
US20160241381A1 (en) * 2011-07-20 2016-08-18 Aviat U.S., Inc. Systems and methods of clock synchronization between devices on a network
US9912465B2 (en) * 2011-07-20 2018-03-06 Aviat U.S., Inc. Systems and methods of clock synchronization between devices on a network
US10594470B2 (en) 2011-07-20 2020-03-17 Aviat U.S., Inc. Systems and methods of clock synchronization between devices on a network
US10608807B2 (en) 2011-07-20 2020-03-31 Aviat U.S., Inc. Systems and methods of clock synchronization between devices on a network
US10349459B2 (en) * 2014-03-07 2019-07-09 Huawei Technologies Co., Ltd. Relay node RN, donor eNodeB DeNB and communication method

Also Published As

Publication number Publication date
EP1220567A1 (en) 2002-07-03
US20020085491A1 (en) 2002-07-04
US6907002B2 (en) 2005-06-14

Similar Documents

Publication Publication Date Title
US6907002B2 (en) Burst switching in a high capacity network
US7606262B1 (en) Universal edge node
US7590109B2 (en) Data burst scheduling
US7817543B2 (en) Regulating data-burst transfer
US8902916B2 (en) Rate controlled opitcal burst switching
US7535841B1 (en) Flow-rate-regulated burst switches
US20090080885A1 (en) Scheduling method and system for optical burst switched networks
US20070110052A1 (en) System and method for the static routing of data packet streams in an interconnect network
US9319310B2 (en) Distributed switchless interconnect
US20020118421A1 (en) Channel scheduling in optical routers
US20100067536A1 (en) Multimodal Data Switch
JP2016501475A (en) Router for passive interconnection and distributed switchless switching
US20120311175A1 (en) Guaranteed bandwidth memory apparatus and method
EP1220497B1 (en) Packet switch
US7212551B1 (en) Time-coordination in a burst-switching network
JP2015536621A (en) Passive connectivity optical module
US7345995B2 (en) Conflict resolution in data stream distribution
KR100903130B1 (en) Switch of mesh type on-chip network and swithing method using thereof
KR100667155B1 (en) Distributed Channel Access Control Method in WDM Ring Network
US20090073968A1 (en) Device with modified round robin arbitration scheme and method for transferring data
JP2000069048A (en) Atm switch
Jia et al. MultiS-Net: A high-capacity, packet-switched, multichannel, single-hop architecture and protocol for a local lightwave network
JPH06205026A (en) Node device and communication network

Legal Events

Date Code Title Description
AS Assignment

Owner name: AFRL/RIJ, NEW YORK

Free format text: EXECUTIVE ORDER 9424, CONFIRMATORY LICENSE;ASSIGNOR:NORTEL NETWORKS LIMITED;REEL/FRAME:020710/0172

Effective date: 20080306

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION