WO2002031620A2 - Improved network restoration capability via dedicated hardware and continuous performance monitoring - Google Patents
Improved network restoration capability via dedicated hardware and continuous performance monitoring Download PDFInfo
- Publication number
- WO2002031620A2 WO2002031620A2 PCT/US2001/030000 US0130000W WO0231620A2 WO 2002031620 A2 WO2002031620 A2 WO 2002031620A2 US 0130000 W US0130000 W US 0130000W WO 0231620 A2 WO0231620 A2 WO 0231620A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- optical
- signal
- parameters
- restoration
- speed bus
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L43/00—Arrangements for monitoring or testing data switching networks
- H04L43/08—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
- H04L43/0805—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
- H04L43/0811—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking connectivity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/06—Management of faults, events, alarms or notifications
- H04L41/0654—Management of faults, events, alarms or notifications using network fault recovery
- H04L41/0663—Performing the actions predefined by failover planning, e.g. switching to standby network elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/40—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using virtualisation of network functions or resources, e.g. SDN or NFV entities
Definitions
- This invention relates to optical data networks, and more particularly relates to fast restoration of data flow in the event of disruption.
- Restoration time tends to be dependent upon how fast an optical switch fabric can be reconfigured and how quickly the optical signal characteristics at the input and output transmission ports can be measured. This reconfiguration process may require several iterations before an optimal signal quality is even achieved, especially if the optical switch fabric is based upon 3D MEMS mirror technology. Thus, a modern high-speed optical data network really cannot function without an exceedingly fast mechanism for signal monitoring and switch reconfiguration to ensure absolute minimum restoration times.
- Figure 1 depicts a typical control architecture for a network node in a modern optical network.
- the depicted system is essentially the current state of the art in modern optical networks.
- the typical steps in identifying a trouble condition and implementing restoration of the signal will be described, along with the temporal costs of each step in the process. It is noted that these temporal costs are high, and significant data will be lost under such a system.
- An incoming optical signal 101 enters an input port in an I/O module
- both copies of the incoming signal now collectively an output signal 101AA and 101 BB, are routed to an output module 104 in which one copy of the signal (101AA or 101 BB) is selected and routed to an output I/O port as outgoing signal 160.
- Signal monitoring can be performed on the incoming optical signal 101 as well as on the outgoing signal 160.
- Such signal monitoring is generally implemented in hardware and thus has minimal execution time, generally on the order of less than 10 milliseconds, and thus adds little temporal cost to the control methodology.
- a trouble condition is detected at the input monitoring point 150, such as a loss of signal condition in the incoming optical signal 101 , or if a trouble condition is detected at the output monitoring point 151 , such as signal degradation in the output signal 160, then an interrupt must be sent to the system controller 110 via the I/O controller 120.
- the system controller 110 reads the I/O pack via I/O controller 120 to examine the state of the key port parameters. This operation is mostly software intensive, with interrupt latency and message buffer transfer times on the order of 500 milliseconds.
- the system controller 110 analyzes the I/O pack data and informs the restoration controller 130 to initiate restoration. These operations are handled in software and generally, in the described state of the art system, require on the order of 10 milliseconds to accomplish.
- the restoration controller 130 computes a new path and port configuration, informs the system controller 110, which then informs the switch controller 135 to reconfigure the switch fabric 103 for the new I/O port connectivity.
- the restoration controller 130 then notifies its nodal neighbors (not shown, as Fig. 1 is depicts a network element in isolation) of the new configuration. This latter step entails software operations and takes on the order of 500 milliseconds to accomplish.
- the total restoration time in the modern state of the art optical data network is comprised of internal processing time at the network element of approximately one second (actually 1.020 seconds or slightly less) plus the tn 2 n, or the external node to node messaging time.
- prior art systems operate by monitoring the incoming optical signal upon entry and prior to being outputted, and if a trouble condition is detected, then and only then is an interrupt sent to a system controller via an I/O controller.
- the system controller receives the interrupt message, reads the I/O pack, and informs a restoration controller to initiate restoration.
- the restoration controller computes new path and port configurations and sends a message to the system controller to reconfigure the switch fabric.
- the restoration controller (“RC") notifies all nodal neighbors of the new configuration. This is thus an alarm-based system where nothing happens unless a trouble condition is detected; then, by a series of interrupts and messages, each with their inherent delays, latencies and processing times, action is taken.
- the present invention provides a novel solution to fast network restoration.
- dedicated hardware elements are utilized to implement restoration, and these elements are linked via a specialized highspeed bus.
- the incoming optical signals to each input port are continually monitored and their status communicated to such dedicated hardware via such high-speed bus. This provides a complete snapshot, in virtually real time, of the state of each input port on the node.
- the specialized hardware automatically detects trouble conditions and reconfigures the switching fabric.
- the hardware comprises a Connection Manager and an Equipment Manager.
- the switching fabric control is also linked via the same high- speed bus, making the changes to input/output port assignments possible in less than a millisecond and thus reducing the overall restoration time.
- the status information is continually updated every 125 microseconds or less, and the switch fabric can be reconfigured in no more than 250 microseconds.
- Figure 1 depicts a typical optical network node control structure
- Figure 2 depicts the optical network node control and restoration structure according to the present invention
- Figure 3 depicts the contents of a status information frame according to the method of the present invention.
- Figure 4 depicts a more detailed view of the structure of Figure 2 in a particular embodiment of the present invention.
- the concept of the present invention is a simple one. As opposed to prior art systems wherein restoration is triggered only upon the detection of a fiber cut or other catastrophic event, and the propagation of the resultant alarm signal through the control architecture and switching fabric, the method and system of the present invention significantly reduce the time it takes for the system to recognize traffic disruption and restore an alternative traffic path by utilizing dedicated hardware and high speed control data links.
- the present invention continually updates the optical signal quality status from all of the optical interface boards bringing incoming optical signals into a network node.
- the high speed control data lines interconnect the optical I/O modules to the system modules concerned with reconfiguring data paths and controlling the switch fabric, thus obviating the temporal costs of propagating an alarm interrupt and the associated intermodule sequential signaling.
- FIG. 2 is a system level drawing of a network node's optical performance monitoring and restoration system.
- the high speed bus 201 which connects the group managers ("GM”s) 202 to the connection manager (“CM") 203, the equipment manager (“EM”) 204 and the switch manager (“SWM”) 205.
- the group managers 202 on the left side of the drawing are each responsible for controlling a number of Input Optical Modules ("IOM"s) 206.
- IOM Input Optical Modules
- each group manager will control 16 input optical modules 206 each having four input lines; with 8 logical group managers 202 in total.
- the term logical in this context, is used to designate the number of GMs actually active at any one time. Various embodiments may use load sharing or some type of protection so as to actually have two or more physical GMs for each logically active GM in the system. Thus, in a preferred embodiment, to support 8 active GMs there will be 16 physical GMs, the excess 8 utilized for load sharing of half the capacity of the logical 8 GMs, or for protection, or for some combination thereof.
- the 8 active GMs each controlling 16 lOMs 206, with each IOM having four input lines, gives a total of 8x16x4 or 512 input lines at the network nodal level.
- Group managers 202 also control output optical modules 207.
- each output optical module 207 having the same number of output lines, namely 4 in this exemplary embodiment, as an input optical module 206 has input lines.
- Any number of group managers 202 could be used, however, as well as any number of optical modules assigned to each GM, and any number of input/output lines per optical input/output module, as design, efficiency, and cost considerations may determine.
- the I/O lines will be bi-directional, and the logical lOMs and OOMs bi-directional as well and thus physically identical.
- An incoming optical signal 200 to the network node terminates on an input optical module or IOM 206.
- the incoming signal is split into two identical copies 200A and 200B and sent to parallel switch fabrics 210.
- both copies of the original input signal 200AA and 200BB, now a pair of output signals are routed to an output optical module or OOM 207, in which a copy of the signal (200AA or 200BB) is selected and routed to an output I/O port as the outgoing signal 221.
- Signal monitoring is performed on an incoming signal at point 220, prior to its entry into the optical module, and on an outgoing signal at point 221 , after its exiting form an optical module. This process is primarily a hardware function, and requires less than 10 milliseconds to accomplish.
- the input side i.e. that measured at point 220
- the output side i.e. that measured at point 221
- transmit the input side
- Devices to monitor the incoming signal are generally well known in the art, and specific examples of specialized and efficient signal monitoring devices are described, for example, in U.S. Patent Application 09/852,582, under common assignment herewith.
- the optical performance monitoring devices measure the various signal parameters, such as optical power (“OP”) and optical signal to noise ratio ("OSNR”) for each input and each output port (these may be physically at the same location in bi-directional port structures), and send this information, via the high-speed bus, to the CM 203, EM 204, and SWM 205.
- Information for the entire set of ports on the shelf i.e. on the entire network node) is updated every F seconds, where F is the frame interval for the frames sent on the high-speed bus. In a preferred embodiment, F is 125 microseconds, corresponding to 8000 frames per second.
- the optical signal performance data rate for the high speed bus can be increased - at increased cost - thus decreasing the frame interval F, and increasing the frequency of a complete port update for all N ports in the system.
- a trouble condition is detected at the input 220 (such as loss of signal) or at the output (such as signal degradation) 221 , then that condition will be reported on the high speed bus 201 and, as described above, will be forwarded to each of the CM 203 and EM 204 in no more than one cycle of the high-speed bus, or frame interval F.
- the frame interval equal to 125 microseconds, reporting occurs within no more than 125 microseconds, and statistically on the average in half that time.
- an entire frame interval F plus transmission time on the high- speed bus is the absolute maximum time it would take for this information to be communicated to the CM 203 and EM 204, inasmuch as if a trouble condition occurs in a given port, say Port N, right after that port's status has been reported, it will be picked up in the immediately next frame following, or within one frame interval F.
- the maximum interval between occurrence of a trouble condition at a given port and its reporting on its high speed bus timeslot to the CM 203 and EM 204 is the frame interval of 125 microseconds, as any transmission time within the bus is negligible.
- EM 204 which continually monitors the data from frame to frame, detects a change-of state, via an interrupt. The CM 203 then initiates an alternate path calculation and notifies neighboring network nodes of the new configuration, while the EM 204 prepares for a switch reconfiguration. This operation involves some software processing, primarily analysis and database lookup, and takes on the order of 5 milliseconds.
- the total restoration time is comprised of the internal processing time of approximately 15.125 milliseconds plus the t n2n , or the external node-to-node message time.
- the high speed bus of the present invention offers a substantial decrease in internal detection and processing times when compared to conventional control and messaging architectures (i.e., 15.125 milliseconds versus 1.020 seconds, or nearly 2 orders of magnitude).
- Figure 4 depicts, from a preferred embodiment of the invention, the system of Figure 2 in more detail. The additional details therein illustrated will next be described.
- the Optical Performance Monitoring (OPM) data is gathered by a dedicated hardware device (e.g., an FPGA, ASIC, or gate array) that is resident on each of the Optical Module (OM) circuit boards.
- the depicted system uses an FPGA 410 located in each OM 415 for this purpose.
- the actual monitoring is accomplished by the OPM device 411.
- the logical IOM and OOM are actually one physical bidirectional OM 415.
- the Figure shows one GM 420 on the top far left, and the remainder at the top center of the figure.
- the interface to the OPM devices 411 is a direct, point-to-point, parallel interface 470 through which the OPM devices 411 are sampled.
- the interface is programmable, is under software control, and in this embodiment can support up to one million 16-bit samples per second.
- the data that is collected is then forwarded from each OM 415 to the OM higher level controller, the Group Manager (GM) 420, through a 155 Mb/s serial data link 480.
- the data is formatted essentially as shown in Figure 3, as described below, with the exception that the Switch Map 302 (with reference to Fig. 3) is not included.
- each of the sixteen OM circuit boards 415 in an I/O shelf (such is the term for the total network nodal, or network element, system, comprising various boards and subsystems) pass their respective data to their Group Manager 420 controller via separate 155 Mb/s serial data links.
- each OM board 415 transmits 88 bytes of data (4 ports worth, recall each OM 415, 415A has four optical ports in the depicted exemplary embodiment) to its GM controller 420.
- this transaction requires about 4.6 microseconds. This data transmission is repeated on each of the other OM boards 415A in the other I/O shelves.
- Each GM 420, 420A contains a link- hub device 460 that terminates the sixteen 155 Mb/s data links 480 from all of its subtending OM circuit boards 415, 415A.
- the link-hub device 460 on the GM 420 is a dedicated hardware device (such as, e.g., a FPGA, ASIC, or Gate Array) that (i) collects the data from all of the 16 OM serial links 480, (ii) adds the current state of the Switch Map (which is sourced by the switch manager (“SWM") 425, 425A and stored in memory on the GM 420, 420A), (iii) formats the data according to the protocol depicted in Figure 3, and (iv) transfers the data to the Dual Port Ram (DPR) 490 on the GM 420 (not shown in the other GMs 420A for simplicity).
- DPR Dual Port Ram
- the DPR memory space 490 where the data is stored acts as a transmit buffer for the high-speed bus, here shown as a GbE interface, whose I/O forms the physical high-speed bus.
- a GbE interface whose I/O forms the physical high-speed bus.
- handshaking between the FPGA 460 and DPR 490 keeps the transmit buffer up-to-date with the current OPM data, while the GbE interface 402 packetizes the data from the buffer and sends it out on to the high-speed bus 401.
- All of the second-level controller GM's 420, 420A and SWM's 425, 425A in the system are equipped with these elements of the high-speed bus interface (uP 491 , DPR 490, GbE 402, link-hub 460, which are shown in detail in the leftmost GM 420).
- the first level controllers, Equipment Manager (EM) 435 and Connection Manager (CM) 430 are also equipped with GbE interfaces that connect with the high-speed bus.
- all of the first level controllers can communicate with one another via a compact PCI bus.
- the Internet Gateway (IG) circuit board 431 which can be considered an extension of the CM 430, provides the restoration communication channels, both electrical and optical, that are used to signal other network nodes in the network. For example, trouble conditions in a local node that are reflected in the high-speed bus data and seen by the CM 430 can trigger the IG 431 restoration communication channels to inform other nodes to take appropriate path rerouting actions (optical switching) or other remedial action, thus propagating the restoration information throughout the data network.
- IG Internet Gateway
- the high-speed bus data is made available to all of the controllers with GbE interfaces, where the packets are received and the payload data (OPM data, switch maps, etc.) is placed in a receive buffer either in on-board memory (as in the case of the CM 430 and EM 435) or in DPR 490 (as in the case of the GM 420, 420A and SWM 425, 425A).
- the payload data OPM data, switch maps, etc.
- DPR 490 data in the receive buffer (DPR 490) on the GM 420, 420A and SWM 425, 425A is extracted by the Link Hub 460 where it is formatted and forwarded to the OM 415 and SW 417 circuit boards over their respective 155 Mb/s serial data links.
- Each serial data link is terminated in the FPGA 410 resident on said OM and SW boards where the link data is available for update to internal registers in the FPGA where, for example, OPM 411 threshold changes (in the case of the OM 415) or cross-connect changes (in the case of the SW 417) can be initiated.
- CM 430 and EM 435 Data in on-board memory (receive buffer) on each of the CM 430 and EM 435 is extracted and processed by the local microprocessor, labeled as "uP" which in turn can initiate restoration messages (via CM 430 and IG 431 ) or reconfigure cross-connects and OPM 411 parameters (via EM 435).
- the high-speed bus is a bidirectional, multinode bus
- contention is managed in a similar fashion to the CSMA/CD protocol that is used in 10/100Base-T LAN networks.
- the high speed bus specification is as follows:
- Transport Medium i) Inter-Shelf: Optical ii) Intra-Shelf: Electrical
- Transport Medium Gigabit Ethernet
- Switch Map 302 Current input/output port association through the optical switch.
- Port Number Bidirectional port number identifier for next set of data.
- Total number of ports (N) in the example is 512.
- XMT OSNR Current optical SNR reading in transmit direction on port currently identified.
- Receive Optical Power Current optical power reading in receive direction on port currently identified.
- Receive Optical Signal-to-Noise Ratio (RCV OSNR): Current optical SNR reading in receive direction on port currently identified.
- Receive Thresholds (RCV THRSH): Indication of optical power and optical SNR threshold crossings in the transmit direction on port currently identified.
- CRC Cyclical Redundancy Checksum over current port data.
- EOP End of Packet identifier
- the following fields comprise one frame and each field has the following bytes assigned to it.
- the first four bytes of each frame have an SOP or start of packet identification; this is depicted as 301 in Figure 3 being bytes B1 through B4.
- the next block of bytes comprises a switch map 302. This gives the totality of port assignments connecting a given input port to a given output port.
- N the total number of ports
- the switch map field as a whole will use 2N bytes, where N equals the total number of ports on a system.
- the next block of bytes consists of the optical signal parameters for Port 1 , and is identified as 305 in Figure 3.
- the first four bytes give the port ID, being bytes B1029 through B1032, as shown in Figure 3.
- the next two bytes, being B1033 and B1034 contain the transmit optical power of port 1 and the following two bytes, B1035 and B1036, give the transmit optical signal to noise ratio.
- the next four bytes, being B1037 through B1040 give the receive optical power and the receive optical signal to noise ratio, respectively, for Port 1. It is noted that transmit values are measured at points 221 in Figure 2, and receive values at points 220 in Figure 2.
- the next four bytes being B1041 through B1044, give the transmit thresholds and the receive thresholds, and the final four bytes give the cyclical redundancy check sum over the entire port data; these are depicted as bytes B1045 through B1048 in Figure 3.
- a given port requires 20 bytes to fully encode the optical signal parameters.
- the interim ports, beings ports 2 through N -1 are not shown, merely designated by a vertical line comprised of dots between bytes B1048 and B11 ,249 in Figure 3.
- Figure 3 ends showing the identical fields for port N as shown for Port 1 , which occupies 20 bytes from B11249 through B 11268; that whole block of 20 bytes is designated as 320 in Figure 3. Finally, at the end of a frame, in parallel fashion to the beginning of the frame, there is an end of packet identifier occupying four bytes, being bytes B11269 through B11272 in Figure 3, therein designated 330.
- the total number of bytes utilized by a frame in the depicted example of Figure 3 does not equal the specified maximum bytes per frame at a bit rate of one gigabyte per second.
- the maximum is 15,625 under the depicted bit rate in this example.
- Increasing the bit rate will, obviously, allow more data per frame or the same frame to be transmitted with a shorter frame interval, as may be desired by the user in given circumstances.
- the bytes per frame can be decreased and the frame interval F decreased as well, thus increasing the update frequency.
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU1126402A AU1126402A (en) | 2000-10-06 | 2001-09-26 | Improved network restoration capability via dedicated hardware and continuous performance monitoring |
Applications Claiming Priority (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US23836400P | 2000-10-06 | 2000-10-06 | |
US23829500P | 2000-10-06 | 2000-10-06 | |
US60/238,364 | 2000-10-06 | ||
US60/238,295 | 2000-10-06 | ||
US09/852,582 US7009210B2 (en) | 2000-10-06 | 2001-05-09 | Method and apparatus for bit-rate and format insensitive performance monitoring of lightwave signals |
US09/852,582 | 2001-05-09 | ||
US09/931,725 US20020133734A1 (en) | 2000-10-06 | 2001-08-17 | Network restoration capability via dedicated hardware and continuous performance monitoring |
US09/931,725 | 2001-08-17 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2002031620A2 true WO2002031620A2 (en) | 2002-04-18 |
WO2002031620A3 WO2002031620A3 (en) | 2002-07-04 |
Family
ID=27499915
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2001/030000 WO2002031620A2 (en) | 2000-10-06 | 2001-09-26 | Improved network restoration capability via dedicated hardware and continuous performance monitoring |
Country Status (2)
Country | Link |
---|---|
AU (1) | AU1126402A (en) |
WO (1) | WO2002031620A2 (en) |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4947459A (en) * | 1988-11-25 | 1990-08-07 | Honeywell, Inc. | Fiber optic link noise measurement and optimization system |
US5317198A (en) * | 1990-06-26 | 1994-05-31 | The Mitre Corporation | Optically controlled remote by-pass switch |
US5436624A (en) * | 1992-01-20 | 1995-07-25 | Madge Network Limited | Communication system with monitoring means connected in parallel to signal carrying medium |
US5522030A (en) * | 1991-10-16 | 1996-05-28 | Fujitsu Limited | Fault control system for communications buses in multi-processor system |
US5590118A (en) * | 1994-08-23 | 1996-12-31 | Alcatel N.V. | Method for rerouting a data stream |
US5884017A (en) * | 1995-12-29 | 1999-03-16 | Mci Communications Corporation | Method and system for optical restoration tributary switching in a fiber network |
US6005694A (en) * | 1995-12-28 | 1999-12-21 | Mci Worldcom, Inc. | Method and system for detecting optical faults within the optical domain of a fiber communication network |
US6130876A (en) * | 1997-09-24 | 2000-10-10 | At&T Corp | Method and apparatus for restoring a network |
US6141319A (en) * | 1996-04-10 | 2000-10-31 | Nec Usa, Inc. | Link based alternative routing scheme for network restoration under failure |
-
2001
- 2001-09-26 WO PCT/US2001/030000 patent/WO2002031620A2/en active Application Filing
- 2001-09-26 AU AU1126402A patent/AU1126402A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4947459A (en) * | 1988-11-25 | 1990-08-07 | Honeywell, Inc. | Fiber optic link noise measurement and optimization system |
US5317198A (en) * | 1990-06-26 | 1994-05-31 | The Mitre Corporation | Optically controlled remote by-pass switch |
US5522030A (en) * | 1991-10-16 | 1996-05-28 | Fujitsu Limited | Fault control system for communications buses in multi-processor system |
US5436624A (en) * | 1992-01-20 | 1995-07-25 | Madge Network Limited | Communication system with monitoring means connected in parallel to signal carrying medium |
US5590118A (en) * | 1994-08-23 | 1996-12-31 | Alcatel N.V. | Method for rerouting a data stream |
US6005694A (en) * | 1995-12-28 | 1999-12-21 | Mci Worldcom, Inc. | Method and system for detecting optical faults within the optical domain of a fiber communication network |
US5884017A (en) * | 1995-12-29 | 1999-03-16 | Mci Communications Corporation | Method and system for optical restoration tributary switching in a fiber network |
US6141319A (en) * | 1996-04-10 | 2000-10-31 | Nec Usa, Inc. | Link based alternative routing scheme for network restoration under failure |
US6130876A (en) * | 1997-09-24 | 2000-10-10 | At&T Corp | Method and apparatus for restoring a network |
Also Published As
Publication number | Publication date |
---|---|
AU1126402A (en) | 2002-04-22 |
WO2002031620A3 (en) | 2002-07-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0986226B1 (en) | Ip packet communication apparatus | |
JP2783164B2 (en) | Communication network | |
US7110394B1 (en) | Packet switching apparatus including cascade ports and method for switching packets | |
US5991264A (en) | Method and apparatus for isolating network failures by applying alarms to failure spans | |
EP1206858B1 (en) | System and method for packet transport in a ring network | |
US20010038471A1 (en) | Fault communication for network distributed restoration | |
JP2005521330A (en) | Supervisory channel in optical network systems | |
US6798991B1 (en) | Optical communication systems, optical communication system terminal facilities, optical communication methods, and methods of communicating within an optical network | |
JPH10233735A (en) | Fault recovery control method | |
JP2002057685A (en) | Access network from one point to multiple points | |
US20020133734A1 (en) | Network restoration capability via dedicated hardware and continuous performance monitoring | |
JP3811007B2 (en) | Virtual connection protection switching | |
US7477595B2 (en) | Selector in switching matrix, line redundant method, and line redundant system | |
CA2129097A1 (en) | Fast packet switch | |
EP1113611B1 (en) | Method and apparatus for passing control information in a bidirectional line switched ring configuration | |
WO2002031620A2 (en) | Improved network restoration capability via dedicated hardware and continuous performance monitoring | |
WO1999001963A1 (en) | Self-healing meshed network | |
JP5357436B2 (en) | Transmission equipment | |
CN100372334C (en) | Device and method for realizing Infini Band data transmission in optical network | |
US5638366A (en) | Data transport for internal messaging | |
US20020122219A1 (en) | Optical supervisory channel | |
US8155515B2 (en) | Method and apparatus for sharing common capacity and using different schemes for restoring telecommunications networks | |
JP3631687B2 (en) | Ethernet communication system and Ethernet device | |
CN1753341B (en) | Protection method based on data business of SDH/SONET and its device | |
US6985443B2 (en) | Method and apparatus for alleviating traffic congestion in a computer network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
AK | Designated states |
Kind code of ref document: A3 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
122 | Ep: pct application non-entry in european phase | ||
NENP | Non-entry into the national phase |
Ref country code: JP |