US20070237085A1 - System and methodology for fast link failover based on remote upstream failures - Google Patents

System and methodology for fast link failover based on remote upstream failures Download PDF

Info

Publication number
US20070237085A1
US20070237085A1 US11/398,519 US39851906A US2007237085A1 US 20070237085 A1 US20070237085 A1 US 20070237085A1 US 39851906 A US39851906 A US 39851906A US 2007237085 A1 US2007237085 A1 US 2007237085A1
Authority
US
United States
Prior art keywords
network
network device
status
notification
change
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/398,519
Other versions
US8886831B2 (en
Inventor
Gokul Tirumalai
Albert Mitchell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cisco Technology Inc
Original Assignee
Cisco Technology Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cisco Technology Inc filed Critical Cisco Technology Inc
Priority to US11/398,519 priority Critical patent/US8886831B2/en
Assigned to CISCO TECHNOLOGY, INC. reassignment CISCO TECHNOLOGY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MITCHELL, ALBERT HYDE, TIRUMALAI, GOKUL S.
Publication of US20070237085A1 publication Critical patent/US20070237085A1/en
Application granted granted Critical
Publication of US8886831B2 publication Critical patent/US8886831B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04QSELECTING
    • H04Q3/00Selecting arrangements
    • H04Q3/0016Arrangements providing connection between exchanges
    • H04Q3/0062Provisions for network management
    • H04Q3/0087Network testing or monitoring arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/22Alternate routing

Definitions

  • the present invention generally relates to management of switched computer networks and, more particularly, to propagating detected changes in upstream connectivity to downstream devices.
  • a computer network generally includes a number of devices, including switches, routers and hubs, connected so as to allow communication among the devices.
  • the devices within a network may be generally categorized into two classes: end stations connected along the edge of a network and network devices. End stations typically include devices that primarily supply or consume information, such as workstations, desktop PCs, printers, servers, hosts, fax machines, and other such devices.
  • Network devices typically include devices that primarily transfer information between other devices (e.g., between end stations and other network devices), such as gateways, switches, routers, and other such devices. In other words, network devices generally provide connectivity between end stations and the network core.
  • FIG. 1 illustrates an exemplary network 100 in which a plurality of end stations 110 are connected to a network core 120 via multiple layers of switches 130 .
  • the network may be modeled in accordance with the Open System Interconnection (OSI) model, which defines a networking framework for implementing protocols in multiple (seven) layers.
  • the switches 130 may represent Data Link or “Layer 2 ” layer switching, that encodes and decodes data packets into bits. Layer 2 switches implement the network transmission protocol and handle errors in the physical layer (Ethernet, optics, etc.), flow control and frame synchronization.
  • switches 130 include aggregation switches 132 (SW 1 and SW 2 ), and access switches 134 (SW 3 and SW 4 ).
  • the end stations 110 may be file servers, data servers, workstations, personal computers (PCs), and the like.
  • end stations 110 may each have multiple network interface cards NICs 112 (or other type network interface/adapter), allowing the end stations 110 to connect to the network through primary links/connections (Lp) and secondary links (Ls).
  • NICs 112 or other type network interface/adapter
  • the end stations may be configured to connect to the network through a primary path (via SW 3 and SW 1 ) by default and switch (or “failover”) to a secondary path (via SW 4 and SW 2 ) if a connectivity problem is detected in the primary path, such as a loss of connectivity to a link L C between switch SW 1 and the network core 120 .
  • end stations do not become aware of a change in upstream connectivity until they discover it on their own. For example, end stations may only discover a loss of upstream connectivity after experiencing consistent protocol packet timeouts. As a result, after a loss in network connectivity on the primary path, end stations may not immediately switch from their primary path to their alternative path despite a loss in network connectivity on the primary path, which may degrade network performance.
  • Embodiments of the present invention provide techniques for propagating changes in upstream connectivity downstream to connected end stations.
  • One embodiment provides a method for managing a network path between at least one end station and a network.
  • the method generally includes detecting, by a first switching device, a change in status of a first upstream communication link between the first switching device and the network, sending, by the first switching device, notification of the change in status to a second switching device in communication with the end station via a first downstream communication link, and, in response to receiving the notification, disabling by the second switch, the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
  • a network switching system generally including at least a first switching device and at least a second switching device.
  • the first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core.
  • the second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable a first downstream communication link between the second switching device and an end station.
  • their may be one or more intermediate switching devices, between the first and second switching devices, that receive the notification of the change in status from the first switching device and propagate the notification (possibly through other intermediate switches) to the second switching device that ultimately disables the first downstream communication link with the end station.
  • Another embodiment provides a system generally including a plurality of end stations, at least a first switching device, and at least a second switching device.
  • the first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core.
  • the second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third switching device.
  • an aggregation switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations.
  • the operations generally include detecting a change in status of a first upstream communication link between the first switching device and a network and sending notification of the change in status to a second switching device in communication with an end station via a first downstream communication link.
  • an access switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations.
  • the operations generally include in response to receiving, from an upstream switching device, notification of a detected change in status between the first switching device and a network, disabling at least one downstream communication link between the access switching device and at least one end station.
  • FIG. 1 illustrates an exemplary multilayer switched network.
  • FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention.
  • FIGS. 3A-3D illustrate propagating, and responding to, changes in upstream connectivity, in accordance with one embodiment of the present invention.
  • FIG. 4 illustrates the routing of traffic through links maintained between access and aggregation devices, in accordance with one embodiment of the present invention.
  • FIG. 5 is a flow diagram of exemplary operations for propagating changes in connectivity, in accordance with embodiments of the present invention.
  • Embodiments of the present invention generally provide methods and apparatus for propagating changes in upstream connectivity downstream.
  • an upstream switch connected to a network core may detect a change in connectivity and, in response, send a message indicating the change to a downstream switch connected to an end station.
  • the downstream switch may disable a link with end station. Disabling the link may cause the end station to promptly switch (fail over) to a different link, in an effort to establish an alternate path to the network.
  • network traffic routed through that switch need not be re-routed and the link can continue to carry traffic destined for other networks.
  • end station generally refers to any device capable of a network connection via one or more network interfaces and need not necessarily reside at an edge of a network. Examples of end stations include a server, PC, fax machine, printer, or the like.
  • switch or switching device generally refers to any type device that switches network traffic, such as a Layer 2 switch (per the OSI model) that switches Ethernet traffic.
  • access device generally refers to a network element that is connected to one or more end stations.
  • aggregation device generally refers to a network element that connects to one or more access devices.
  • upstream generally refers to a path (or device along a path) in the direction from an end station towards the core of the network.
  • downstream generally refers to a path (or device along a path) in the direction from the core of the network to the edge (e.g., towards an end station).
  • link or connection may be used herein interchangeably and generally refer to any connection between devices, such as a Layer 2 network connection between two devices.
  • failure generally refers to a loss of connectivity to a link from a device, regardless of the particular cause, including if the link is intentionally disabled by another device.
  • Embodiments of the present invention allow changes in upstream connectivity to be communicated downstream, thereby allowing an end station to quickly make decisions regarding which paths to use to the network.
  • the techniques described herein may be implemented, for example, as software methods (e.g., executable instructions residing in memory) running on network devices (e.g., aggregation devices and/or access devices) that are directly or indirectly connected to a network interface of an end station.
  • the indirectly connected switches may track the link status on upstream interfaces and communicate any change in connectivity (loss/restoration) to a set of downstream interfaces, for example, by sending a message.
  • Such a message may be intercepted by downstream network devices, such as access switches directly connected to network interfaces of the end stations.
  • these switches may take appropriate action based on the message communicated, for example, disabling corresponding active links if the message indicates loss of connectivity. This allows conventional end stations with redundant network interfaces to detect that their primary connection is down and to automatically switch over to their secondary network interfaces.
  • FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention.
  • the operations 200 may be collectively performed by a combination of switches, such as aggregation switches connected to a network core and access switches connected to an end station.
  • switches such as aggregation switches connected to a network core and access switches connected to an end station.
  • FIGS. 3A-3D illustrate the status of different links in the network as connectivity changes therein.
  • FIGS. 3A-3D To facilitate understanding, only a limited portion of the network shown in FIG. 1 is illustrated in FIGS. 3A-3D . However, those skilled in the art will recognize that the concepts described herein may be advantageously scaled to accommodate any size and type of switching network having any number of multiple switching layers (e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements), with potentially hundreds of end stations.
  • any number of multiple switching layers e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements
  • the operations begin, at step 202 , by detecting a change in upstream connectivity between a first switch and a network core.
  • the exemplary network shown includes a layer of aggregation switches 332 SW 1 and SW 2 and a layer of access switches 334 SW 3 and SW 4 .
  • a primary path between end stations 110 and the network core 120 may be established through SW 3 and SW 1 via first network interface cards 112 (NIC 1 ).
  • NIC 1 first network interface cards
  • a loss of connectivity between SW 1 and the network core may occur for any number of reasons (e.g., device failure, such as failure of a laser in a fiber optic connection, failure in the physical connection/cable, or loss of power).
  • switches SW 3 may have downstream connections to both end stations and other switches, not just to one another.
  • the change in connectivity is communicated to a second switch located downstream between the first switch and an end station.
  • the first switch SW 1 may communicate the change in connectivity (a loss in this example), to second switch SW 3 via a change in connectivity message.
  • the contents of the message may vary, as may the recipients.
  • a user e.g., network administrator
  • a downstream connection between the second switch and the end station is modified in response to the change in connectivity.
  • SW 3 may decipher the message from SW 1 and, in response, disable (bring down) the link between the primary NICs of the end stations 110 .
  • Exactly how the link is disabled may depend on the corresponding type of physical medium. For example, for an Ethernet connection this may entail simply ceasing to transmit a signal (e.g., by disabling a laser for a fiber optic Ethernet connection or bringing down the voltage level for a copper Ethernet connection).
  • end stations equipped with redundant network interfaces may be configured to react to such a loss in connectivity to a primary link by automatically switching (or “failing over”) to a secondary interface in order to establish an alternate path to the network.
  • end stations 110 may automatically switch over to secondary NICs to establish an alternate path to the core network via SW 4 and SW 2 .
  • switch SW 1 may forward another message to SW 3 (and/or other connected devices) indicating such.
  • SW 3 may bring the link with the connected end station back up, restoring it to an operational state.
  • one or more of the end stations 110 may switch back to the primary connections immediately, or at any appropriate later time, for example, depending on the particular circumstance.
  • end stations may switchover much more rapidly than in conventional systems that rely on waiting for a higher layer protocol running on these end stations to detect that its link to the network is not working properly.
  • These higher layer protocols typically rely upon timers to fail, and also have to account for network congestion to avoid erroneously detecting “false” failures.
  • upstream switches e.g., SW 1 and SW 2
  • downstream devices e.g., SW 3 and SW 4
  • L X links
  • the particular manner in which changes in upstream connectivity are propagated down to an end station may vary, depending on a particular implementation.
  • the particular format and contents of the message sent by an upstream device to communicate a change in upstream connectivity may vary.
  • the contents of the message may include at least an indication of the detected change (e.g., loss or restoration), as well as any other suitable information, such as identification of an intended recipient (destination).
  • the type and/or number of devices to receive a message indicating a change in connectivity may also vary depending on a particular implementation and, for some embodiments, may be configurable by a user (e.g., a network administrator).
  • messages may vary with different embodiments and different types of messages may be sent (e.g., broadcast to many devices or sent directly to identified recipients).
  • messages may include a common set of fields including, but not limited to one or more of the following: 1) a Switch identifier (source of the msg); 2) a Msg_type/Action_type (e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if a user reconfigures a link from being the cause of downstream interfaces being brought down); 3) an Upstream Link Identifier (which may be at least be unique on the source switch); 4) a SW Version of some type; 5) a msg counter for a number of msgs of this type sent; and 6) a time stamp of action occurrence.
  • a Switch identifier source of the msg
  • Msg_type/Action_type e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if
  • the Upstream Link Identifier may be a field in the message that identifies the upstream interface.
  • the message may include an identifier of a group that is effected by the change in connectivity of an upstream interface.
  • the upstream interface may be a single physical interface, or it could be a logical grouping of physical interfaces.
  • the particular contents of a message may depend on the particular type of message being sent. For some embodiments, there may be at least two types of messages, for example, broadcast or direct, with each having different contents. Regardless of the implementation, msg contents would most likely contain a common set of fields, such as those listed above. The field contents, however, would likely change depending up on the action and the source switch.
  • a message may be broadcast to multiple devices connected to a switch detecting a change in communication status. How the message is treated may depend on the particular recipient. As an example, a switch, such as SW 3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto.
  • a switch such as SW 3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto.
  • a user may be able to configure a particular set of downstream interfaces that should be brought down upon receiving a message indicating a loss of upstream connectivity. For example, as illustrated in FIG. 5 , a user may be able to define a first group 510 1 of end stations with primary interfaces connected to a switch SW 3 that should be brought down in the event that connectivity between SW 1 and the network core (L C ) is lost. This grouping may define a logical connection between these end stations and an upstream interface. As a result, in response to detecting a change in connectivity of the upstream interface, switch SW 3 may disable the end stations in the first group 510 1 . In response to the primary interface being disabled, the end stations may automatically switch over to secondary interfaces connected to SW 4 .
  • a second group of end stations 510 2 may not have a secondary NIC and, thus should not be brought down in the event that upstream connectivity is lost.
  • Such end stations e.g., older “legacy” end stations
  • the downstream interfaces connecting to such end stations may be configured to be maintained (not brought down) by the user in the case of an upstream link losing connectivity.
  • end stations in Group 2 may continue to use the same interface with SW 3 and their traffic may be routed to the network core via a redundant link (L Z ) between SW 1 and SW 2 to reach the core. While the bandwidth of the redundant link L Z may not be as high as the direct link to the core (link L C ) and may not be meant to carry sustained traffic to the core, it may still allow the end station to remain active.
  • the techniques described herein may be configurable by a user, for example, via a user command to run automatically, on demand, or be disabled. In some cases, a user may be allowed to define logical grouping of upstream interfaces to downstream interfaces. Further, for some embodiments, in addition to, or as an alternative to selecting groups, the techniques described herein could be configurable to be enabled in a variety of ways with varying scope, such as on a “per port” basis, via a particular connection (e.g., a VLAN), or switch-wide.
  • the downstream device may bring down a connection with an end station. Bringing down the connection may result in the end station promptly switching over to a redundant connection. Further, by maintaining the link between the intermediate switch and the upstream device, network traffic routed therethrough need not be re-routed.

Abstract

A method and apparatus that promote expedited fast link failover decisions by and end station with redundant multilayer switching paths to a network are disclosed. In response to detecting a change in connectivity with an upstream link to a network, an upstream switch may send a notification to an intermediate switch with a downstream link to the end station. In response to the notification, the intermediate switch may disable (“bring down”) the downstream link to the end station. Bringing the downstream link down in this manner may cause the end station to switch (“failover”) to a different downstream link, for example, allowing the end station to establish connectivity to the network via a redundant path.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention generally relates to management of switched computer networks and, more particularly, to propagating detected changes in upstream connectivity to downstream devices.
  • 2. Description of the Related Art
  • A computer network generally includes a number of devices, including switches, routers and hubs, connected so as to allow communication among the devices. The devices within a network may be generally categorized into two classes: end stations connected along the edge of a network and network devices. End stations typically include devices that primarily supply or consume information, such as workstations, desktop PCs, printers, servers, hosts, fax machines, and other such devices. Network devices typically include devices that primarily transfer information between other devices (e.g., between end stations and other network devices), such as gateways, switches, routers, and other such devices. In other words, network devices generally provide connectivity between end stations and the network core.
  • FIG. 1 illustrates an exemplary network 100 in which a plurality of end stations 110 are connected to a network core 120 via multiple layers of switches 130. In some cases, the network may be modeled in accordance with the Open System Interconnection (OSI) model, which defines a networking framework for implementing protocols in multiple (seven) layers. The switches 130 may represent Data Link or “Layer 2” layer switching, that encodes and decodes data packets into bits. Layer 2 switches implement the network transmission protocol and handle errors in the physical layer (Ethernet, optics, etc.), flow control and frame synchronization. In the illustrated example, switches 130 include aggregation switches 132 (SW1 and SW2), and access switches 134 (SW3 and SW4). The end stations 110 may be file servers, data servers, workstations, personal computers (PCs), and the like.
  • In an effort to provide greater reliability, such networks are often designed to provide redundant Layer 2 connectivity between end stations 110 and the network core 120. To this end, at least some end stations 110 may each have multiple network interface cards NICs 112 (or other type network interface/adapter), allowing the end stations 110 to connect to the network through primary links/connections (Lp) and secondary links (Ls). For example, the end stations may be configured to connect to the network through a primary path (via SW3 and SW1) by default and switch (or “failover”) to a secondary path (via SW4 and SW2) if a connectivity problem is detected in the primary path, such as a loss of connectivity to a link LC between switch SW1 and the network core 120.
  • In conventional systems, end stations do not become aware of a change in upstream connectivity until they discover it on their own. For example, end stations may only discover a loss of upstream connectivity after experiencing consistent protocol packet timeouts. As a result, after a loss in network connectivity on the primary path, end stations may not immediately switch from their primary path to their alternative path despite a loss in network connectivity on the primary path, which may degrade network performance.
  • Therefore, there is a need for techniques that cause end stations to switch over to alternative paths more rapidly in response to remote upstream connectivity changes.
  • SUMMARY OF THE INVENTION
  • Embodiments of the present invention provide techniques for propagating changes in upstream connectivity downstream to connected end stations.
  • One embodiment provides a method for managing a network path between at least one end station and a network. The method generally includes detecting, by a first switching device, a change in status of a first upstream communication link between the first switching device and the network, sending, by the first switching device, notification of the change in status to a second switching device in communication with the end station via a first downstream communication link, and, in response to receiving the notification, disabling by the second switch, the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
  • Another embodiment provides a network switching system generally including at least a first switching device and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable a first downstream communication link between the second switching device and an end station. For some embodiments, their may be one or more intermediate switching devices, between the first and second switching devices, that receive the notification of the change in status from the first switching device and propagate the notification (possibly through other intermediate switches) to the second switching device that ultimately disables the first downstream communication link with the end station.
  • Another embodiment provides a system generally including a plurality of end stations, at least a first switching device, and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third switching device.
  • Another embodiment provides an aggregation switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include detecting a change in status of a first upstream communication link between the first switching device and a network and sending notification of the change in status to a second switching device in communication with an end station via a first downstream communication link.
  • Another embodiment provides an access switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include in response to receiving, from an upstream switching device, notification of a detected change in status between the first switching device and a network, disabling at least one downstream communication link between the access switching device and at least one end station.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • So that the manner in which the above recited features of the present invention can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to embodiments, some of which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
  • FIG. 1 illustrates an exemplary multilayer switched network.
  • FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention.
  • FIGS. 3A-3D illustrate propagating, and responding to, changes in upstream connectivity, in accordance with one embodiment of the present invention.
  • FIG. 4 illustrates the routing of traffic through links maintained between access and aggregation devices, in accordance with one embodiment of the present invention.
  • FIG. 5 is a flow diagram of exemplary operations for propagating changes in connectivity, in accordance with embodiments of the present invention.
  • DETAILED DESCRIPTION
  • Embodiments of the present invention generally provide methods and apparatus for propagating changes in upstream connectivity downstream. For some embodiments, an upstream switch connected to a network core may detect a change in connectivity and, in response, send a message indicating the change to a downstream switch connected to an end station. In response to the message, the downstream switch may disable a link with end station. Disabling the link may cause the end station to promptly switch (fail over) to a different link, in an effort to establish an alternate path to the network. Further, by maintaining the link between the intermediate or downstream switch and the upstream device, network traffic routed through that switch need not be re-routed and the link can continue to carry traffic destined for other networks.
  • Methods and apparatus for managing a multilayer switching path are described. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be appreciated, however, to one skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form (absent specific details) in order to avoid unnecessarily obscuring the present invention.
  • To facilitate understanding, the following terms, as used herein, have the following meaning. As used herein, the term end station generally refers to any device capable of a network connection via one or more network interfaces and need not necessarily reside at an edge of a network. Examples of end stations include a server, PC, fax machine, printer, or the like. The term switch (or switching device) generally refers to any type device that switches network traffic, such as a Layer 2 switch (per the OSI model) that switches Ethernet traffic. The term access device generally refers to a network element that is connected to one or more end stations. The term aggregation device generally refers to a network element that connects to one or more access devices. The term upstream generally refers to a path (or device along a path) in the direction from an end station towards the core of the network. The term downstream generally refers to a path (or device along a path) in the direction from the core of the network to the edge (e.g., towards an end station). The terms link or connection may be used herein interchangeably and generally refer to any connection between devices, such as a Layer 2 network connection between two devices. The term failure generally refers to a loss of connectivity to a link from a device, regardless of the particular cause, including if the link is intentionally disabled by another device.
  • Propagating (Downstream) Changes in Upstream Connectivity
  • Embodiments of the present invention allow changes in upstream connectivity to be communicated downstream, thereby allowing an end station to quickly make decisions regarding which paths to use to the network. The techniques described herein may be implemented, for example, as software methods (e.g., executable instructions residing in memory) running on network devices (e.g., aggregation devices and/or access devices) that are directly or indirectly connected to a network interface of an end station. As will be described herein, the indirectly connected switches may track the link status on upstream interfaces and communicate any change in connectivity (loss/restoration) to a set of downstream interfaces, for example, by sending a message. Such a message may be intercepted by downstream network devices, such as access switches directly connected to network interfaces of the end stations. In response, these switches may take appropriate action based on the message communicated, for example, disabling corresponding active links if the message indicates loss of connectivity. This allows conventional end stations with redundant network interfaces to detect that their primary connection is down and to automatically switch over to their secondary network interfaces.
  • FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention. For some embodiments, the operations 200 may be collectively performed by a combination of switches, such as aggregation switches connected to a network core and access switches connected to an end station. To illustrate the potential operation of the switching devices, the operations 200 may be described with reference to network diagrams of FIGS. 3A-3D, which illustrate the status of different links in the network as connectivity changes therein.
  • To facilitate understanding, only a limited portion of the network shown in FIG. 1 is illustrated in FIGS. 3A-3D. However, those skilled in the art will recognize that the concepts described herein may be advantageously scaled to accommodate any size and type of switching network having any number of multiple switching layers (e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements), with potentially hundreds of end stations.
  • The operations begin, at step 202, by detecting a change in upstream connectivity between a first switch and a network core. Referring to FIG. 3A, the exemplary network shown includes a layer of aggregation switches 332 SW1 and SW2 and a layer of access switches 334 SW3 and SW4. As illustrated, a primary path between end stations 110 and the network core 120 may be established through SW3 and SW1 via first network interface cards 112 (NIC1). As illustrated in FIG. 3, however, a loss of connectivity between SW1 and the network core may occur for any number of reasons (e.g., device failure, such as failure of a laser in a fiber optic connection, failure in the physical connection/cable, or loss of power). For some embodiments, switches SW3 may have downstream connections to both end stations and other switches, not just to one another.
  • At step 204, the change in connectivity is communicated to a second switch located downstream between the first switch and an end station. As further illustrated in FIG. 3B, the first switch SW1 may communicate the change in connectivity (a loss in this example), to second switch SW3 via a change in connectivity message. As will be described in greater detail below, the contents of the message may vary, as may the recipients. For some embodiments, a user (e.g., network administrator) may be able to configure the manner in which the message is sent, for example, broadcasting to all devices to respond/ignore or only sending the message to identified devices (e.g., along a path to an end device).
  • At step 206, a downstream connection between the second switch and the end station is modified in response to the change in connectivity. For example, as illustrated in FIG. 3C, SW3 may decipher the message from SW1 and, in response, disable (bring down) the link between the primary NICs of the end stations 110. Exactly how the link is disabled may depend on the corresponding type of physical medium. For example, for an Ethernet connection this may entail simply ceasing to transmit a signal (e.g., by disabling a laser for a fiber optic Ethernet connection or bringing down the voltage level for a copper Ethernet connection).
  • In any case, end stations equipped with redundant network interfaces may be configured to react to such a loss in connectivity to a primary link by automatically switching (or “failing over”) to a secondary interface in order to establish an alternate path to the network. Thus, in the illustrated example as shown in FIG. 3D, end stations 110 may automatically switch over to secondary NICs to establish an alternate path to the core network via SW4 and SW2.
  • For some embodiments, should the connectivity between SW1 and the network core be restored, switch SW1 may forward another message to SW3 (and/or other connected devices) indicating such. In response to the message, SW3 may bring the link with the connected end station back up, restoring it to an operational state. In response to detecting that the primary link has been brought back up, one or more of the end stations 110 may switch back to the primary connections immediately, or at any appropriate later time, for example, depending on the particular circumstance.
  • By propagating an upstream Layer 2 connectivity failure in this manner, switchover times for the end stations may be reduced. In other words, end stations may switchover much more rapidly than in conventional systems that rely on waiting for a higher layer protocol running on these end stations to detect that its link to the network is not working properly. These higher layer protocols typically rely upon timers to fail, and also have to account for network congestion to avoid erroneously detecting “false” failures.
  • In addition to improving switchover times, the techniques described herein also allow upstream switches (e.g., SW1 and SW2) to maintain their interfaces to downstream devices (e.g., SW3 and SW4). This is advantageous in that other network traffic may continue to pass through the interface from access devices to aggregation devices and allows multiple networks to be active that would otherwise be inactive if this link were brought down. These links (LX) may continue to carry traffic for management, as well as other local network traffic.
  • Configuring Upstream Failure Propagation
  • The particular manner in which changes in upstream connectivity are propagated down to an end station may vary, depending on a particular implementation. As an example, the particular format and contents of the message sent by an upstream device to communicate a change in upstream connectivity may vary. For some embodiments, the contents of the message may include at least an indication of the detected change (e.g., loss or restoration), as well as any other suitable information, such as identification of an intended recipient (destination). The type and/or number of devices to receive a message indicating a change in connectivity may also vary depending on a particular implementation and, for some embodiments, may be configurable by a user (e.g., a network administrator).
  • The contents of messages may vary with different embodiments and different types of messages may be sent (e.g., broadcast to many devices or sent directly to identified recipients). In any case, messages may include a common set of fields including, but not limited to one or more of the following: 1) a Switch identifier (source of the msg); 2) a Msg_type/Action_type (e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if a user reconfigures a link from being the cause of downstream interfaces being brought down); 3) an Upstream Link Identifier (which may be at least be unique on the source switch); 4) a SW Version of some type; 5) a msg counter for a number of msgs of this type sent; and 6) a time stamp of action occurrence. For some embodiments, the Upstream Link Identifier may be a field in the message that identifies the upstream interface. As an alternative, the message may include an identifier of a group that is effected by the change in connectivity of an upstream interface. The upstream interface may be a single physical interface, or it could be a logical grouping of physical interfaces.
  • The particular contents of a message may depend on the particular type of message being sent. For some embodiments, there may be at least two types of messages, for example, broadcast or direct, with each having different contents. Regardless of the implementation, msg contents would most likely contain a common set of fields, such as those listed above. The field contents, however, would likely change depending up on the action and the source switch.
  • As illustrated in FIG. 4, for some embodiments, a message may be broadcast to multiple devices connected to a switch detecting a change in communication status. How the message is treated may depend on the particular recipient. As an example, a switch, such as SW3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto.
  • For some embodiments, a user may be able to configure a particular set of downstream interfaces that should be brought down upon receiving a message indicating a loss of upstream connectivity. For example, as illustrated in FIG. 5, a user may be able to define a first group 510 1 of end stations with primary interfaces connected to a switch SW3 that should be brought down in the event that connectivity between SW1 and the network core (LC) is lost. This grouping may define a logical connection between these end stations and an upstream interface. As a result, in response to detecting a change in connectivity of the upstream interface, switch SW3 may disable the end stations in the first group 510 1. In response to the primary interface being disabled, the end stations may automatically switch over to secondary interfaces connected to SW4.
  • A second group of end stations 510 2 may not have a secondary NIC and, thus should not be brought down in the event that upstream connectivity is lost. Such end stations (e.g., older “legacy” end stations) may not have the capability to switch over from a primary to secondary interface. As such, the downstream interfaces connecting to such end stations may be configured to be maintained (not brought down) by the user in the case of an upstream link losing connectivity. As illustrated, end stations in Group 2 may continue to use the same interface with SW3 and their traffic may be routed to the network core via a redundant link (LZ) between SW1 and SW2 to reach the core. While the bandwidth of the redundant link LZ may not be as high as the direct link to the core (link LC) and may not be meant to carry sustained traffic to the core, it may still allow the end station to remain active.
  • For some embodiments, the techniques described herein may be configurable by a user, for example, via a user command to run automatically, on demand, or be disabled. In some cases, a user may be allowed to define logical grouping of upstream interfaces to downstream interfaces. Further, for some embodiments, in addition to, or as an alternative to selecting groups, the techniques described herein could be configurable to be enabled in a variety of ways with varying scope, such as on a “per port” basis, via a particular connection (e.g., a VLAN), or switch-wide.
  • CONCLUSION
  • By propagating upstream connectivity losses detected by an upstream device to a downstream device connected, the downstream device may bring down a connection with an end station. Bringing down the connection may result in the end station promptly switching over to a redundant connection. Further, by maintaining the link between the intermediate switch and the upstream device, network traffic routed therethrough need not be re-routed.
  • While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.

Claims (28)

1. A method for managing a network path between at least one end station and a network, comprising:
detecting, by a first network device, a change in status of a first upstream communication link between the first network device and the network; and
sending, by the first network device, notification of the change in status to a second network device in communication with the end station via a first downstream communication link, the notification designed to cause the second network device to disable the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
2. The method of claim 1, wherein one or more intermediate network devices, between the first and second network devices, receives the notification of the change in status from the first network device and forwards the notification to the second network device.
3. The method of claim 1, wherein sending notification, by the first network device, of the change in status to the second network device comprises:
sending a message from the first network device to one or more network devices in communication with the first network device.
4. The method of claim 1, wherein sending a message from the first network device to one or more network devices in communication with the first network device comprises sending a message to one or more network devices specified by a user.
5. The method of claim 1, wherein the second switch is configured to disable, in response to receiving the notification, a plurality of downstream communication links between the second switch and a plurality of end stations.
6. The method of claim 1, wherein the second switch is configured to forward a message containing the notification to one or more network devices specified by a user.
7. The method of claim 6, wherein the second network device is configured to forward the message without interpreting the message.
8. The method of claim 1, wherein the plurality of end stations are specified by a user.
9. The method of claim 1, wherein the change in communication status comprises a loss in connectivity between the first network device and the network and the method further comprises:
detecting, by the first network device, recovery of connectivity between the first network device and the network; and
sending, by the first network device, notification of the recovery in status to the second network device; and
in response to receiving the notification of the recovery in status, enabling by the second switch, the first downstream communication link.
10. A network switching system, comprising:
at least a first network device configured to detect a change in status of a first upstream communication link between the first network device and a network core; and
at least a second network device configured to receive, from the first network device, notification of the change in status from the first network device and, in response, disable a first downstream communication link between the second network device and an end station.
11. The system of claim 10, further comprising one or more intermediate network devices, between the first and second network devices, configured to receive the notification of the change in status from the first network device and forward the notification to the second network device.
12. The system of claim 10, wherein the first network device is configured to send notification of the change in status to a plurality of network devices specified by a user.
13. The system of claim 10, wherein the second network device is configured to, in response to receiving notification of the change in status from the first network device, disable first downstream communication links between the second switch and a plurality of end stations.
14. The system of claim 13, wherein the plurality of end stations are specified by a user.
15. A system, comprising:
a plurality of end stations;
at least a first network device configured to detect a change in status of a first upstream communication link between the first network device and a network core; and
at least a second network device configured to receive, from the first network device, notification of the change in status from the first network device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third network device.
16. The system of claim 15, wherein the plurality of end stations comprises at least one end station without a second network interface.
17. The system of claim 15, wherein the first network device is configured to send notification of the change in status to a plurality of network devices specified by a user.
18. The system of claim 15, wherein the second network device is configured to, in response to receiving notification of the change in status from the first network device, disable first downstream communication links between the second switch and a limited subset of the plurality of end stations specified by a user.
19. An aggregation device comprising:
a processor; and
a computer readable medium containing instructions which, when executed by the processor, cause the aggregation device to perform operations comprising, detecting a change in status of a first upstream communication link between the aggregation device and a network, and
sending notification of the change in status to a network device in communication with an end station via a first downstream communication link.
20. The aggregation device of claim 19, wherein the operations further comprise:
detecting recovery of connectivity between the aggregation device and the network; and
sending notification of the recovery in status to the network device.
21. The aggregation device of claim 19, wherein the operations comprise:
obtaining a list of network devices specified by a user; and
sending notification of the change in status to the network devices specified by the user.
22. An access device, comprising:
a processor; and
a computer readable medium containing instructions which, when executed by the processor, cause the access device to perform operations comprising, in response to receiving notification of a detected change in status between an upstream network device and a network, disabling at least one downstream communication link between the access device and at least one end station.
23. The access device of claim 22, wherein disabling at least one downstream communication link between the access device and at least one end station comprises disabling downstream communication links between the access device a plurality of end stations.
24. The access device of claim 22, wherein the operations further comprise receiving a user-specified list of the plurality of end stations to be disabled in response to receiving notification of the detected change in status from the upstream switching device.
25. A network device, comprising:
means for receiving notification of a detected change in status between an upstream network device and a network; and
means for disabling at least one downstream communication link between the access device and at least one end station.
26. The network device of claim 25, wherein the means for disabling at least one downstream communication link between the access device and at least one end station comprises an optical driver circuit.
27. The network device of claim 25, wherein the means for receiving notification of a detected change in status between an upstream network device and a network comprises means for receiving a message broadcast from the upstream network device.
28. A network device comprising:
means for detecting a change in status of a first upstream communication link between the network device and a network; and
means for sending notification of the change in status to another network device in communication with an end station via a downstream communication link to cause the other network device to disable the downstream communication link.
US11/398,519 2006-04-05 2006-04-05 System and methodology for fast link failover based on remote upstream failures Expired - Fee Related US8886831B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/398,519 US8886831B2 (en) 2006-04-05 2006-04-05 System and methodology for fast link failover based on remote upstream failures

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/398,519 US8886831B2 (en) 2006-04-05 2006-04-05 System and methodology for fast link failover based on remote upstream failures

Publications (2)

Publication Number Publication Date
US20070237085A1 true US20070237085A1 (en) 2007-10-11
US8886831B2 US8886831B2 (en) 2014-11-11

Family

ID=38575123

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/398,519 Expired - Fee Related US8886831B2 (en) 2006-04-05 2006-04-05 System and methodology for fast link failover based on remote upstream failures

Country Status (1)

Country Link
US (1) US8886831B2 (en)

Cited By (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070277828A1 (en) * 2006-06-05 2007-12-06 Ho Peter C F Flexible connector
US20080049631A1 (en) * 2006-08-22 2008-02-28 Morrill Robert J System and method for monitoring interlayer devices and optimizing network performance
WO2009097615A1 (en) * 2008-02-01 2009-08-06 Allied Telesis Holdings Kabushiki Kaisha Method and system for dynamic link failover management
US20100020680A1 (en) * 2008-07-28 2010-01-28 Salam Samer M Multi-chassis ethernet link aggregation
US20100296390A1 (en) * 2006-11-27 2010-11-25 Honeywell International Inc. Fault tolerance in wireless networks
US7843831B2 (en) 2006-08-22 2010-11-30 Embarq Holdings Company Llc System and method for routing data on a packet network
US7940735B2 (en) 2006-08-22 2011-05-10 Embarq Holdings Company, Llc System and method for selecting an access point
US7948909B2 (en) 2006-06-30 2011-05-24 Embarq Holdings Company, Llc System and method for resetting counters counting network performance information at network communications devices on a packet network
US8000318B2 (en) 2006-06-30 2011-08-16 Embarq Holdings Company, Llc System and method for call routing based on transmission performance of a packet network
US20110211585A1 (en) * 2010-02-26 2011-09-01 Alaxala Networks Corporation Packet relay device
US8015294B2 (en) 2006-08-22 2011-09-06 Embarq Holdings Company, LP Pin-hole firewall for communicating data packets on a packet network
US8040811B2 (en) 2006-08-22 2011-10-18 Embarq Holdings Company, Llc System and method for collecting and managing network performance information
US8064391B2 (en) 2006-08-22 2011-11-22 Embarq Holdings Company, Llc System and method for monitoring and optimizing network performance to a wireless device
US8068425B2 (en) 2008-04-09 2011-11-29 Embarq Holdings Company, Llc System and method for using network performance information to determine improved measures of path states
US8098579B2 (en) 2006-08-22 2012-01-17 Embarq Holdings Company, LP System and method for adjusting the window size of a TCP packet through remote network elements
US8102770B2 (en) 2006-08-22 2012-01-24 Embarq Holdings Company, LP System and method for monitoring and optimizing network performance with vector performance tables and engines
US8107366B2 (en) 2006-08-22 2012-01-31 Embarq Holdings Company, LP System and method for using centralized network performance tables to manage network communications
US8111692B2 (en) 2007-05-31 2012-02-07 Embarq Holdings Company Llc System and method for modifying network traffic
US8125897B2 (en) 2006-08-22 2012-02-28 Embarq Holdings Company Lp System and method for monitoring and optimizing network performance with user datagram protocol network performance information packets
US8130793B2 (en) 2006-08-22 2012-03-06 Embarq Holdings Company, Llc System and method for enabling reciprocal billing for different types of communications over a packet network
US8144586B2 (en) 2006-08-22 2012-03-27 Embarq Holdings Company, Llc System and method for controlling network bandwidth with a connection admission control engine
US8144587B2 (en) 2006-08-22 2012-03-27 Embarq Holdings Company, Llc System and method for load balancing network resources using a connection admission control engine
US8184549B2 (en) 2006-06-30 2012-05-22 Embarq Holdings Company, LLP System and method for selecting network egress
US8189468B2 (en) 2006-10-25 2012-05-29 Embarq Holdings, Company, LLC System and method for regulating messages between networks
US8194643B2 (en) 2006-10-19 2012-06-05 Embarq Holdings Company, Llc System and method for monitoring the connection of an end-user to a remote network
US8194555B2 (en) 2006-08-22 2012-06-05 Embarq Holdings Company, Llc System and method for using distributed network performance information tables to manage network communications
US8199653B2 (en) 2006-08-22 2012-06-12 Embarq Holdings Company, Llc System and method for communicating network performance information over a packet network
US8224255B2 (en) 2006-08-22 2012-07-17 Embarq Holdings Company, Llc System and method for managing radio frequency windows
US8228791B2 (en) 2006-08-22 2012-07-24 Embarq Holdings Company, Llc System and method for routing communications between packet networks based on intercarrier agreements
US8274905B2 (en) 2006-08-22 2012-09-25 Embarq Holdings Company, Llc System and method for displaying a graph representative of network performance over a time period
US8289965B2 (en) 2006-10-19 2012-10-16 Embarq Holdings Company, Llc System and method for establishing a communications session with an end-user based on the state of a network connection
US8307065B2 (en) 2006-08-22 2012-11-06 Centurylink Intellectual Property Llc System and method for remotely controlling network operators
US8358580B2 (en) 2006-08-22 2013-01-22 Centurylink Intellectual Property Llc System and method for adjusting the window size of a TCP packet through network elements
US8407765B2 (en) 2006-08-22 2013-03-26 Centurylink Intellectual Property Llc System and method for restricting access to network performance information tables
US8488447B2 (en) 2006-06-30 2013-07-16 Centurylink Intellectual Property Llc System and method for adjusting code speed in a transmission path during call set-up due to reduced transmission performance
US20130232277A1 (en) * 2011-08-31 2013-09-05 Metaswitch Networks Ltd. Transmitting and Forwarding Data
US8531954B2 (en) 2006-08-22 2013-09-10 Centurylink Intellectual Property Llc System and method for handling reservation requests with a connection admission control engine
US8537695B2 (en) 2006-08-22 2013-09-17 Centurylink Intellectual Property Llc System and method for establishing a call being received by a trunk on a packet network
US8549405B2 (en) 2006-08-22 2013-10-01 Centurylink Intellectual Property Llc System and method for displaying a graphical representation of a network to identify nodes and node segments on the network that are not operating normally
US8576722B2 (en) 2006-08-22 2013-11-05 Centurylink Intellectual Property Llc System and method for modifying connectivity fault management packets
US8619600B2 (en) 2006-08-22 2013-12-31 Centurylink Intellectual Property Llc System and method for establishing calls over a call path having best path metrics
US8717911B2 (en) 2006-06-30 2014-05-06 Centurylink Intellectual Property Llc System and method for collecting network performance information
US8743703B2 (en) 2006-08-22 2014-06-03 Centurylink Intellectual Property Llc System and method for tracking application resource usage
US8743700B2 (en) 2006-08-22 2014-06-03 Centurylink Intellectual Property Llc System and method for provisioning resources of a packet network based on collected network performance information
US8750158B2 (en) 2006-08-22 2014-06-10 Centurylink Intellectual Property Llc System and method for differentiated billing
US8824276B2 (en) * 2012-06-01 2014-09-02 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of MOFRR with dataplane notifications
US8913482B2 (en) 2012-06-01 2014-12-16 Telefonaktiebolaget L M Ericsson (Publ) Enhancements to PIM fast re-route with upstream activation packets
US20150172070A1 (en) * 2012-06-01 2015-06-18 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of mofrr with dataplane notifications
US9094257B2 (en) 2006-06-30 2015-07-28 Centurylink Intellectual Property Llc System and method for selecting a content delivery network
US20150295861A1 (en) * 2013-01-06 2015-10-15 Hangzhou H3C Technologies Co., Ltd. Link problem handling
US9479341B2 (en) 2006-08-22 2016-10-25 Centurylink Intellectual Property Llc System and method for initiating diagnostics on a packet network node
WO2017160306A1 (en) * 2016-03-18 2017-09-21 Hewlett Packard Enterprise Development Lp Responses to loops in networks having a ring topology
US11012369B2 (en) * 2019-07-05 2021-05-18 Dell Products L.P. Aggregated switch path optimization system
CN114679348A (en) * 2020-12-24 2022-06-28 华为技术有限公司 Fault processing method and related product
US11551684B1 (en) * 2017-12-08 2023-01-10 Amazon Technologies, Inc. State detection and responses for electronic devices

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9866470B2 (en) * 2014-01-24 2018-01-09 Red Hat, Inc. Multiple active link aggregators
US9838302B1 (en) * 2015-06-10 2017-12-05 Amazon Technologies, Inc. Managing loss of network connectivity in traffic forwarding systems
US10237157B1 (en) 2015-06-10 2019-03-19 Amazon Technologies, Inc. Managing host failures in a traffic forwarding system
FR3041842A1 (en) * 2015-09-30 2017-03-31 Orange SERVICE RESTORATION SYSTEM PROVIDED BY A RESIDENTIAL GATEWAY
US10097462B2 (en) * 2016-04-02 2018-10-09 Niciria, Inc. Throughput resilience during link failover
US11533231B2 (en) 2019-11-29 2022-12-20 Amazon Technologies, Inc. Configuration and management of scalable global private networks
US11729077B2 (en) 2019-11-29 2023-08-15 Amazon Technologies, Inc. Configuration and management of scalable global private networks
US10999169B1 (en) * 2019-11-29 2021-05-04 Amazon Technologies, Inc. Configuration and management of scalable global private networks
US11336528B2 (en) 2019-11-29 2022-05-17 Amazon Technologies, Inc. Configuration and management of scalable global private networks

Citations (61)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5835696A (en) * 1995-11-22 1998-11-10 Lucent Technologies Inc. Data router backup feature
US6032194A (en) * 1997-12-24 2000-02-29 Cisco Technology, Inc. Method and apparatus for rapidly reconfiguring computer networks
US6330229B1 (en) * 1998-11-09 2001-12-11 3Com Corporation Spanning tree with rapid forwarding database updates
US6392990B1 (en) * 1999-07-23 2002-05-21 Glenayre Electronics, Inc. Method for implementing interface redundancy in a computer network
US20020112072A1 (en) * 2001-02-12 2002-08-15 Maple Optical Systems, Inc. System and method for fast-rerouting of data in a data communication network
US6442133B2 (en) * 2000-05-24 2002-08-27 Nec Corporation Switchboard having a dual switching system
US20020152320A1 (en) * 2001-02-14 2002-10-17 Lau Pui Lun System and method for rapidly switching between redundant networks
US20020167895A1 (en) * 2001-05-08 2002-11-14 Jack Zhu Method for restoring diversely routed circuits
US20020184387A1 (en) * 2001-05-31 2002-12-05 The Furukawa Electric Co., Ltd. Method for connecting between networks, virtual router, and system for connecting between networks by using this virtual router
US20030016624A1 (en) * 1998-05-04 2003-01-23 Bare Ballard C. Path recovery on failure in load balancing switch protocols
US6513129B1 (en) * 1999-06-30 2003-01-28 Objective Systems Integrators, Inc. System and method for managing faults using a gateway
US20030021226A1 (en) * 2001-07-24 2003-01-30 Gal Mor Interconnect and gateway protection in bidirectional ring networks
US6542934B1 (en) * 1997-05-30 2003-04-01 International Business Machines Corporation Non-disruptively rerouting network communications from a secondary network path to a primary path
US20030097470A1 (en) * 2001-09-10 2003-05-22 Roger Lapuh System, device, and method for improving communication network reliability using trunk splitting
US20030137932A1 (en) * 2002-01-23 2003-07-24 Nec Corporation Path fault recovery method, switching-back method after recovery from fault, and node using the same
US20030158964A1 (en) * 2002-02-15 2003-08-21 Naofumi Kobayashi Upper and lower relays and network system
US20030161260A1 (en) * 2002-02-25 2003-08-28 Sundara Murugan Method and apparatus for implementing automatic protection switching functionality in a distributed processor data router
US20030179700A1 (en) * 1999-01-15 2003-09-25 Saleh Ali Najib Method for restoring a virtual path in an optical network using 1‘protection
US6628661B1 (en) * 1998-08-27 2003-09-30 Intel Corporation Spanning tree recovery in computer networks
US20030185225A1 (en) * 2002-03-29 2003-10-02 Wirth Brian Michael Switch and a switching apparatus for a communication network
US20040001449A1 (en) * 2002-06-28 2004-01-01 Rostron Andy E. System and method for supporting automatic protection switching between multiple node pairs using common agent architecture
US6678241B1 (en) * 1999-11-30 2004-01-13 Cisc Technology, Inc. Fast convergence with topology switching
US6697339B1 (en) * 1999-03-04 2004-02-24 3Com Corporation High availability spanning tree with rapid reconfiguration with alternate port selection
US20040047336A1 (en) * 1999-08-23 2004-03-11 Avaya Communication Israel Ltd. Modular bridging-device
US6721275B1 (en) * 1999-05-03 2004-04-13 Hewlett-Packard Development Company, L.P. Bridged network stations location revision
US20040098501A1 (en) * 2002-10-29 2004-05-20 Finn Norman W. Multi-bridge lan aggregation
US20040105390A1 (en) * 2001-05-28 2004-06-03 Nokia Corporation Method and system for implementing a fast recovery process in a local area network
US20040133693A1 (en) * 2000-01-28 2004-07-08 3Com Corporation Multi-point link aggregation spoofing
US6801506B1 (en) * 1999-03-31 2004-10-05 Cisco Technology, Inc. Method and apparatus for providing fast spanning tree re-starts
US20040264364A1 (en) * 2003-06-27 2004-12-30 Nec Corporation Network system for building redundancy within groups
US20050063395A1 (en) * 2003-09-18 2005-03-24 Cisco Technology, Inc. Virtual network device
US6891808B2 (en) * 2000-03-10 2005-05-10 Anritsu Corporation Spanning tree bridge and route change method using the same
US20050108401A1 (en) * 2003-11-13 2005-05-19 Gonda Rumi S. Method for supporting SDH/SONET OAMP on Ethernet
US20050111350A1 (en) * 2003-11-20 2005-05-26 Shinya Kano Method and apparatus for designing backup communication path, and computer product
US6956824B2 (en) * 2001-06-14 2005-10-18 Tropic Networks Inc. Extension of link aggregation protocols over the network
US6987740B1 (en) * 2000-09-11 2006-01-17 Cisco Technology, Inc. STP root guard
US20060023630A1 (en) * 2003-05-19 2006-02-02 Hiroyuki Akiyama SDH transmission apparatus and signal transmission method
US7016351B1 (en) * 2000-02-29 2006-03-21 Cisco Technology, Inc. Small group multicast in a computer network
US20060159011A1 (en) * 2005-01-14 2006-07-20 Mitesh Dalal Detecting unavailable network connections
US20060206602A1 (en) * 2005-03-14 2006-09-14 International Business Machines Corporation Network switch link failover in a redundant switch configuration
US20060227704A1 (en) * 2005-04-12 2006-10-12 Fujitsu Limited Special marker message for link aggregation marker protocol
US20060256712A1 (en) * 2003-02-21 2006-11-16 Nippon Telegraph And Telephone Corporation Device and method for correcting a path trouble in a communication network
US20070041313A1 (en) * 2005-08-17 2007-02-22 Honeywell International, Inc Apparatus and methods for managing nodes on a fault tolerant network
US20070047436A1 (en) * 2005-08-24 2007-03-01 Masaya Arai Network relay device and control method
US20070047578A1 (en) * 2005-08-29 2007-03-01 Fujitsu Limited Bandwidth control method and transmission equipment
US7197660B1 (en) * 2002-06-26 2007-03-27 Juniper Networks, Inc. High availability network security systems
US20070070886A1 (en) * 2005-09-29 2007-03-29 Seth Zirin Modifying an endpoint node connection associated with a destination
US20070076727A1 (en) * 2005-09-30 2007-04-05 Tekelec Adaptive redundancy protection scheme
US20070183347A1 (en) * 2004-08-29 2007-08-09 Huawei Technologies Co., Ltd. Method for implementing dual-homing
US20070280103A1 (en) * 2006-06-02 2007-12-06 Cisco Technology, Inc. Transparent automatic protection switching for a chassis deployment
US20080016402A1 (en) * 2006-07-11 2008-01-17 Corrigent Systems Ltd. Connectivity fault management (CFM) in networks with link aggregation group connections
US20080181196A1 (en) * 2007-01-31 2008-07-31 Alcatel Lucent Link aggregation across multiple chassis
US7453900B2 (en) * 2003-03-05 2008-11-18 Cisco Technology, Inc. System and method for monitoring noise associated with a communication link
US7480283B1 (en) * 2002-03-26 2009-01-20 Nortel Networks Limited Virtual trunking over packet networks
US7518986B1 (en) * 2005-11-16 2009-04-14 Juniper Networks, Inc. Push-based hierarchical state propagation within a multi-chassis network device
US20090096592A1 (en) * 2007-10-12 2009-04-16 Alpha Networks Inc. Power line adapter and method of controlling power line adapter operated in power-saving mode
US20090109998A1 (en) * 2007-10-29 2009-04-30 Alcatel Lucent Facilitating self configuring link aggregation using link aggregation control protocol
US7580349B1 (en) * 2001-11-02 2009-08-25 Nortel Networks Limited Content-aware dynamic network resource allocation
US7593319B1 (en) * 2002-10-15 2009-09-22 Garrettcom, Inc. LAN switch with rapid fault recovery
US7616561B1 (en) * 2005-01-19 2009-11-10 Juniper Networks, Inc. Systems and methods for routing data in a communications network
US20100020680A1 (en) * 2008-07-28 2010-01-28 Salam Samer M Multi-chassis ethernet link aggregation

Patent Citations (62)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5835696A (en) * 1995-11-22 1998-11-10 Lucent Technologies Inc. Data router backup feature
US6542934B1 (en) * 1997-05-30 2003-04-01 International Business Machines Corporation Non-disruptively rerouting network communications from a secondary network path to a primary path
US6535491B2 (en) * 1997-12-24 2003-03-18 Cisco Technology, Inc. Method and apparatus for rapidly reconfiguring computer networks using a spanning tree algorithm
US6032194A (en) * 1997-12-24 2000-02-29 Cisco Technology, Inc. Method and apparatus for rapidly reconfiguring computer networks
US20030016624A1 (en) * 1998-05-04 2003-01-23 Bare Ballard C. Path recovery on failure in load balancing switch protocols
US6628661B1 (en) * 1998-08-27 2003-09-30 Intel Corporation Spanning tree recovery in computer networks
US6330229B1 (en) * 1998-11-09 2001-12-11 3Com Corporation Spanning tree with rapid forwarding database updates
US20030179700A1 (en) * 1999-01-15 2003-09-25 Saleh Ali Najib Method for restoring a virtual path in an optical network using 1‘protection
US6697339B1 (en) * 1999-03-04 2004-02-24 3Com Corporation High availability spanning tree with rapid reconfiguration with alternate port selection
US6801506B1 (en) * 1999-03-31 2004-10-05 Cisco Technology, Inc. Method and apparatus for providing fast spanning tree re-starts
US6721275B1 (en) * 1999-05-03 2004-04-13 Hewlett-Packard Development Company, L.P. Bridged network stations location revision
US6513129B1 (en) * 1999-06-30 2003-01-28 Objective Systems Integrators, Inc. System and method for managing faults using a gateway
US6392990B1 (en) * 1999-07-23 2002-05-21 Glenayre Electronics, Inc. Method for implementing interface redundancy in a computer network
US20040047336A1 (en) * 1999-08-23 2004-03-11 Avaya Communication Israel Ltd. Modular bridging-device
US6678241B1 (en) * 1999-11-30 2004-01-13 Cisc Technology, Inc. Fast convergence with topology switching
US20040133693A1 (en) * 2000-01-28 2004-07-08 3Com Corporation Multi-point link aggregation spoofing
US7016351B1 (en) * 2000-02-29 2006-03-21 Cisco Technology, Inc. Small group multicast in a computer network
US6891808B2 (en) * 2000-03-10 2005-05-10 Anritsu Corporation Spanning tree bridge and route change method using the same
US6442133B2 (en) * 2000-05-24 2002-08-27 Nec Corporation Switchboard having a dual switching system
US6987740B1 (en) * 2000-09-11 2006-01-17 Cisco Technology, Inc. STP root guard
US20020112072A1 (en) * 2001-02-12 2002-08-15 Maple Optical Systems, Inc. System and method for fast-rerouting of data in a data communication network
US20020152320A1 (en) * 2001-02-14 2002-10-17 Lau Pui Lun System and method for rapidly switching between redundant networks
US20020167895A1 (en) * 2001-05-08 2002-11-14 Jack Zhu Method for restoring diversely routed circuits
US20040105390A1 (en) * 2001-05-28 2004-06-03 Nokia Corporation Method and system for implementing a fast recovery process in a local area network
US20020184387A1 (en) * 2001-05-31 2002-12-05 The Furukawa Electric Co., Ltd. Method for connecting between networks, virtual router, and system for connecting between networks by using this virtual router
US6956824B2 (en) * 2001-06-14 2005-10-18 Tropic Networks Inc. Extension of link aggregation protocols over the network
US20030021226A1 (en) * 2001-07-24 2003-01-30 Gal Mor Interconnect and gateway protection in bidirectional ring networks
US20030097470A1 (en) * 2001-09-10 2003-05-22 Roger Lapuh System, device, and method for improving communication network reliability using trunk splitting
US7580349B1 (en) * 2001-11-02 2009-08-25 Nortel Networks Limited Content-aware dynamic network resource allocation
US20030137932A1 (en) * 2002-01-23 2003-07-24 Nec Corporation Path fault recovery method, switching-back method after recovery from fault, and node using the same
US20030158964A1 (en) * 2002-02-15 2003-08-21 Naofumi Kobayashi Upper and lower relays and network system
US20030161260A1 (en) * 2002-02-25 2003-08-28 Sundara Murugan Method and apparatus for implementing automatic protection switching functionality in a distributed processor data router
US7480283B1 (en) * 2002-03-26 2009-01-20 Nortel Networks Limited Virtual trunking over packet networks
US20030185225A1 (en) * 2002-03-29 2003-10-02 Wirth Brian Michael Switch and a switching apparatus for a communication network
US7197660B1 (en) * 2002-06-26 2007-03-27 Juniper Networks, Inc. High availability network security systems
US20040001449A1 (en) * 2002-06-28 2004-01-01 Rostron Andy E. System and method for supporting automatic protection switching between multiple node pairs using common agent architecture
US7593319B1 (en) * 2002-10-15 2009-09-22 Garrettcom, Inc. LAN switch with rapid fault recovery
US20040098501A1 (en) * 2002-10-29 2004-05-20 Finn Norman W. Multi-bridge lan aggregation
US20060256712A1 (en) * 2003-02-21 2006-11-16 Nippon Telegraph And Telephone Corporation Device and method for correcting a path trouble in a communication network
US7453900B2 (en) * 2003-03-05 2008-11-18 Cisco Technology, Inc. System and method for monitoring noise associated with a communication link
US20060023630A1 (en) * 2003-05-19 2006-02-02 Hiroyuki Akiyama SDH transmission apparatus and signal transmission method
US20040264364A1 (en) * 2003-06-27 2004-12-30 Nec Corporation Network system for building redundancy within groups
US20050063395A1 (en) * 2003-09-18 2005-03-24 Cisco Technology, Inc. Virtual network device
US20050108401A1 (en) * 2003-11-13 2005-05-19 Gonda Rumi S. Method for supporting SDH/SONET OAMP on Ethernet
US20050111350A1 (en) * 2003-11-20 2005-05-26 Shinya Kano Method and apparatus for designing backup communication path, and computer product
US20070183347A1 (en) * 2004-08-29 2007-08-09 Huawei Technologies Co., Ltd. Method for implementing dual-homing
US20060159011A1 (en) * 2005-01-14 2006-07-20 Mitesh Dalal Detecting unavailable network connections
US7616561B1 (en) * 2005-01-19 2009-11-10 Juniper Networks, Inc. Systems and methods for routing data in a communications network
US20060206602A1 (en) * 2005-03-14 2006-09-14 International Business Machines Corporation Network switch link failover in a redundant switch configuration
US20060227704A1 (en) * 2005-04-12 2006-10-12 Fujitsu Limited Special marker message for link aggregation marker protocol
US20070041313A1 (en) * 2005-08-17 2007-02-22 Honeywell International, Inc Apparatus and methods for managing nodes on a fault tolerant network
US20070047436A1 (en) * 2005-08-24 2007-03-01 Masaya Arai Network relay device and control method
US20070047578A1 (en) * 2005-08-29 2007-03-01 Fujitsu Limited Bandwidth control method and transmission equipment
US20070070886A1 (en) * 2005-09-29 2007-03-29 Seth Zirin Modifying an endpoint node connection associated with a destination
US20070076727A1 (en) * 2005-09-30 2007-04-05 Tekelec Adaptive redundancy protection scheme
US7518986B1 (en) * 2005-11-16 2009-04-14 Juniper Networks, Inc. Push-based hierarchical state propagation within a multi-chassis network device
US20070280103A1 (en) * 2006-06-02 2007-12-06 Cisco Technology, Inc. Transparent automatic protection switching for a chassis deployment
US20080016402A1 (en) * 2006-07-11 2008-01-17 Corrigent Systems Ltd. Connectivity fault management (CFM) in networks with link aggregation group connections
US20080181196A1 (en) * 2007-01-31 2008-07-31 Alcatel Lucent Link aggregation across multiple chassis
US20090096592A1 (en) * 2007-10-12 2009-04-16 Alpha Networks Inc. Power line adapter and method of controlling power line adapter operated in power-saving mode
US20090109998A1 (en) * 2007-10-29 2009-04-30 Alcatel Lucent Facilitating self configuring link aggregation using link aggregation control protocol
US20100020680A1 (en) * 2008-07-28 2010-01-28 Salam Samer M Multi-chassis ethernet link aggregation

Cited By (120)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070277828A1 (en) * 2006-06-05 2007-12-06 Ho Peter C F Flexible connector
US9094257B2 (en) 2006-06-30 2015-07-28 Centurylink Intellectual Property Llc System and method for selecting a content delivery network
US9549004B2 (en) 2006-06-30 2017-01-17 Centurylink Intellectual Property Llc System and method for re-routing calls
US7948909B2 (en) 2006-06-30 2011-05-24 Embarq Holdings Company, Llc System and method for resetting counters counting network performance information at network communications devices on a packet network
US9749399B2 (en) 2006-06-30 2017-08-29 Centurylink Intellectual Property Llc System and method for selecting a content delivery network
US8000318B2 (en) 2006-06-30 2011-08-16 Embarq Holdings Company, Llc System and method for call routing based on transmission performance of a packet network
US9154634B2 (en) 2006-06-30 2015-10-06 Centurylink Intellectual Property Llc System and method for managing network communications
US9118583B2 (en) 2006-06-30 2015-08-25 Centurylink Intellectual Property Llc System and method for re-routing calls
US9054915B2 (en) 2006-06-30 2015-06-09 Centurylink Intellectual Property Llc System and method for adjusting CODEC speed in a transmission path during call set-up due to reduced transmission performance
US9838440B2 (en) 2006-06-30 2017-12-05 Centurylink Intellectual Property Llc Managing voice over internet protocol (VoIP) communications
US10230788B2 (en) 2006-06-30 2019-03-12 Centurylink Intellectual Property Llc System and method for selecting a content delivery network
US8184549B2 (en) 2006-06-30 2012-05-22 Embarq Holdings Company, LLP System and method for selecting network egress
US8976665B2 (en) 2006-06-30 2015-03-10 Centurylink Intellectual Property Llc System and method for re-routing calls
US8717911B2 (en) 2006-06-30 2014-05-06 Centurylink Intellectual Property Llc System and method for collecting network performance information
US10560494B2 (en) 2006-06-30 2020-02-11 Centurylink Intellectual Property Llc Managing voice over internet protocol (VoIP) communications
US8570872B2 (en) 2006-06-30 2013-10-29 Centurylink Intellectual Property Llc System and method for selecting network ingress and egress
US8488447B2 (en) 2006-06-30 2013-07-16 Centurylink Intellectual Property Llc System and method for adjusting code speed in a transmission path during call set-up due to reduced transmission performance
US8477614B2 (en) 2006-06-30 2013-07-02 Centurylink Intellectual Property Llc System and method for routing calls if potential call paths are impaired or congested
US8670313B2 (en) 2006-08-22 2014-03-11 Centurylink Intellectual Property Llc System and method for adjusting the window size of a TCP packet through network elements
US8576722B2 (en) 2006-08-22 2013-11-05 Centurylink Intellectual Property Llc System and method for modifying connectivity fault management packets
US8125897B2 (en) 2006-08-22 2012-02-28 Embarq Holdings Company Lp System and method for monitoring and optimizing network performance with user datagram protocol network performance information packets
US8130793B2 (en) 2006-08-22 2012-03-06 Embarq Holdings Company, Llc System and method for enabling reciprocal billing for different types of communications over a packet network
US8144586B2 (en) 2006-08-22 2012-03-27 Embarq Holdings Company, Llc System and method for controlling network bandwidth with a connection admission control engine
US8144587B2 (en) 2006-08-22 2012-03-27 Embarq Holdings Company, Llc System and method for load balancing network resources using a connection admission control engine
US8107366B2 (en) 2006-08-22 2012-01-31 Embarq Holdings Company, LP System and method for using centralized network performance tables to manage network communications
US9479341B2 (en) 2006-08-22 2016-10-25 Centurylink Intellectual Property Llc System and method for initiating diagnostics on a packet network node
US9253661B2 (en) 2006-08-22 2016-02-02 Centurylink Intellectual Property Llc System and method for modifying connectivity fault management packets
US8194555B2 (en) 2006-08-22 2012-06-05 Embarq Holdings Company, Llc System and method for using distributed network performance information tables to manage network communications
US8199653B2 (en) 2006-08-22 2012-06-12 Embarq Holdings Company, Llc System and method for communicating network performance information over a packet network
US8213366B2 (en) 2006-08-22 2012-07-03 Embarq Holdings Company, Llc System and method for monitoring and optimizing network performance to a wireless device
US8224255B2 (en) 2006-08-22 2012-07-17 Embarq Holdings Company, Llc System and method for managing radio frequency windows
US8223654B2 (en) 2006-08-22 2012-07-17 Embarq Holdings Company, Llc Application-specific integrated circuit for monitoring and optimizing interlayer network performance
US8228791B2 (en) 2006-08-22 2012-07-24 Embarq Holdings Company, Llc System and method for routing communications between packet networks based on intercarrier agreements
US8238253B2 (en) * 2006-08-22 2012-08-07 Embarq Holdings Company, Llc System and method for monitoring interlayer devices and optimizing network performance
US20080049631A1 (en) * 2006-08-22 2008-02-28 Morrill Robert J System and method for monitoring interlayer devices and optimizing network performance
US8274905B2 (en) 2006-08-22 2012-09-25 Embarq Holdings Company, Llc System and method for displaying a graph representative of network performance over a time period
US9602265B2 (en) 2006-08-22 2017-03-21 Centurylink Intellectual Property Llc System and method for handling communications requests
US20120269089A1 (en) * 2006-08-22 2012-10-25 Morrill Robert J System and method for monitoring interlayer devices and optimizing network performance
US10469385B2 (en) 2006-08-22 2019-11-05 Centurylink Intellectual Property Llc System and method for improving network performance using a connection admission control engine
US8307065B2 (en) 2006-08-22 2012-11-06 Centurylink Intellectual Property Llc System and method for remotely controlling network operators
US9241277B2 (en) 2006-08-22 2016-01-19 Centurylink Intellectual Property Llc System and method for monitoring and optimizing network performance to a wireless device
US8358580B2 (en) 2006-08-22 2013-01-22 Centurylink Intellectual Property Llc System and method for adjusting the window size of a TCP packet through network elements
US8374090B2 (en) 2006-08-22 2013-02-12 Centurylink Intellectual Property Llc System and method for routing data on a packet network
US8407765B2 (en) 2006-08-22 2013-03-26 Centurylink Intellectual Property Llc System and method for restricting access to network performance information tables
US8472326B2 (en) * 2006-08-22 2013-06-25 Centurylink Intellectual Property Llc System and method for monitoring interlayer devices and optimizing network performance
US8102770B2 (en) 2006-08-22 2012-01-24 Embarq Holdings Company, LP System and method for monitoring and optimizing network performance with vector performance tables and engines
US8488495B2 (en) 2006-08-22 2013-07-16 Centurylink Intellectual Property Llc System and method for routing communications between packet networks based on real time pricing
US8098579B2 (en) 2006-08-22 2012-01-17 Embarq Holdings Company, LP System and method for adjusting the window size of a TCP packet through remote network elements
US8509082B2 (en) 2006-08-22 2013-08-13 Centurylink Intellectual Property Llc System and method for load balancing network resources using a connection admission control engine
US8520603B2 (en) 2006-08-22 2013-08-27 Centurylink Intellectual Property Llc System and method for monitoring and optimizing network performance to a wireless device
US10298476B2 (en) 2006-08-22 2019-05-21 Centurylink Intellectual Property Llc System and method for tracking application resource usage
US8531954B2 (en) 2006-08-22 2013-09-10 Centurylink Intellectual Property Llc System and method for handling reservation requests with a connection admission control engine
US8537695B2 (en) 2006-08-22 2013-09-17 Centurylink Intellectual Property Llc System and method for establishing a call being received by a trunk on a packet network
US8549405B2 (en) 2006-08-22 2013-10-01 Centurylink Intellectual Property Llc System and method for displaying a graphical representation of a network to identify nodes and node segments on the network that are not operating normally
US10075351B2 (en) 2006-08-22 2018-09-11 Centurylink Intellectual Property Llc System and method for improving network performance
US9241271B2 (en) 2006-08-22 2016-01-19 Centurylink Intellectual Property Llc System and method for restricting access to network performance information
US8619820B2 (en) 2006-08-22 2013-12-31 Centurylink Intellectual Property Llc System and method for enabling communications over a number of packet networks
US8619596B2 (en) 2006-08-22 2013-12-31 Centurylink Intellectual Property Llc System and method for using centralized network performance tables to manage network communications
US8619600B2 (en) 2006-08-22 2013-12-31 Centurylink Intellectual Property Llc System and method for establishing calls over a call path having best path metrics
US8064391B2 (en) 2006-08-22 2011-11-22 Embarq Holdings Company, Llc System and method for monitoring and optimizing network performance to a wireless device
US8687614B2 (en) 2006-08-22 2014-04-01 Centurylink Intellectual Property Llc System and method for adjusting radio frequency parameters
US8040811B2 (en) 2006-08-22 2011-10-18 Embarq Holdings Company, Llc System and method for collecting and managing network performance information
US8743703B2 (en) 2006-08-22 2014-06-03 Centurylink Intellectual Property Llc System and method for tracking application resource usage
US8743700B2 (en) 2006-08-22 2014-06-03 Centurylink Intellectual Property Llc System and method for provisioning resources of a packet network based on collected network performance information
US8750158B2 (en) 2006-08-22 2014-06-10 Centurylink Intellectual Property Llc System and method for differentiated billing
US8811160B2 (en) 2006-08-22 2014-08-19 Centurylink Intellectual Property Llc System and method for routing data on a packet network
US9992348B2 (en) 2006-08-22 2018-06-05 Century Link Intellectual Property LLC System and method for establishing a call on a packet network
US9929923B2 (en) 2006-08-22 2018-03-27 Centurylink Intellectual Property Llc System and method for provisioning resources of a packet network based on collected network performance information
US9832090B2 (en) 2006-08-22 2017-11-28 Centurylink Intellectual Property Llc System, method for compiling network performancing information for communications with customer premise equipment
US9813320B2 (en) 2006-08-22 2017-11-07 Centurylink Intellectual Property Llc System and method for generating a graphical user interface representative of network performance
US9806972B2 (en) 2006-08-22 2017-10-31 Centurylink Intellectual Property Llc System and method for monitoring and altering performance of a packet network
US8015294B2 (en) 2006-08-22 2011-09-06 Embarq Holdings Company, LP Pin-hole firewall for communicating data packets on a packet network
US9014204B2 (en) 2006-08-22 2015-04-21 Centurylink Intellectual Property Llc System and method for managing network communications
US9042370B2 (en) 2006-08-22 2015-05-26 Centurylink Intellectual Property Llc System and method for establishing calls over a call path having best path metrics
US9712445B2 (en) 2006-08-22 2017-07-18 Centurylink Intellectual Property Llc System and method for routing data on a packet network
US9054986B2 (en) 2006-08-22 2015-06-09 Centurylink Intellectual Property Llc System and method for enabling communications over a number of packet networks
US9660917B2 (en) 2006-08-22 2017-05-23 Centurylink Intellectual Property Llc System and method for remotely controlling network operators
US9094261B2 (en) 2006-08-22 2015-07-28 Centurylink Intellectual Property Llc System and method for establishing a call being received by a trunk on a packet network
US7940735B2 (en) 2006-08-22 2011-05-10 Embarq Holdings Company, Llc System and method for selecting an access point
US9112734B2 (en) 2006-08-22 2015-08-18 Centurylink Intellectual Property Llc System and method for generating a graphical user interface representative of network performance
US7843831B2 (en) 2006-08-22 2010-11-30 Embarq Holdings Company Llc System and method for routing data on a packet network
US9240906B2 (en) 2006-08-22 2016-01-19 Centurylink Intellectual Property Llc System and method for monitoring and altering performance of a packet network
US9661514B2 (en) 2006-08-22 2017-05-23 Centurylink Intellectual Property Llc System and method for adjusting communication parameters
US9621361B2 (en) 2006-08-22 2017-04-11 Centurylink Intellectual Property Llc Pin-hole firewall for communicating data packets on a packet network
US9225609B2 (en) 2006-08-22 2015-12-29 Centurylink Intellectual Property Llc System and method for remotely controlling network operators
US9225646B2 (en) 2006-08-22 2015-12-29 Centurylink Intellectual Property Llc System and method for improving network performance using a connection admission control engine
US8289965B2 (en) 2006-10-19 2012-10-16 Embarq Holdings Company, Llc System and method for establishing a communications session with an end-user based on the state of a network connection
US8194643B2 (en) 2006-10-19 2012-06-05 Embarq Holdings Company, Llc System and method for monitoring the connection of an end-user to a remote network
US8189468B2 (en) 2006-10-25 2012-05-29 Embarq Holdings, Company, LLC System and method for regulating messages between networks
US9521150B2 (en) 2006-10-25 2016-12-13 Centurylink Intellectual Property Llc System and method for automatically regulating messages between networks
US20100296390A1 (en) * 2006-11-27 2010-11-25 Honeywell International Inc. Fault tolerance in wireless networks
US8340008B2 (en) * 2006-11-27 2012-12-25 Honeywell International Inc. Fault tolerance in wireless networks
US8111692B2 (en) 2007-05-31 2012-02-07 Embarq Holdings Company Llc System and method for modifying network traffic
EP2243255A1 (en) * 2008-02-01 2010-10-27 Allied Telesis Holdings K.K. Method and system for dynamic link failover management
EP2243255A4 (en) * 2008-02-01 2012-08-29 Allied Telesis Holdings Kk Method and system for dynamic link failover management
WO2009097615A1 (en) * 2008-02-01 2009-08-06 Allied Telesis Holdings Kabushiki Kaisha Method and system for dynamic link failover management
US20090249115A1 (en) * 2008-02-01 2009-10-01 Allied Telesis Holdings Kabushiki Kaisha Method and system for dynamic link failover management
US8068425B2 (en) 2008-04-09 2011-11-29 Embarq Holdings Company, Llc System and method for using network performance information to determine improved measures of path states
US8879391B2 (en) 2008-04-09 2014-11-04 Centurylink Intellectual Property Llc System and method for using network derivations to determine path states
US20100020680A1 (en) * 2008-07-28 2010-01-28 Salam Samer M Multi-chassis ethernet link aggregation
US8300523B2 (en) 2008-07-28 2012-10-30 Cisco Technology, Inc. Multi-chasis ethernet link aggregation
US20110211585A1 (en) * 2010-02-26 2011-09-01 Alaxala Networks Corporation Packet relay device
US8971337B2 (en) * 2010-02-26 2015-03-03 Alaxala Networks Corporation Packet relay device
US20130232277A1 (en) * 2011-08-31 2013-09-05 Metaswitch Networks Ltd. Transmitting and Forwarding Data
US9246796B2 (en) * 2011-08-31 2016-01-26 Metaswitch Networks Ltd Transmitting and forwarding data
US9197547B2 (en) * 2012-06-01 2015-11-24 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of MoFRR with dataplane notifications
US9628285B2 (en) * 2012-06-01 2017-04-18 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of MoFRR with dataplane notifications
US9736061B2 (en) 2012-06-01 2017-08-15 Telefonaktiebolaget L M Ericsson (Publ) Enhancements to PIM fast re-route with upstream activation packets
US20150172070A1 (en) * 2012-06-01 2015-06-18 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of mofrr with dataplane notifications
US8824276B2 (en) * 2012-06-01 2014-09-02 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of MOFRR with dataplane notifications
US8913482B2 (en) 2012-06-01 2014-12-16 Telefonaktiebolaget L M Ericsson (Publ) Enhancements to PIM fast re-route with upstream activation packets
US20140355422A1 (en) * 2012-06-01 2014-12-04 Telefonaktiebolaget L M Ericsson (Publ) Increasing failure coverage of mofrr with dataplane notifications
US20150295861A1 (en) * 2013-01-06 2015-10-15 Hangzhou H3C Technologies Co., Ltd. Link problem handling
US9559989B2 (en) * 2013-01-06 2017-01-31 Hewlett Packard Enterprise Development Lp Link problem handling
US20190028299A1 (en) * 2016-03-18 2019-01-24 Hewlett Packard Enterprise Development Lp Responses to loops in networks having a ring topology
WO2017160306A1 (en) * 2016-03-18 2017-09-21 Hewlett Packard Enterprise Development Lp Responses to loops in networks having a ring topology
US10812289B2 (en) 2016-03-18 2020-10-20 Hewlett Packard Enterprise Development Lp Responses to loops in networks having a ring topology
US11551684B1 (en) * 2017-12-08 2023-01-10 Amazon Technologies, Inc. State detection and responses for electronic devices
US11012369B2 (en) * 2019-07-05 2021-05-18 Dell Products L.P. Aggregated switch path optimization system
CN114679348A (en) * 2020-12-24 2022-06-28 华为技术有限公司 Fault processing method and related product

Also Published As

Publication number Publication date
US8886831B2 (en) 2014-11-11

Similar Documents

Publication Publication Date Title
US8886831B2 (en) System and methodology for fast link failover based on remote upstream failures
US8117337B2 (en) Method and device for implementing link pass through in point-to-multipoint network
US8804534B2 (en) Interworking between MPLS/IP and Ethernet OAM mechanisms
EP2087712B1 (en) Method and apparatus for computing alternate multicast/broadcast paths in a routed network
US9203732B2 (en) Recovery of traffic in a connection-oriented network
EP1919138B1 (en) A method for implementing backup of the uplink
JP4884184B2 (en) Communication device and program for link aggregation
EP1895724A1 (en) A method for implementing active/standby gateway device in the network and a system thereof
EP1958364B1 (en) Vpls remote failure indication
US20080112333A1 (en) Communicating an operational state of a transport service
US9385944B2 (en) Communication system, path switching method and communication device
KR102088298B1 (en) Method and appratus for protection switching in packet transport system
JP2007088949A (en) Information processing apparatus, communication load diffusing method and communication load diffusion program
JP2008078893A (en) Redundant method for network and medium switching equipment
JP2007525895A (en) Recovery mechanism for network topology
EP3029883B1 (en) Network protection method and apparatus, next-ring node, and system
EP2254289B1 (en) Method, device, and system for establishing label switching path in fast rerouting switching
CN102857316B (en) A kind of method and system realizing source looped network protection
US7174389B2 (en) Tandem node system and a method therefor
CN102882779A (en) VRRP (Virtual Router Redundancy Protocol) advertisement link protection method and system
WO2011011934A1 (en) Method and apparatus for ethernet tunnel segmentation protection
CN113037622B (en) System and method for preventing BFD from vibrating
JP5612560B2 (en) Network relay device and network relay system
JP2003338831A (en) Switching method for mpls network
JP4035820B2 (en) Line switching apparatus, data transmission system, and line switching method

Legal Events

Date Code Title Description
AS Assignment

Owner name: CISCO TECHNOLOGY, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TIRUMALAI, GOKUL S.;MITCHELL, ALBERT HYDE;REEL/FRAME:017762/0523

Effective date: 20060405

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20221111