US20070237085A1 - System and methodology for fast link failover based on remote upstream failures - Google Patents
System and methodology for fast link failover based on remote upstream failures Download PDFInfo
- Publication number
- US20070237085A1 US20070237085A1 US11/398,519 US39851906A US2007237085A1 US 20070237085 A1 US20070237085 A1 US 20070237085A1 US 39851906 A US39851906 A US 39851906A US 2007237085 A1 US2007237085 A1 US 2007237085A1
- Authority
- US
- United States
- Prior art keywords
- network
- network device
- status
- notification
- change
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04Q—SELECTING
- H04Q3/00—Selecting arrangements
- H04Q3/0016—Arrangements providing connection between exchanges
- H04Q3/0062—Provisions for network management
- H04Q3/0087—Network testing or monitoring arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L45/00—Routing or path finding of packets in data switching networks
- H04L45/22—Alternate routing
Definitions
- the present invention generally relates to management of switched computer networks and, more particularly, to propagating detected changes in upstream connectivity to downstream devices.
- a computer network generally includes a number of devices, including switches, routers and hubs, connected so as to allow communication among the devices.
- the devices within a network may be generally categorized into two classes: end stations connected along the edge of a network and network devices. End stations typically include devices that primarily supply or consume information, such as workstations, desktop PCs, printers, servers, hosts, fax machines, and other such devices.
- Network devices typically include devices that primarily transfer information between other devices (e.g., between end stations and other network devices), such as gateways, switches, routers, and other such devices. In other words, network devices generally provide connectivity between end stations and the network core.
- FIG. 1 illustrates an exemplary network 100 in which a plurality of end stations 110 are connected to a network core 120 via multiple layers of switches 130 .
- the network may be modeled in accordance with the Open System Interconnection (OSI) model, which defines a networking framework for implementing protocols in multiple (seven) layers.
- the switches 130 may represent Data Link or “Layer 2 ” layer switching, that encodes and decodes data packets into bits. Layer 2 switches implement the network transmission protocol and handle errors in the physical layer (Ethernet, optics, etc.), flow control and frame synchronization.
- switches 130 include aggregation switches 132 (SW 1 and SW 2 ), and access switches 134 (SW 3 and SW 4 ).
- the end stations 110 may be file servers, data servers, workstations, personal computers (PCs), and the like.
- end stations 110 may each have multiple network interface cards NICs 112 (or other type network interface/adapter), allowing the end stations 110 to connect to the network through primary links/connections (Lp) and secondary links (Ls).
- NICs 112 or other type network interface/adapter
- the end stations may be configured to connect to the network through a primary path (via SW 3 and SW 1 ) by default and switch (or “failover”) to a secondary path (via SW 4 and SW 2 ) if a connectivity problem is detected in the primary path, such as a loss of connectivity to a link L C between switch SW 1 and the network core 120 .
- end stations do not become aware of a change in upstream connectivity until they discover it on their own. For example, end stations may only discover a loss of upstream connectivity after experiencing consistent protocol packet timeouts. As a result, after a loss in network connectivity on the primary path, end stations may not immediately switch from their primary path to their alternative path despite a loss in network connectivity on the primary path, which may degrade network performance.
- Embodiments of the present invention provide techniques for propagating changes in upstream connectivity downstream to connected end stations.
- One embodiment provides a method for managing a network path between at least one end station and a network.
- the method generally includes detecting, by a first switching device, a change in status of a first upstream communication link between the first switching device and the network, sending, by the first switching device, notification of the change in status to a second switching device in communication with the end station via a first downstream communication link, and, in response to receiving the notification, disabling by the second switch, the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
- a network switching system generally including at least a first switching device and at least a second switching device.
- the first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core.
- the second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable a first downstream communication link between the second switching device and an end station.
- their may be one or more intermediate switching devices, between the first and second switching devices, that receive the notification of the change in status from the first switching device and propagate the notification (possibly through other intermediate switches) to the second switching device that ultimately disables the first downstream communication link with the end station.
- Another embodiment provides a system generally including a plurality of end stations, at least a first switching device, and at least a second switching device.
- the first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core.
- the second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third switching device.
- an aggregation switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations.
- the operations generally include detecting a change in status of a first upstream communication link between the first switching device and a network and sending notification of the change in status to a second switching device in communication with an end station via a first downstream communication link.
- an access switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations.
- the operations generally include in response to receiving, from an upstream switching device, notification of a detected change in status between the first switching device and a network, disabling at least one downstream communication link between the access switching device and at least one end station.
- FIG. 1 illustrates an exemplary multilayer switched network.
- FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention.
- FIGS. 3A-3D illustrate propagating, and responding to, changes in upstream connectivity, in accordance with one embodiment of the present invention.
- FIG. 4 illustrates the routing of traffic through links maintained between access and aggregation devices, in accordance with one embodiment of the present invention.
- FIG. 5 is a flow diagram of exemplary operations for propagating changes in connectivity, in accordance with embodiments of the present invention.
- Embodiments of the present invention generally provide methods and apparatus for propagating changes in upstream connectivity downstream.
- an upstream switch connected to a network core may detect a change in connectivity and, in response, send a message indicating the change to a downstream switch connected to an end station.
- the downstream switch may disable a link with end station. Disabling the link may cause the end station to promptly switch (fail over) to a different link, in an effort to establish an alternate path to the network.
- network traffic routed through that switch need not be re-routed and the link can continue to carry traffic destined for other networks.
- end station generally refers to any device capable of a network connection via one or more network interfaces and need not necessarily reside at an edge of a network. Examples of end stations include a server, PC, fax machine, printer, or the like.
- switch or switching device generally refers to any type device that switches network traffic, such as a Layer 2 switch (per the OSI model) that switches Ethernet traffic.
- access device generally refers to a network element that is connected to one or more end stations.
- aggregation device generally refers to a network element that connects to one or more access devices.
- upstream generally refers to a path (or device along a path) in the direction from an end station towards the core of the network.
- downstream generally refers to a path (or device along a path) in the direction from the core of the network to the edge (e.g., towards an end station).
- link or connection may be used herein interchangeably and generally refer to any connection between devices, such as a Layer 2 network connection between two devices.
- failure generally refers to a loss of connectivity to a link from a device, regardless of the particular cause, including if the link is intentionally disabled by another device.
- Embodiments of the present invention allow changes in upstream connectivity to be communicated downstream, thereby allowing an end station to quickly make decisions regarding which paths to use to the network.
- the techniques described herein may be implemented, for example, as software methods (e.g., executable instructions residing in memory) running on network devices (e.g., aggregation devices and/or access devices) that are directly or indirectly connected to a network interface of an end station.
- the indirectly connected switches may track the link status on upstream interfaces and communicate any change in connectivity (loss/restoration) to a set of downstream interfaces, for example, by sending a message.
- Such a message may be intercepted by downstream network devices, such as access switches directly connected to network interfaces of the end stations.
- these switches may take appropriate action based on the message communicated, for example, disabling corresponding active links if the message indicates loss of connectivity. This allows conventional end stations with redundant network interfaces to detect that their primary connection is down and to automatically switch over to their secondary network interfaces.
- FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention.
- the operations 200 may be collectively performed by a combination of switches, such as aggregation switches connected to a network core and access switches connected to an end station.
- switches such as aggregation switches connected to a network core and access switches connected to an end station.
- FIGS. 3A-3D illustrate the status of different links in the network as connectivity changes therein.
- FIGS. 3A-3D To facilitate understanding, only a limited portion of the network shown in FIG. 1 is illustrated in FIGS. 3A-3D . However, those skilled in the art will recognize that the concepts described herein may be advantageously scaled to accommodate any size and type of switching network having any number of multiple switching layers (e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements), with potentially hundreds of end stations.
- any number of multiple switching layers e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements
- the operations begin, at step 202 , by detecting a change in upstream connectivity between a first switch and a network core.
- the exemplary network shown includes a layer of aggregation switches 332 SW 1 and SW 2 and a layer of access switches 334 SW 3 and SW 4 .
- a primary path between end stations 110 and the network core 120 may be established through SW 3 and SW 1 via first network interface cards 112 (NIC 1 ).
- NIC 1 first network interface cards
- a loss of connectivity between SW 1 and the network core may occur for any number of reasons (e.g., device failure, such as failure of a laser in a fiber optic connection, failure in the physical connection/cable, or loss of power).
- switches SW 3 may have downstream connections to both end stations and other switches, not just to one another.
- the change in connectivity is communicated to a second switch located downstream between the first switch and an end station.
- the first switch SW 1 may communicate the change in connectivity (a loss in this example), to second switch SW 3 via a change in connectivity message.
- the contents of the message may vary, as may the recipients.
- a user e.g., network administrator
- a downstream connection between the second switch and the end station is modified in response to the change in connectivity.
- SW 3 may decipher the message from SW 1 and, in response, disable (bring down) the link between the primary NICs of the end stations 110 .
- Exactly how the link is disabled may depend on the corresponding type of physical medium. For example, for an Ethernet connection this may entail simply ceasing to transmit a signal (e.g., by disabling a laser for a fiber optic Ethernet connection or bringing down the voltage level for a copper Ethernet connection).
- end stations equipped with redundant network interfaces may be configured to react to such a loss in connectivity to a primary link by automatically switching (or “failing over”) to a secondary interface in order to establish an alternate path to the network.
- end stations 110 may automatically switch over to secondary NICs to establish an alternate path to the core network via SW 4 and SW 2 .
- switch SW 1 may forward another message to SW 3 (and/or other connected devices) indicating such.
- SW 3 may bring the link with the connected end station back up, restoring it to an operational state.
- one or more of the end stations 110 may switch back to the primary connections immediately, or at any appropriate later time, for example, depending on the particular circumstance.
- end stations may switchover much more rapidly than in conventional systems that rely on waiting for a higher layer protocol running on these end stations to detect that its link to the network is not working properly.
- These higher layer protocols typically rely upon timers to fail, and also have to account for network congestion to avoid erroneously detecting “false” failures.
- upstream switches e.g., SW 1 and SW 2
- downstream devices e.g., SW 3 and SW 4
- L X links
- the particular manner in which changes in upstream connectivity are propagated down to an end station may vary, depending on a particular implementation.
- the particular format and contents of the message sent by an upstream device to communicate a change in upstream connectivity may vary.
- the contents of the message may include at least an indication of the detected change (e.g., loss or restoration), as well as any other suitable information, such as identification of an intended recipient (destination).
- the type and/or number of devices to receive a message indicating a change in connectivity may also vary depending on a particular implementation and, for some embodiments, may be configurable by a user (e.g., a network administrator).
- messages may vary with different embodiments and different types of messages may be sent (e.g., broadcast to many devices or sent directly to identified recipients).
- messages may include a common set of fields including, but not limited to one or more of the following: 1) a Switch identifier (source of the msg); 2) a Msg_type/Action_type (e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if a user reconfigures a link from being the cause of downstream interfaces being brought down); 3) an Upstream Link Identifier (which may be at least be unique on the source switch); 4) a SW Version of some type; 5) a msg counter for a number of msgs of this type sent; and 6) a time stamp of action occurrence.
- a Switch identifier source of the msg
- Msg_type/Action_type e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if
- the Upstream Link Identifier may be a field in the message that identifies the upstream interface.
- the message may include an identifier of a group that is effected by the change in connectivity of an upstream interface.
- the upstream interface may be a single physical interface, or it could be a logical grouping of physical interfaces.
- the particular contents of a message may depend on the particular type of message being sent. For some embodiments, there may be at least two types of messages, for example, broadcast or direct, with each having different contents. Regardless of the implementation, msg contents would most likely contain a common set of fields, such as those listed above. The field contents, however, would likely change depending up on the action and the source switch.
- a message may be broadcast to multiple devices connected to a switch detecting a change in communication status. How the message is treated may depend on the particular recipient. As an example, a switch, such as SW 3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto.
- a switch such as SW 3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto.
- a user may be able to configure a particular set of downstream interfaces that should be brought down upon receiving a message indicating a loss of upstream connectivity. For example, as illustrated in FIG. 5 , a user may be able to define a first group 510 1 of end stations with primary interfaces connected to a switch SW 3 that should be brought down in the event that connectivity between SW 1 and the network core (L C ) is lost. This grouping may define a logical connection between these end stations and an upstream interface. As a result, in response to detecting a change in connectivity of the upstream interface, switch SW 3 may disable the end stations in the first group 510 1 . In response to the primary interface being disabled, the end stations may automatically switch over to secondary interfaces connected to SW 4 .
- a second group of end stations 510 2 may not have a secondary NIC and, thus should not be brought down in the event that upstream connectivity is lost.
- Such end stations e.g., older “legacy” end stations
- the downstream interfaces connecting to such end stations may be configured to be maintained (not brought down) by the user in the case of an upstream link losing connectivity.
- end stations in Group 2 may continue to use the same interface with SW 3 and their traffic may be routed to the network core via a redundant link (L Z ) between SW 1 and SW 2 to reach the core. While the bandwidth of the redundant link L Z may not be as high as the direct link to the core (link L C ) and may not be meant to carry sustained traffic to the core, it may still allow the end station to remain active.
- the techniques described herein may be configurable by a user, for example, via a user command to run automatically, on demand, or be disabled. In some cases, a user may be allowed to define logical grouping of upstream interfaces to downstream interfaces. Further, for some embodiments, in addition to, or as an alternative to selecting groups, the techniques described herein could be configurable to be enabled in a variety of ways with varying scope, such as on a “per port” basis, via a particular connection (e.g., a VLAN), or switch-wide.
- the downstream device may bring down a connection with an end station. Bringing down the connection may result in the end station promptly switching over to a redundant connection. Further, by maintaining the link between the intermediate switch and the upstream device, network traffic routed therethrough need not be re-routed.
Abstract
Description
- 1. Field of the Invention
- The present invention generally relates to management of switched computer networks and, more particularly, to propagating detected changes in upstream connectivity to downstream devices.
- 2. Description of the Related Art
- A computer network generally includes a number of devices, including switches, routers and hubs, connected so as to allow communication among the devices. The devices within a network may be generally categorized into two classes: end stations connected along the edge of a network and network devices. End stations typically include devices that primarily supply or consume information, such as workstations, desktop PCs, printers, servers, hosts, fax machines, and other such devices. Network devices typically include devices that primarily transfer information between other devices (e.g., between end stations and other network devices), such as gateways, switches, routers, and other such devices. In other words, network devices generally provide connectivity between end stations and the network core.
-
FIG. 1 illustrates anexemplary network 100 in which a plurality ofend stations 110 are connected to anetwork core 120 via multiple layers ofswitches 130. In some cases, the network may be modeled in accordance with the Open System Interconnection (OSI) model, which defines a networking framework for implementing protocols in multiple (seven) layers. Theswitches 130 may represent Data Link or “Layer 2” layer switching, that encodes and decodes data packets into bits.Layer 2 switches implement the network transmission protocol and handle errors in the physical layer (Ethernet, optics, etc.), flow control and frame synchronization. In the illustrated example,switches 130 include aggregation switches 132 (SW1 and SW2), and access switches 134 (SW3 and SW4). Theend stations 110 may be file servers, data servers, workstations, personal computers (PCs), and the like. - In an effort to provide greater reliability, such networks are often designed to provide
redundant Layer 2 connectivity betweenend stations 110 and thenetwork core 120. To this end, at least someend stations 110 may each have multiple network interface cards NICs 112 (or other type network interface/adapter), allowing theend stations 110 to connect to the network through primary links/connections (Lp) and secondary links (Ls). For example, the end stations may be configured to connect to the network through a primary path (via SW3 and SW1) by default and switch (or “failover”) to a secondary path (via SW4 and SW2) if a connectivity problem is detected in the primary path, such as a loss of connectivity to a link LC between switch SW1 and thenetwork core 120. - In conventional systems, end stations do not become aware of a change in upstream connectivity until they discover it on their own. For example, end stations may only discover a loss of upstream connectivity after experiencing consistent protocol packet timeouts. As a result, after a loss in network connectivity on the primary path, end stations may not immediately switch from their primary path to their alternative path despite a loss in network connectivity on the primary path, which may degrade network performance.
- Therefore, there is a need for techniques that cause end stations to switch over to alternative paths more rapidly in response to remote upstream connectivity changes.
- Embodiments of the present invention provide techniques for propagating changes in upstream connectivity downstream to connected end stations.
- One embodiment provides a method for managing a network path between at least one end station and a network. The method generally includes detecting, by a first switching device, a change in status of a first upstream communication link between the first switching device and the network, sending, by the first switching device, notification of the change in status to a second switching device in communication with the end station via a first downstream communication link, and, in response to receiving the notification, disabling by the second switch, the first downstream communication link in a manner that causes the end station to failover to a second downstream link between a third switch and the end station.
- Another embodiment provides a network switching system generally including at least a first switching device and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable a first downstream communication link between the second switching device and an end station. For some embodiments, their may be one or more intermediate switching devices, between the first and second switching devices, that receive the notification of the change in status from the first switching device and propagate the notification (possibly through other intermediate switches) to the second switching device that ultimately disables the first downstream communication link with the end station.
- Another embodiment provides a system generally including a plurality of end stations, at least a first switching device, and at least a second switching device. The first switching device is generally configured to detect a change in status of a first upstream communication link between the first switching device and a network core. The second switching device is generally configured to receive, from the first switching device, notification of the change in status from the first switching device and, in response, disable first downstream communication links between first network interface devices on the plurality of end stations, thereby causing at least some of the plurality of end stations to switch to second downstream communications links between second network interface devices thereon and at least a third switching device.
- Another embodiment provides an aggregation switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include detecting a change in status of a first upstream communication link between the first switching device and a network and sending notification of the change in status to a second switching device in communication with an end station via a first downstream communication link.
- Another embodiment provides an access switching device generally including a processor and a computer readable medium containing instructions which, when executed by the processor, cause the switching device to perform operations. The operations generally include in response to receiving, from an upstream switching device, notification of a detected change in status between the first switching device and a network, disabling at least one downstream communication link between the access switching device and at least one end station.
- So that the manner in which the above recited features of the present invention can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to embodiments, some of which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
-
FIG. 1 illustrates an exemplary multilayer switched network. -
FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention. -
FIGS. 3A-3D illustrate propagating, and responding to, changes in upstream connectivity, in accordance with one embodiment of the present invention. -
FIG. 4 illustrates the routing of traffic through links maintained between access and aggregation devices, in accordance with one embodiment of the present invention. -
FIG. 5 is a flow diagram of exemplary operations for propagating changes in connectivity, in accordance with embodiments of the present invention. - Embodiments of the present invention generally provide methods and apparatus for propagating changes in upstream connectivity downstream. For some embodiments, an upstream switch connected to a network core may detect a change in connectivity and, in response, send a message indicating the change to a downstream switch connected to an end station. In response to the message, the downstream switch may disable a link with end station. Disabling the link may cause the end station to promptly switch (fail over) to a different link, in an effort to establish an alternate path to the network. Further, by maintaining the link between the intermediate or downstream switch and the upstream device, network traffic routed through that switch need not be re-routed and the link can continue to carry traffic destined for other networks.
- Methods and apparatus for managing a multilayer switching path are described. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be appreciated, however, to one skilled in the art that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form (absent specific details) in order to avoid unnecessarily obscuring the present invention.
- To facilitate understanding, the following terms, as used herein, have the following meaning. As used herein, the term end station generally refers to any device capable of a network connection via one or more network interfaces and need not necessarily reside at an edge of a network. Examples of end stations include a server, PC, fax machine, printer, or the like. The term switch (or switching device) generally refers to any type device that switches network traffic, such as a
Layer 2 switch (per the OSI model) that switches Ethernet traffic. The term access device generally refers to a network element that is connected to one or more end stations. The term aggregation device generally refers to a network element that connects to one or more access devices. The term upstream generally refers to a path (or device along a path) in the direction from an end station towards the core of the network. The term downstream generally refers to a path (or device along a path) in the direction from the core of the network to the edge (e.g., towards an end station). The terms link or connection may be used herein interchangeably and generally refer to any connection between devices, such as aLayer 2 network connection between two devices. The term failure generally refers to a loss of connectivity to a link from a device, regardless of the particular cause, including if the link is intentionally disabled by another device. - Propagating (Downstream) Changes in Upstream Connectivity
- Embodiments of the present invention allow changes in upstream connectivity to be communicated downstream, thereby allowing an end station to quickly make decisions regarding which paths to use to the network. The techniques described herein may be implemented, for example, as software methods (e.g., executable instructions residing in memory) running on network devices (e.g., aggregation devices and/or access devices) that are directly or indirectly connected to a network interface of an end station. As will be described herein, the indirectly connected switches may track the link status on upstream interfaces and communicate any change in connectivity (loss/restoration) to a set of downstream interfaces, for example, by sending a message. Such a message may be intercepted by downstream network devices, such as access switches directly connected to network interfaces of the end stations. In response, these switches may take appropriate action based on the message communicated, for example, disabling corresponding active links if the message indicates loss of connectivity. This allows conventional end stations with redundant network interfaces to detect that their primary connection is down and to automatically switch over to their secondary network interfaces.
-
FIG. 2 is a flow diagram of exemplary operations for propagating upstream connectivity information to a downstream device, in accordance with embodiments of the present invention. For some embodiments, theoperations 200 may be collectively performed by a combination of switches, such as aggregation switches connected to a network core and access switches connected to an end station. To illustrate the potential operation of the switching devices, theoperations 200 may be described with reference to network diagrams ofFIGS. 3A-3D , which illustrate the status of different links in the network as connectivity changes therein. - To facilitate understanding, only a limited portion of the network shown in
FIG. 1 is illustrated inFIGS. 3A-3D . However, those skilled in the art will recognize that the concepts described herein may be advantageously scaled to accommodate any size and type of switching network having any number of multiple switching layers (e.g., any combination of aggregate switches, access switches, routers, and any other types of network elements), with potentially hundreds of end stations. - The operations begin, at
step 202, by detecting a change in upstream connectivity between a first switch and a network core. Referring toFIG. 3A , the exemplary network shown includes a layer ofaggregation switches 332 SW1 and SW2 and a layer ofaccess switches 334 SW3 and SW4. As illustrated, a primary path betweenend stations 110 and thenetwork core 120 may be established through SW3 and SW1 via first network interface cards 112 (NIC1). As illustrated inFIG. 3 , however, a loss of connectivity between SW1 and the network core may occur for any number of reasons (e.g., device failure, such as failure of a laser in a fiber optic connection, failure in the physical connection/cable, or loss of power). For some embodiments, switches SW3 may have downstream connections to both end stations and other switches, not just to one another. - At
step 204, the change in connectivity is communicated to a second switch located downstream between the first switch and an end station. As further illustrated inFIG. 3B , the first switch SW1 may communicate the change in connectivity (a loss in this example), to second switch SW3 via a change in connectivity message. As will be described in greater detail below, the contents of the message may vary, as may the recipients. For some embodiments, a user (e.g., network administrator) may be able to configure the manner in which the message is sent, for example, broadcasting to all devices to respond/ignore or only sending the message to identified devices (e.g., along a path to an end device). - At
step 206, a downstream connection between the second switch and the end station is modified in response to the change in connectivity. For example, as illustrated inFIG. 3C , SW3 may decipher the message from SW1 and, in response, disable (bring down) the link between the primary NICs of theend stations 110. Exactly how the link is disabled may depend on the corresponding type of physical medium. For example, for an Ethernet connection this may entail simply ceasing to transmit a signal (e.g., by disabling a laser for a fiber optic Ethernet connection or bringing down the voltage level for a copper Ethernet connection). - In any case, end stations equipped with redundant network interfaces may be configured to react to such a loss in connectivity to a primary link by automatically switching (or “failing over”) to a secondary interface in order to establish an alternate path to the network. Thus, in the illustrated example as shown in
FIG. 3D ,end stations 110 may automatically switch over to secondary NICs to establish an alternate path to the core network via SW4 and SW2. - For some embodiments, should the connectivity between SW1 and the network core be restored, switch SW1 may forward another message to SW3 (and/or other connected devices) indicating such. In response to the message, SW3 may bring the link with the connected end station back up, restoring it to an operational state. In response to detecting that the primary link has been brought back up, one or more of the
end stations 110 may switch back to the primary connections immediately, or at any appropriate later time, for example, depending on the particular circumstance. - By propagating an
upstream Layer 2 connectivity failure in this manner, switchover times for the end stations may be reduced. In other words, end stations may switchover much more rapidly than in conventional systems that rely on waiting for a higher layer protocol running on these end stations to detect that its link to the network is not working properly. These higher layer protocols typically rely upon timers to fail, and also have to account for network congestion to avoid erroneously detecting “false” failures. - In addition to improving switchover times, the techniques described herein also allow upstream switches (e.g., SW1 and SW2) to maintain their interfaces to downstream devices (e.g., SW3 and SW4). This is advantageous in that other network traffic may continue to pass through the interface from access devices to aggregation devices and allows multiple networks to be active that would otherwise be inactive if this link were brought down. These links (LX) may continue to carry traffic for management, as well as other local network traffic.
- The particular manner in which changes in upstream connectivity are propagated down to an end station may vary, depending on a particular implementation. As an example, the particular format and contents of the message sent by an upstream device to communicate a change in upstream connectivity may vary. For some embodiments, the contents of the message may include at least an indication of the detected change (e.g., loss or restoration), as well as any other suitable information, such as identification of an intended recipient (destination). The type and/or number of devices to receive a message indicating a change in connectivity may also vary depending on a particular implementation and, for some embodiments, may be configurable by a user (e.g., a network administrator).
- The contents of messages may vary with different embodiments and different types of messages may be sent (e.g., broadcast to many devices or sent directly to identified recipients). In any case, messages may include a common set of fields including, but not limited to one or more of the following: 1) a Switch identifier (source of the msg); 2) a Msg_type/Action_type (e.g., Link Lost, Link Restored, or link removed from being monitored, for example, if a user reconfigures a link from being the cause of downstream interfaces being brought down); 3) an Upstream Link Identifier (which may be at least be unique on the source switch); 4) a SW Version of some type; 5) a msg counter for a number of msgs of this type sent; and 6) a time stamp of action occurrence. For some embodiments, the Upstream Link Identifier may be a field in the message that identifies the upstream interface. As an alternative, the message may include an identifier of a group that is effected by the change in connectivity of an upstream interface. The upstream interface may be a single physical interface, or it could be a logical grouping of physical interfaces.
- The particular contents of a message may depend on the particular type of message being sent. For some embodiments, there may be at least two types of messages, for example, broadcast or direct, with each having different contents. Regardless of the implementation, msg contents would most likely contain a common set of fields, such as those listed above. The field contents, however, would likely change depending up on the action and the source switch.
- As illustrated in
FIG. 4 , for some embodiments, a message may be broadcast to multiple devices connected to a switch detecting a change in communication status. How the message is treated may depend on the particular recipient. As an example, a switch, such as SW3 in a downstream path between the switch sending the message may intercept the message and respond accordingly, while other devices may ignore the message or forward it on to other devices connected thereto. - For some embodiments, a user may be able to configure a particular set of downstream interfaces that should be brought down upon receiving a message indicating a loss of upstream connectivity. For example, as illustrated in
FIG. 5 , a user may be able to define a first group 510 1 of end stations with primary interfaces connected to a switch SW3 that should be brought down in the event that connectivity between SW1 and the network core (LC) is lost. This grouping may define a logical connection between these end stations and an upstream interface. As a result, in response to detecting a change in connectivity of the upstream interface, switch SW3 may disable the end stations in the first group 510 1. In response to the primary interface being disabled, the end stations may automatically switch over to secondary interfaces connected to SW4. - A second group of end stations 510 2 may not have a secondary NIC and, thus should not be brought down in the event that upstream connectivity is lost. Such end stations (e.g., older “legacy” end stations) may not have the capability to switch over from a primary to secondary interface. As such, the downstream interfaces connecting to such end stations may be configured to be maintained (not brought down) by the user in the case of an upstream link losing connectivity. As illustrated, end stations in
Group 2 may continue to use the same interface with SW3 and their traffic may be routed to the network core via a redundant link (LZ) between SW1 and SW2 to reach the core. While the bandwidth of the redundant link LZ may not be as high as the direct link to the core (link LC) and may not be meant to carry sustained traffic to the core, it may still allow the end station to remain active. - For some embodiments, the techniques described herein may be configurable by a user, for example, via a user command to run automatically, on demand, or be disabled. In some cases, a user may be allowed to define logical grouping of upstream interfaces to downstream interfaces. Further, for some embodiments, in addition to, or as an alternative to selecting groups, the techniques described herein could be configurable to be enabled in a variety of ways with varying scope, such as on a “per port” basis, via a particular connection (e.g., a VLAN), or switch-wide.
- By propagating upstream connectivity losses detected by an upstream device to a downstream device connected, the downstream device may bring down a connection with an end station. Bringing down the connection may result in the end station promptly switching over to a redundant connection. Further, by maintaining the link between the intermediate switch and the upstream device, network traffic routed therethrough need not be re-routed.
- While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Claims (28)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/398,519 US8886831B2 (en) | 2006-04-05 | 2006-04-05 | System and methodology for fast link failover based on remote upstream failures |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/398,519 US8886831B2 (en) | 2006-04-05 | 2006-04-05 | System and methodology for fast link failover based on remote upstream failures |
Publications (2)
Publication Number | Publication Date |
---|---|
US20070237085A1 true US20070237085A1 (en) | 2007-10-11 |
US8886831B2 US8886831B2 (en) | 2014-11-11 |
Family
ID=38575123
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/398,519 Expired - Fee Related US8886831B2 (en) | 2006-04-05 | 2006-04-05 | System and methodology for fast link failover based on remote upstream failures |
Country Status (1)
Country | Link |
---|---|
US (1) | US8886831B2 (en) |
Cited By (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070277828A1 (en) * | 2006-06-05 | 2007-12-06 | Ho Peter C F | Flexible connector |
US20080049631A1 (en) * | 2006-08-22 | 2008-02-28 | Morrill Robert J | System and method for monitoring interlayer devices and optimizing network performance |
WO2009097615A1 (en) * | 2008-02-01 | 2009-08-06 | Allied Telesis Holdings Kabushiki Kaisha | Method and system for dynamic link failover management |
US20100020680A1 (en) * | 2008-07-28 | 2010-01-28 | Salam Samer M | Multi-chassis ethernet link aggregation |
US20100296390A1 (en) * | 2006-11-27 | 2010-11-25 | Honeywell International Inc. | Fault tolerance in wireless networks |
US7843831B2 (en) | 2006-08-22 | 2010-11-30 | Embarq Holdings Company Llc | System and method for routing data on a packet network |
US7940735B2 (en) | 2006-08-22 | 2011-05-10 | Embarq Holdings Company, Llc | System and method for selecting an access point |
US7948909B2 (en) | 2006-06-30 | 2011-05-24 | Embarq Holdings Company, Llc | System and method for resetting counters counting network performance information at network communications devices on a packet network |
US8000318B2 (en) | 2006-06-30 | 2011-08-16 | Embarq Holdings Company, Llc | System and method for call routing based on transmission performance of a packet network |
US20110211585A1 (en) * | 2010-02-26 | 2011-09-01 | Alaxala Networks Corporation | Packet relay device |
US8015294B2 (en) | 2006-08-22 | 2011-09-06 | Embarq Holdings Company, LP | Pin-hole firewall for communicating data packets on a packet network |
US8040811B2 (en) | 2006-08-22 | 2011-10-18 | Embarq Holdings Company, Llc | System and method for collecting and managing network performance information |
US8064391B2 (en) | 2006-08-22 | 2011-11-22 | Embarq Holdings Company, Llc | System and method for monitoring and optimizing network performance to a wireless device |
US8068425B2 (en) | 2008-04-09 | 2011-11-29 | Embarq Holdings Company, Llc | System and method for using network performance information to determine improved measures of path states |
US8098579B2 (en) | 2006-08-22 | 2012-01-17 | Embarq Holdings Company, LP | System and method for adjusting the window size of a TCP packet through remote network elements |
US8102770B2 (en) | 2006-08-22 | 2012-01-24 | Embarq Holdings Company, LP | System and method for monitoring and optimizing network performance with vector performance tables and engines |
US8107366B2 (en) | 2006-08-22 | 2012-01-31 | Embarq Holdings Company, LP | System and method for using centralized network performance tables to manage network communications |
US8111692B2 (en) | 2007-05-31 | 2012-02-07 | Embarq Holdings Company Llc | System and method for modifying network traffic |
US8125897B2 (en) | 2006-08-22 | 2012-02-28 | Embarq Holdings Company Lp | System and method for monitoring and optimizing network performance with user datagram protocol network performance information packets |
US8130793B2 (en) | 2006-08-22 | 2012-03-06 | Embarq Holdings Company, Llc | System and method for enabling reciprocal billing for different types of communications over a packet network |
US8144586B2 (en) | 2006-08-22 | 2012-03-27 | Embarq Holdings Company, Llc | System and method for controlling network bandwidth with a connection admission control engine |
US8144587B2 (en) | 2006-08-22 | 2012-03-27 | Embarq Holdings Company, Llc | System and method for load balancing network resources using a connection admission control engine |
US8184549B2 (en) | 2006-06-30 | 2012-05-22 | Embarq Holdings Company, LLP | System and method for selecting network egress |
US8189468B2 (en) | 2006-10-25 | 2012-05-29 | Embarq Holdings, Company, LLC | System and method for regulating messages between networks |
US8194643B2 (en) | 2006-10-19 | 2012-06-05 | Embarq Holdings Company, Llc | System and method for monitoring the connection of an end-user to a remote network |
US8194555B2 (en) | 2006-08-22 | 2012-06-05 | Embarq Holdings Company, Llc | System and method for using distributed network performance information tables to manage network communications |
US8199653B2 (en) | 2006-08-22 | 2012-06-12 | Embarq Holdings Company, Llc | System and method for communicating network performance information over a packet network |
US8224255B2 (en) | 2006-08-22 | 2012-07-17 | Embarq Holdings Company, Llc | System and method for managing radio frequency windows |
US8228791B2 (en) | 2006-08-22 | 2012-07-24 | Embarq Holdings Company, Llc | System and method for routing communications between packet networks based on intercarrier agreements |
US8274905B2 (en) | 2006-08-22 | 2012-09-25 | Embarq Holdings Company, Llc | System and method for displaying a graph representative of network performance over a time period |
US8289965B2 (en) | 2006-10-19 | 2012-10-16 | Embarq Holdings Company, Llc | System and method for establishing a communications session with an end-user based on the state of a network connection |
US8307065B2 (en) | 2006-08-22 | 2012-11-06 | Centurylink Intellectual Property Llc | System and method for remotely controlling network operators |
US8358580B2 (en) | 2006-08-22 | 2013-01-22 | Centurylink Intellectual Property Llc | System and method for adjusting the window size of a TCP packet through network elements |
US8407765B2 (en) | 2006-08-22 | 2013-03-26 | Centurylink Intellectual Property Llc | System and method for restricting access to network performance information tables |
US8488447B2 (en) | 2006-06-30 | 2013-07-16 | Centurylink Intellectual Property Llc | System and method for adjusting code speed in a transmission path during call set-up due to reduced transmission performance |
US20130232277A1 (en) * | 2011-08-31 | 2013-09-05 | Metaswitch Networks Ltd. | Transmitting and Forwarding Data |
US8531954B2 (en) | 2006-08-22 | 2013-09-10 | Centurylink Intellectual Property Llc | System and method for handling reservation requests with a connection admission control engine |
US8537695B2 (en) | 2006-08-22 | 2013-09-17 | Centurylink Intellectual Property Llc | System and method for establishing a call being received by a trunk on a packet network |
US8549405B2 (en) | 2006-08-22 | 2013-10-01 | Centurylink Intellectual Property Llc | System and method for displaying a graphical representation of a network to identify nodes and node segments on the network that are not operating normally |
US8576722B2 (en) | 2006-08-22 | 2013-11-05 | Centurylink Intellectual Property Llc | System and method for modifying connectivity fault management packets |
US8619600B2 (en) | 2006-08-22 | 2013-12-31 | Centurylink Intellectual Property Llc | System and method for establishing calls over a call path having best path metrics |
US8717911B2 (en) | 2006-06-30 | 2014-05-06 | Centurylink Intellectual Property Llc | System and method for collecting network performance information |
US8743703B2 (en) | 2006-08-22 | 2014-06-03 | Centurylink Intellectual Property Llc | System and method for tracking application resource usage |
US8743700B2 (en) | 2006-08-22 | 2014-06-03 | Centurylink Intellectual Property Llc | System and method for provisioning resources of a packet network based on collected network performance information |
US8750158B2 (en) | 2006-08-22 | 2014-06-10 | Centurylink Intellectual Property Llc | System and method for differentiated billing |
US8824276B2 (en) * | 2012-06-01 | 2014-09-02 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of MOFRR with dataplane notifications |
US8913482B2 (en) | 2012-06-01 | 2014-12-16 | Telefonaktiebolaget L M Ericsson (Publ) | Enhancements to PIM fast re-route with upstream activation packets |
US20150172070A1 (en) * | 2012-06-01 | 2015-06-18 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of mofrr with dataplane notifications |
US9094257B2 (en) | 2006-06-30 | 2015-07-28 | Centurylink Intellectual Property Llc | System and method for selecting a content delivery network |
US20150295861A1 (en) * | 2013-01-06 | 2015-10-15 | Hangzhou H3C Technologies Co., Ltd. | Link problem handling |
US9479341B2 (en) | 2006-08-22 | 2016-10-25 | Centurylink Intellectual Property Llc | System and method for initiating diagnostics on a packet network node |
WO2017160306A1 (en) * | 2016-03-18 | 2017-09-21 | Hewlett Packard Enterprise Development Lp | Responses to loops in networks having a ring topology |
US11012369B2 (en) * | 2019-07-05 | 2021-05-18 | Dell Products L.P. | Aggregated switch path optimization system |
CN114679348A (en) * | 2020-12-24 | 2022-06-28 | 华为技术有限公司 | Fault processing method and related product |
US11551684B1 (en) * | 2017-12-08 | 2023-01-10 | Amazon Technologies, Inc. | State detection and responses for electronic devices |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9866470B2 (en) * | 2014-01-24 | 2018-01-09 | Red Hat, Inc. | Multiple active link aggregators |
US9838302B1 (en) * | 2015-06-10 | 2017-12-05 | Amazon Technologies, Inc. | Managing loss of network connectivity in traffic forwarding systems |
US10237157B1 (en) | 2015-06-10 | 2019-03-19 | Amazon Technologies, Inc. | Managing host failures in a traffic forwarding system |
FR3041842A1 (en) * | 2015-09-30 | 2017-03-31 | Orange | SERVICE RESTORATION SYSTEM PROVIDED BY A RESIDENTIAL GATEWAY |
US10097462B2 (en) * | 2016-04-02 | 2018-10-09 | Niciria, Inc. | Throughput resilience during link failover |
US11533231B2 (en) | 2019-11-29 | 2022-12-20 | Amazon Technologies, Inc. | Configuration and management of scalable global private networks |
US11729077B2 (en) | 2019-11-29 | 2023-08-15 | Amazon Technologies, Inc. | Configuration and management of scalable global private networks |
US10999169B1 (en) * | 2019-11-29 | 2021-05-04 | Amazon Technologies, Inc. | Configuration and management of scalable global private networks |
US11336528B2 (en) | 2019-11-29 | 2022-05-17 | Amazon Technologies, Inc. | Configuration and management of scalable global private networks |
Citations (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5835696A (en) * | 1995-11-22 | 1998-11-10 | Lucent Technologies Inc. | Data router backup feature |
US6032194A (en) * | 1997-12-24 | 2000-02-29 | Cisco Technology, Inc. | Method and apparatus for rapidly reconfiguring computer networks |
US6330229B1 (en) * | 1998-11-09 | 2001-12-11 | 3Com Corporation | Spanning tree with rapid forwarding database updates |
US6392990B1 (en) * | 1999-07-23 | 2002-05-21 | Glenayre Electronics, Inc. | Method for implementing interface redundancy in a computer network |
US20020112072A1 (en) * | 2001-02-12 | 2002-08-15 | Maple Optical Systems, Inc. | System and method for fast-rerouting of data in a data communication network |
US6442133B2 (en) * | 2000-05-24 | 2002-08-27 | Nec Corporation | Switchboard having a dual switching system |
US20020152320A1 (en) * | 2001-02-14 | 2002-10-17 | Lau Pui Lun | System and method for rapidly switching between redundant networks |
US20020167895A1 (en) * | 2001-05-08 | 2002-11-14 | Jack Zhu | Method for restoring diversely routed circuits |
US20020184387A1 (en) * | 2001-05-31 | 2002-12-05 | The Furukawa Electric Co., Ltd. | Method for connecting between networks, virtual router, and system for connecting between networks by using this virtual router |
US20030016624A1 (en) * | 1998-05-04 | 2003-01-23 | Bare Ballard C. | Path recovery on failure in load balancing switch protocols |
US6513129B1 (en) * | 1999-06-30 | 2003-01-28 | Objective Systems Integrators, Inc. | System and method for managing faults using a gateway |
US20030021226A1 (en) * | 2001-07-24 | 2003-01-30 | Gal Mor | Interconnect and gateway protection in bidirectional ring networks |
US6542934B1 (en) * | 1997-05-30 | 2003-04-01 | International Business Machines Corporation | Non-disruptively rerouting network communications from a secondary network path to a primary path |
US20030097470A1 (en) * | 2001-09-10 | 2003-05-22 | Roger Lapuh | System, device, and method for improving communication network reliability using trunk splitting |
US20030137932A1 (en) * | 2002-01-23 | 2003-07-24 | Nec Corporation | Path fault recovery method, switching-back method after recovery from fault, and node using the same |
US20030158964A1 (en) * | 2002-02-15 | 2003-08-21 | Naofumi Kobayashi | Upper and lower relays and network system |
US20030161260A1 (en) * | 2002-02-25 | 2003-08-28 | Sundara Murugan | Method and apparatus for implementing automatic protection switching functionality in a distributed processor data router |
US20030179700A1 (en) * | 1999-01-15 | 2003-09-25 | Saleh Ali Najib | Method for restoring a virtual path in an optical network using 1‘protection |
US6628661B1 (en) * | 1998-08-27 | 2003-09-30 | Intel Corporation | Spanning tree recovery in computer networks |
US20030185225A1 (en) * | 2002-03-29 | 2003-10-02 | Wirth Brian Michael | Switch and a switching apparatus for a communication network |
US20040001449A1 (en) * | 2002-06-28 | 2004-01-01 | Rostron Andy E. | System and method for supporting automatic protection switching between multiple node pairs using common agent architecture |
US6678241B1 (en) * | 1999-11-30 | 2004-01-13 | Cisc Technology, Inc. | Fast convergence with topology switching |
US6697339B1 (en) * | 1999-03-04 | 2004-02-24 | 3Com Corporation | High availability spanning tree with rapid reconfiguration with alternate port selection |
US20040047336A1 (en) * | 1999-08-23 | 2004-03-11 | Avaya Communication Israel Ltd. | Modular bridging-device |
US6721275B1 (en) * | 1999-05-03 | 2004-04-13 | Hewlett-Packard Development Company, L.P. | Bridged network stations location revision |
US20040098501A1 (en) * | 2002-10-29 | 2004-05-20 | Finn Norman W. | Multi-bridge lan aggregation |
US20040105390A1 (en) * | 2001-05-28 | 2004-06-03 | Nokia Corporation | Method and system for implementing a fast recovery process in a local area network |
US20040133693A1 (en) * | 2000-01-28 | 2004-07-08 | 3Com Corporation | Multi-point link aggregation spoofing |
US6801506B1 (en) * | 1999-03-31 | 2004-10-05 | Cisco Technology, Inc. | Method and apparatus for providing fast spanning tree re-starts |
US20040264364A1 (en) * | 2003-06-27 | 2004-12-30 | Nec Corporation | Network system for building redundancy within groups |
US20050063395A1 (en) * | 2003-09-18 | 2005-03-24 | Cisco Technology, Inc. | Virtual network device |
US6891808B2 (en) * | 2000-03-10 | 2005-05-10 | Anritsu Corporation | Spanning tree bridge and route change method using the same |
US20050108401A1 (en) * | 2003-11-13 | 2005-05-19 | Gonda Rumi S. | Method for supporting SDH/SONET OAMP on Ethernet |
US20050111350A1 (en) * | 2003-11-20 | 2005-05-26 | Shinya Kano | Method and apparatus for designing backup communication path, and computer product |
US6956824B2 (en) * | 2001-06-14 | 2005-10-18 | Tropic Networks Inc. | Extension of link aggregation protocols over the network |
US6987740B1 (en) * | 2000-09-11 | 2006-01-17 | Cisco Technology, Inc. | STP root guard |
US20060023630A1 (en) * | 2003-05-19 | 2006-02-02 | Hiroyuki Akiyama | SDH transmission apparatus and signal transmission method |
US7016351B1 (en) * | 2000-02-29 | 2006-03-21 | Cisco Technology, Inc. | Small group multicast in a computer network |
US20060159011A1 (en) * | 2005-01-14 | 2006-07-20 | Mitesh Dalal | Detecting unavailable network connections |
US20060206602A1 (en) * | 2005-03-14 | 2006-09-14 | International Business Machines Corporation | Network switch link failover in a redundant switch configuration |
US20060227704A1 (en) * | 2005-04-12 | 2006-10-12 | Fujitsu Limited | Special marker message for link aggregation marker protocol |
US20060256712A1 (en) * | 2003-02-21 | 2006-11-16 | Nippon Telegraph And Telephone Corporation | Device and method for correcting a path trouble in a communication network |
US20070041313A1 (en) * | 2005-08-17 | 2007-02-22 | Honeywell International, Inc | Apparatus and methods for managing nodes on a fault tolerant network |
US20070047436A1 (en) * | 2005-08-24 | 2007-03-01 | Masaya Arai | Network relay device and control method |
US20070047578A1 (en) * | 2005-08-29 | 2007-03-01 | Fujitsu Limited | Bandwidth control method and transmission equipment |
US7197660B1 (en) * | 2002-06-26 | 2007-03-27 | Juniper Networks, Inc. | High availability network security systems |
US20070070886A1 (en) * | 2005-09-29 | 2007-03-29 | Seth Zirin | Modifying an endpoint node connection associated with a destination |
US20070076727A1 (en) * | 2005-09-30 | 2007-04-05 | Tekelec | Adaptive redundancy protection scheme |
US20070183347A1 (en) * | 2004-08-29 | 2007-08-09 | Huawei Technologies Co., Ltd. | Method for implementing dual-homing |
US20070280103A1 (en) * | 2006-06-02 | 2007-12-06 | Cisco Technology, Inc. | Transparent automatic protection switching for a chassis deployment |
US20080016402A1 (en) * | 2006-07-11 | 2008-01-17 | Corrigent Systems Ltd. | Connectivity fault management (CFM) in networks with link aggregation group connections |
US20080181196A1 (en) * | 2007-01-31 | 2008-07-31 | Alcatel Lucent | Link aggregation across multiple chassis |
US7453900B2 (en) * | 2003-03-05 | 2008-11-18 | Cisco Technology, Inc. | System and method for monitoring noise associated with a communication link |
US7480283B1 (en) * | 2002-03-26 | 2009-01-20 | Nortel Networks Limited | Virtual trunking over packet networks |
US7518986B1 (en) * | 2005-11-16 | 2009-04-14 | Juniper Networks, Inc. | Push-based hierarchical state propagation within a multi-chassis network device |
US20090096592A1 (en) * | 2007-10-12 | 2009-04-16 | Alpha Networks Inc. | Power line adapter and method of controlling power line adapter operated in power-saving mode |
US20090109998A1 (en) * | 2007-10-29 | 2009-04-30 | Alcatel Lucent | Facilitating self configuring link aggregation using link aggregation control protocol |
US7580349B1 (en) * | 2001-11-02 | 2009-08-25 | Nortel Networks Limited | Content-aware dynamic network resource allocation |
US7593319B1 (en) * | 2002-10-15 | 2009-09-22 | Garrettcom, Inc. | LAN switch with rapid fault recovery |
US7616561B1 (en) * | 2005-01-19 | 2009-11-10 | Juniper Networks, Inc. | Systems and methods for routing data in a communications network |
US20100020680A1 (en) * | 2008-07-28 | 2010-01-28 | Salam Samer M | Multi-chassis ethernet link aggregation |
-
2006
- 2006-04-05 US US11/398,519 patent/US8886831B2/en not_active Expired - Fee Related
Patent Citations (62)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5835696A (en) * | 1995-11-22 | 1998-11-10 | Lucent Technologies Inc. | Data router backup feature |
US6542934B1 (en) * | 1997-05-30 | 2003-04-01 | International Business Machines Corporation | Non-disruptively rerouting network communications from a secondary network path to a primary path |
US6535491B2 (en) * | 1997-12-24 | 2003-03-18 | Cisco Technology, Inc. | Method and apparatus for rapidly reconfiguring computer networks using a spanning tree algorithm |
US6032194A (en) * | 1997-12-24 | 2000-02-29 | Cisco Technology, Inc. | Method and apparatus for rapidly reconfiguring computer networks |
US20030016624A1 (en) * | 1998-05-04 | 2003-01-23 | Bare Ballard C. | Path recovery on failure in load balancing switch protocols |
US6628661B1 (en) * | 1998-08-27 | 2003-09-30 | Intel Corporation | Spanning tree recovery in computer networks |
US6330229B1 (en) * | 1998-11-09 | 2001-12-11 | 3Com Corporation | Spanning tree with rapid forwarding database updates |
US20030179700A1 (en) * | 1999-01-15 | 2003-09-25 | Saleh Ali Najib | Method for restoring a virtual path in an optical network using 1‘protection |
US6697339B1 (en) * | 1999-03-04 | 2004-02-24 | 3Com Corporation | High availability spanning tree with rapid reconfiguration with alternate port selection |
US6801506B1 (en) * | 1999-03-31 | 2004-10-05 | Cisco Technology, Inc. | Method and apparatus for providing fast spanning tree re-starts |
US6721275B1 (en) * | 1999-05-03 | 2004-04-13 | Hewlett-Packard Development Company, L.P. | Bridged network stations location revision |
US6513129B1 (en) * | 1999-06-30 | 2003-01-28 | Objective Systems Integrators, Inc. | System and method for managing faults using a gateway |
US6392990B1 (en) * | 1999-07-23 | 2002-05-21 | Glenayre Electronics, Inc. | Method for implementing interface redundancy in a computer network |
US20040047336A1 (en) * | 1999-08-23 | 2004-03-11 | Avaya Communication Israel Ltd. | Modular bridging-device |
US6678241B1 (en) * | 1999-11-30 | 2004-01-13 | Cisc Technology, Inc. | Fast convergence with topology switching |
US20040133693A1 (en) * | 2000-01-28 | 2004-07-08 | 3Com Corporation | Multi-point link aggregation spoofing |
US7016351B1 (en) * | 2000-02-29 | 2006-03-21 | Cisco Technology, Inc. | Small group multicast in a computer network |
US6891808B2 (en) * | 2000-03-10 | 2005-05-10 | Anritsu Corporation | Spanning tree bridge and route change method using the same |
US6442133B2 (en) * | 2000-05-24 | 2002-08-27 | Nec Corporation | Switchboard having a dual switching system |
US6987740B1 (en) * | 2000-09-11 | 2006-01-17 | Cisco Technology, Inc. | STP root guard |
US20020112072A1 (en) * | 2001-02-12 | 2002-08-15 | Maple Optical Systems, Inc. | System and method for fast-rerouting of data in a data communication network |
US20020152320A1 (en) * | 2001-02-14 | 2002-10-17 | Lau Pui Lun | System and method for rapidly switching between redundant networks |
US20020167895A1 (en) * | 2001-05-08 | 2002-11-14 | Jack Zhu | Method for restoring diversely routed circuits |
US20040105390A1 (en) * | 2001-05-28 | 2004-06-03 | Nokia Corporation | Method and system for implementing a fast recovery process in a local area network |
US20020184387A1 (en) * | 2001-05-31 | 2002-12-05 | The Furukawa Electric Co., Ltd. | Method for connecting between networks, virtual router, and system for connecting between networks by using this virtual router |
US6956824B2 (en) * | 2001-06-14 | 2005-10-18 | Tropic Networks Inc. | Extension of link aggregation protocols over the network |
US20030021226A1 (en) * | 2001-07-24 | 2003-01-30 | Gal Mor | Interconnect and gateway protection in bidirectional ring networks |
US20030097470A1 (en) * | 2001-09-10 | 2003-05-22 | Roger Lapuh | System, device, and method for improving communication network reliability using trunk splitting |
US7580349B1 (en) * | 2001-11-02 | 2009-08-25 | Nortel Networks Limited | Content-aware dynamic network resource allocation |
US20030137932A1 (en) * | 2002-01-23 | 2003-07-24 | Nec Corporation | Path fault recovery method, switching-back method after recovery from fault, and node using the same |
US20030158964A1 (en) * | 2002-02-15 | 2003-08-21 | Naofumi Kobayashi | Upper and lower relays and network system |
US20030161260A1 (en) * | 2002-02-25 | 2003-08-28 | Sundara Murugan | Method and apparatus for implementing automatic protection switching functionality in a distributed processor data router |
US7480283B1 (en) * | 2002-03-26 | 2009-01-20 | Nortel Networks Limited | Virtual trunking over packet networks |
US20030185225A1 (en) * | 2002-03-29 | 2003-10-02 | Wirth Brian Michael | Switch and a switching apparatus for a communication network |
US7197660B1 (en) * | 2002-06-26 | 2007-03-27 | Juniper Networks, Inc. | High availability network security systems |
US20040001449A1 (en) * | 2002-06-28 | 2004-01-01 | Rostron Andy E. | System and method for supporting automatic protection switching between multiple node pairs using common agent architecture |
US7593319B1 (en) * | 2002-10-15 | 2009-09-22 | Garrettcom, Inc. | LAN switch with rapid fault recovery |
US20040098501A1 (en) * | 2002-10-29 | 2004-05-20 | Finn Norman W. | Multi-bridge lan aggregation |
US20060256712A1 (en) * | 2003-02-21 | 2006-11-16 | Nippon Telegraph And Telephone Corporation | Device and method for correcting a path trouble in a communication network |
US7453900B2 (en) * | 2003-03-05 | 2008-11-18 | Cisco Technology, Inc. | System and method for monitoring noise associated with a communication link |
US20060023630A1 (en) * | 2003-05-19 | 2006-02-02 | Hiroyuki Akiyama | SDH transmission apparatus and signal transmission method |
US20040264364A1 (en) * | 2003-06-27 | 2004-12-30 | Nec Corporation | Network system for building redundancy within groups |
US20050063395A1 (en) * | 2003-09-18 | 2005-03-24 | Cisco Technology, Inc. | Virtual network device |
US20050108401A1 (en) * | 2003-11-13 | 2005-05-19 | Gonda Rumi S. | Method for supporting SDH/SONET OAMP on Ethernet |
US20050111350A1 (en) * | 2003-11-20 | 2005-05-26 | Shinya Kano | Method and apparatus for designing backup communication path, and computer product |
US20070183347A1 (en) * | 2004-08-29 | 2007-08-09 | Huawei Technologies Co., Ltd. | Method for implementing dual-homing |
US20060159011A1 (en) * | 2005-01-14 | 2006-07-20 | Mitesh Dalal | Detecting unavailable network connections |
US7616561B1 (en) * | 2005-01-19 | 2009-11-10 | Juniper Networks, Inc. | Systems and methods for routing data in a communications network |
US20060206602A1 (en) * | 2005-03-14 | 2006-09-14 | International Business Machines Corporation | Network switch link failover in a redundant switch configuration |
US20060227704A1 (en) * | 2005-04-12 | 2006-10-12 | Fujitsu Limited | Special marker message for link aggregation marker protocol |
US20070041313A1 (en) * | 2005-08-17 | 2007-02-22 | Honeywell International, Inc | Apparatus and methods for managing nodes on a fault tolerant network |
US20070047436A1 (en) * | 2005-08-24 | 2007-03-01 | Masaya Arai | Network relay device and control method |
US20070047578A1 (en) * | 2005-08-29 | 2007-03-01 | Fujitsu Limited | Bandwidth control method and transmission equipment |
US20070070886A1 (en) * | 2005-09-29 | 2007-03-29 | Seth Zirin | Modifying an endpoint node connection associated with a destination |
US20070076727A1 (en) * | 2005-09-30 | 2007-04-05 | Tekelec | Adaptive redundancy protection scheme |
US7518986B1 (en) * | 2005-11-16 | 2009-04-14 | Juniper Networks, Inc. | Push-based hierarchical state propagation within a multi-chassis network device |
US20070280103A1 (en) * | 2006-06-02 | 2007-12-06 | Cisco Technology, Inc. | Transparent automatic protection switching for a chassis deployment |
US20080016402A1 (en) * | 2006-07-11 | 2008-01-17 | Corrigent Systems Ltd. | Connectivity fault management (CFM) in networks with link aggregation group connections |
US20080181196A1 (en) * | 2007-01-31 | 2008-07-31 | Alcatel Lucent | Link aggregation across multiple chassis |
US20090096592A1 (en) * | 2007-10-12 | 2009-04-16 | Alpha Networks Inc. | Power line adapter and method of controlling power line adapter operated in power-saving mode |
US20090109998A1 (en) * | 2007-10-29 | 2009-04-30 | Alcatel Lucent | Facilitating self configuring link aggregation using link aggregation control protocol |
US20100020680A1 (en) * | 2008-07-28 | 2010-01-28 | Salam Samer M | Multi-chassis ethernet link aggregation |
Cited By (120)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070277828A1 (en) * | 2006-06-05 | 2007-12-06 | Ho Peter C F | Flexible connector |
US9094257B2 (en) | 2006-06-30 | 2015-07-28 | Centurylink Intellectual Property Llc | System and method for selecting a content delivery network |
US9549004B2 (en) | 2006-06-30 | 2017-01-17 | Centurylink Intellectual Property Llc | System and method for re-routing calls |
US7948909B2 (en) | 2006-06-30 | 2011-05-24 | Embarq Holdings Company, Llc | System and method for resetting counters counting network performance information at network communications devices on a packet network |
US9749399B2 (en) | 2006-06-30 | 2017-08-29 | Centurylink Intellectual Property Llc | System and method for selecting a content delivery network |
US8000318B2 (en) | 2006-06-30 | 2011-08-16 | Embarq Holdings Company, Llc | System and method for call routing based on transmission performance of a packet network |
US9154634B2 (en) | 2006-06-30 | 2015-10-06 | Centurylink Intellectual Property Llc | System and method for managing network communications |
US9118583B2 (en) | 2006-06-30 | 2015-08-25 | Centurylink Intellectual Property Llc | System and method for re-routing calls |
US9054915B2 (en) | 2006-06-30 | 2015-06-09 | Centurylink Intellectual Property Llc | System and method for adjusting CODEC speed in a transmission path during call set-up due to reduced transmission performance |
US9838440B2 (en) | 2006-06-30 | 2017-12-05 | Centurylink Intellectual Property Llc | Managing voice over internet protocol (VoIP) communications |
US10230788B2 (en) | 2006-06-30 | 2019-03-12 | Centurylink Intellectual Property Llc | System and method for selecting a content delivery network |
US8184549B2 (en) | 2006-06-30 | 2012-05-22 | Embarq Holdings Company, LLP | System and method for selecting network egress |
US8976665B2 (en) | 2006-06-30 | 2015-03-10 | Centurylink Intellectual Property Llc | System and method for re-routing calls |
US8717911B2 (en) | 2006-06-30 | 2014-05-06 | Centurylink Intellectual Property Llc | System and method for collecting network performance information |
US10560494B2 (en) | 2006-06-30 | 2020-02-11 | Centurylink Intellectual Property Llc | Managing voice over internet protocol (VoIP) communications |
US8570872B2 (en) | 2006-06-30 | 2013-10-29 | Centurylink Intellectual Property Llc | System and method for selecting network ingress and egress |
US8488447B2 (en) | 2006-06-30 | 2013-07-16 | Centurylink Intellectual Property Llc | System and method for adjusting code speed in a transmission path during call set-up due to reduced transmission performance |
US8477614B2 (en) | 2006-06-30 | 2013-07-02 | Centurylink Intellectual Property Llc | System and method for routing calls if potential call paths are impaired or congested |
US8670313B2 (en) | 2006-08-22 | 2014-03-11 | Centurylink Intellectual Property Llc | System and method for adjusting the window size of a TCP packet through network elements |
US8576722B2 (en) | 2006-08-22 | 2013-11-05 | Centurylink Intellectual Property Llc | System and method for modifying connectivity fault management packets |
US8125897B2 (en) | 2006-08-22 | 2012-02-28 | Embarq Holdings Company Lp | System and method for monitoring and optimizing network performance with user datagram protocol network performance information packets |
US8130793B2 (en) | 2006-08-22 | 2012-03-06 | Embarq Holdings Company, Llc | System and method for enabling reciprocal billing for different types of communications over a packet network |
US8144586B2 (en) | 2006-08-22 | 2012-03-27 | Embarq Holdings Company, Llc | System and method for controlling network bandwidth with a connection admission control engine |
US8144587B2 (en) | 2006-08-22 | 2012-03-27 | Embarq Holdings Company, Llc | System and method for load balancing network resources using a connection admission control engine |
US8107366B2 (en) | 2006-08-22 | 2012-01-31 | Embarq Holdings Company, LP | System and method for using centralized network performance tables to manage network communications |
US9479341B2 (en) | 2006-08-22 | 2016-10-25 | Centurylink Intellectual Property Llc | System and method for initiating diagnostics on a packet network node |
US9253661B2 (en) | 2006-08-22 | 2016-02-02 | Centurylink Intellectual Property Llc | System and method for modifying connectivity fault management packets |
US8194555B2 (en) | 2006-08-22 | 2012-06-05 | Embarq Holdings Company, Llc | System and method for using distributed network performance information tables to manage network communications |
US8199653B2 (en) | 2006-08-22 | 2012-06-12 | Embarq Holdings Company, Llc | System and method for communicating network performance information over a packet network |
US8213366B2 (en) | 2006-08-22 | 2012-07-03 | Embarq Holdings Company, Llc | System and method for monitoring and optimizing network performance to a wireless device |
US8224255B2 (en) | 2006-08-22 | 2012-07-17 | Embarq Holdings Company, Llc | System and method for managing radio frequency windows |
US8223654B2 (en) | 2006-08-22 | 2012-07-17 | Embarq Holdings Company, Llc | Application-specific integrated circuit for monitoring and optimizing interlayer network performance |
US8228791B2 (en) | 2006-08-22 | 2012-07-24 | Embarq Holdings Company, Llc | System and method for routing communications between packet networks based on intercarrier agreements |
US8238253B2 (en) * | 2006-08-22 | 2012-08-07 | Embarq Holdings Company, Llc | System and method for monitoring interlayer devices and optimizing network performance |
US20080049631A1 (en) * | 2006-08-22 | 2008-02-28 | Morrill Robert J | System and method for monitoring interlayer devices and optimizing network performance |
US8274905B2 (en) | 2006-08-22 | 2012-09-25 | Embarq Holdings Company, Llc | System and method for displaying a graph representative of network performance over a time period |
US9602265B2 (en) | 2006-08-22 | 2017-03-21 | Centurylink Intellectual Property Llc | System and method for handling communications requests |
US20120269089A1 (en) * | 2006-08-22 | 2012-10-25 | Morrill Robert J | System and method for monitoring interlayer devices and optimizing network performance |
US10469385B2 (en) | 2006-08-22 | 2019-11-05 | Centurylink Intellectual Property Llc | System and method for improving network performance using a connection admission control engine |
US8307065B2 (en) | 2006-08-22 | 2012-11-06 | Centurylink Intellectual Property Llc | System and method for remotely controlling network operators |
US9241277B2 (en) | 2006-08-22 | 2016-01-19 | Centurylink Intellectual Property Llc | System and method for monitoring and optimizing network performance to a wireless device |
US8358580B2 (en) | 2006-08-22 | 2013-01-22 | Centurylink Intellectual Property Llc | System and method for adjusting the window size of a TCP packet through network elements |
US8374090B2 (en) | 2006-08-22 | 2013-02-12 | Centurylink Intellectual Property Llc | System and method for routing data on a packet network |
US8407765B2 (en) | 2006-08-22 | 2013-03-26 | Centurylink Intellectual Property Llc | System and method for restricting access to network performance information tables |
US8472326B2 (en) * | 2006-08-22 | 2013-06-25 | Centurylink Intellectual Property Llc | System and method for monitoring interlayer devices and optimizing network performance |
US8102770B2 (en) | 2006-08-22 | 2012-01-24 | Embarq Holdings Company, LP | System and method for monitoring and optimizing network performance with vector performance tables and engines |
US8488495B2 (en) | 2006-08-22 | 2013-07-16 | Centurylink Intellectual Property Llc | System and method for routing communications between packet networks based on real time pricing |
US8098579B2 (en) | 2006-08-22 | 2012-01-17 | Embarq Holdings Company, LP | System and method for adjusting the window size of a TCP packet through remote network elements |
US8509082B2 (en) | 2006-08-22 | 2013-08-13 | Centurylink Intellectual Property Llc | System and method for load balancing network resources using a connection admission control engine |
US8520603B2 (en) | 2006-08-22 | 2013-08-27 | Centurylink Intellectual Property Llc | System and method for monitoring and optimizing network performance to a wireless device |
US10298476B2 (en) | 2006-08-22 | 2019-05-21 | Centurylink Intellectual Property Llc | System and method for tracking application resource usage |
US8531954B2 (en) | 2006-08-22 | 2013-09-10 | Centurylink Intellectual Property Llc | System and method for handling reservation requests with a connection admission control engine |
US8537695B2 (en) | 2006-08-22 | 2013-09-17 | Centurylink Intellectual Property Llc | System and method for establishing a call being received by a trunk on a packet network |
US8549405B2 (en) | 2006-08-22 | 2013-10-01 | Centurylink Intellectual Property Llc | System and method for displaying a graphical representation of a network to identify nodes and node segments on the network that are not operating normally |
US10075351B2 (en) | 2006-08-22 | 2018-09-11 | Centurylink Intellectual Property Llc | System and method for improving network performance |
US9241271B2 (en) | 2006-08-22 | 2016-01-19 | Centurylink Intellectual Property Llc | System and method for restricting access to network performance information |
US8619820B2 (en) | 2006-08-22 | 2013-12-31 | Centurylink Intellectual Property Llc | System and method for enabling communications over a number of packet networks |
US8619596B2 (en) | 2006-08-22 | 2013-12-31 | Centurylink Intellectual Property Llc | System and method for using centralized network performance tables to manage network communications |
US8619600B2 (en) | 2006-08-22 | 2013-12-31 | Centurylink Intellectual Property Llc | System and method for establishing calls over a call path having best path metrics |
US8064391B2 (en) | 2006-08-22 | 2011-11-22 | Embarq Holdings Company, Llc | System and method for monitoring and optimizing network performance to a wireless device |
US8687614B2 (en) | 2006-08-22 | 2014-04-01 | Centurylink Intellectual Property Llc | System and method for adjusting radio frequency parameters |
US8040811B2 (en) | 2006-08-22 | 2011-10-18 | Embarq Holdings Company, Llc | System and method for collecting and managing network performance information |
US8743703B2 (en) | 2006-08-22 | 2014-06-03 | Centurylink Intellectual Property Llc | System and method for tracking application resource usage |
US8743700B2 (en) | 2006-08-22 | 2014-06-03 | Centurylink Intellectual Property Llc | System and method for provisioning resources of a packet network based on collected network performance information |
US8750158B2 (en) | 2006-08-22 | 2014-06-10 | Centurylink Intellectual Property Llc | System and method for differentiated billing |
US8811160B2 (en) | 2006-08-22 | 2014-08-19 | Centurylink Intellectual Property Llc | System and method for routing data on a packet network |
US9992348B2 (en) | 2006-08-22 | 2018-06-05 | Century Link Intellectual Property LLC | System and method for establishing a call on a packet network |
US9929923B2 (en) | 2006-08-22 | 2018-03-27 | Centurylink Intellectual Property Llc | System and method for provisioning resources of a packet network based on collected network performance information |
US9832090B2 (en) | 2006-08-22 | 2017-11-28 | Centurylink Intellectual Property Llc | System, method for compiling network performancing information for communications with customer premise equipment |
US9813320B2 (en) | 2006-08-22 | 2017-11-07 | Centurylink Intellectual Property Llc | System and method for generating a graphical user interface representative of network performance |
US9806972B2 (en) | 2006-08-22 | 2017-10-31 | Centurylink Intellectual Property Llc | System and method for monitoring and altering performance of a packet network |
US8015294B2 (en) | 2006-08-22 | 2011-09-06 | Embarq Holdings Company, LP | Pin-hole firewall for communicating data packets on a packet network |
US9014204B2 (en) | 2006-08-22 | 2015-04-21 | Centurylink Intellectual Property Llc | System and method for managing network communications |
US9042370B2 (en) | 2006-08-22 | 2015-05-26 | Centurylink Intellectual Property Llc | System and method for establishing calls over a call path having best path metrics |
US9712445B2 (en) | 2006-08-22 | 2017-07-18 | Centurylink Intellectual Property Llc | System and method for routing data on a packet network |
US9054986B2 (en) | 2006-08-22 | 2015-06-09 | Centurylink Intellectual Property Llc | System and method for enabling communications over a number of packet networks |
US9660917B2 (en) | 2006-08-22 | 2017-05-23 | Centurylink Intellectual Property Llc | System and method for remotely controlling network operators |
US9094261B2 (en) | 2006-08-22 | 2015-07-28 | Centurylink Intellectual Property Llc | System and method for establishing a call being received by a trunk on a packet network |
US7940735B2 (en) | 2006-08-22 | 2011-05-10 | Embarq Holdings Company, Llc | System and method for selecting an access point |
US9112734B2 (en) | 2006-08-22 | 2015-08-18 | Centurylink Intellectual Property Llc | System and method for generating a graphical user interface representative of network performance |
US7843831B2 (en) | 2006-08-22 | 2010-11-30 | Embarq Holdings Company Llc | System and method for routing data on a packet network |
US9240906B2 (en) | 2006-08-22 | 2016-01-19 | Centurylink Intellectual Property Llc | System and method for monitoring and altering performance of a packet network |
US9661514B2 (en) | 2006-08-22 | 2017-05-23 | Centurylink Intellectual Property Llc | System and method for adjusting communication parameters |
US9621361B2 (en) | 2006-08-22 | 2017-04-11 | Centurylink Intellectual Property Llc | Pin-hole firewall for communicating data packets on a packet network |
US9225609B2 (en) | 2006-08-22 | 2015-12-29 | Centurylink Intellectual Property Llc | System and method for remotely controlling network operators |
US9225646B2 (en) | 2006-08-22 | 2015-12-29 | Centurylink Intellectual Property Llc | System and method for improving network performance using a connection admission control engine |
US8289965B2 (en) | 2006-10-19 | 2012-10-16 | Embarq Holdings Company, Llc | System and method for establishing a communications session with an end-user based on the state of a network connection |
US8194643B2 (en) | 2006-10-19 | 2012-06-05 | Embarq Holdings Company, Llc | System and method for monitoring the connection of an end-user to a remote network |
US8189468B2 (en) | 2006-10-25 | 2012-05-29 | Embarq Holdings, Company, LLC | System and method for regulating messages between networks |
US9521150B2 (en) | 2006-10-25 | 2016-12-13 | Centurylink Intellectual Property Llc | System and method for automatically regulating messages between networks |
US20100296390A1 (en) * | 2006-11-27 | 2010-11-25 | Honeywell International Inc. | Fault tolerance in wireless networks |
US8340008B2 (en) * | 2006-11-27 | 2012-12-25 | Honeywell International Inc. | Fault tolerance in wireless networks |
US8111692B2 (en) | 2007-05-31 | 2012-02-07 | Embarq Holdings Company Llc | System and method for modifying network traffic |
EP2243255A1 (en) * | 2008-02-01 | 2010-10-27 | Allied Telesis Holdings K.K. | Method and system for dynamic link failover management |
EP2243255A4 (en) * | 2008-02-01 | 2012-08-29 | Allied Telesis Holdings Kk | Method and system for dynamic link failover management |
WO2009097615A1 (en) * | 2008-02-01 | 2009-08-06 | Allied Telesis Holdings Kabushiki Kaisha | Method and system for dynamic link failover management |
US20090249115A1 (en) * | 2008-02-01 | 2009-10-01 | Allied Telesis Holdings Kabushiki Kaisha | Method and system for dynamic link failover management |
US8068425B2 (en) | 2008-04-09 | 2011-11-29 | Embarq Holdings Company, Llc | System and method for using network performance information to determine improved measures of path states |
US8879391B2 (en) | 2008-04-09 | 2014-11-04 | Centurylink Intellectual Property Llc | System and method for using network derivations to determine path states |
US20100020680A1 (en) * | 2008-07-28 | 2010-01-28 | Salam Samer M | Multi-chassis ethernet link aggregation |
US8300523B2 (en) | 2008-07-28 | 2012-10-30 | Cisco Technology, Inc. | Multi-chasis ethernet link aggregation |
US20110211585A1 (en) * | 2010-02-26 | 2011-09-01 | Alaxala Networks Corporation | Packet relay device |
US8971337B2 (en) * | 2010-02-26 | 2015-03-03 | Alaxala Networks Corporation | Packet relay device |
US20130232277A1 (en) * | 2011-08-31 | 2013-09-05 | Metaswitch Networks Ltd. | Transmitting and Forwarding Data |
US9246796B2 (en) * | 2011-08-31 | 2016-01-26 | Metaswitch Networks Ltd | Transmitting and forwarding data |
US9197547B2 (en) * | 2012-06-01 | 2015-11-24 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of MoFRR with dataplane notifications |
US9628285B2 (en) * | 2012-06-01 | 2017-04-18 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of MoFRR with dataplane notifications |
US9736061B2 (en) | 2012-06-01 | 2017-08-15 | Telefonaktiebolaget L M Ericsson (Publ) | Enhancements to PIM fast re-route with upstream activation packets |
US20150172070A1 (en) * | 2012-06-01 | 2015-06-18 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of mofrr with dataplane notifications |
US8824276B2 (en) * | 2012-06-01 | 2014-09-02 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of MOFRR with dataplane notifications |
US8913482B2 (en) | 2012-06-01 | 2014-12-16 | Telefonaktiebolaget L M Ericsson (Publ) | Enhancements to PIM fast re-route with upstream activation packets |
US20140355422A1 (en) * | 2012-06-01 | 2014-12-04 | Telefonaktiebolaget L M Ericsson (Publ) | Increasing failure coverage of mofrr with dataplane notifications |
US20150295861A1 (en) * | 2013-01-06 | 2015-10-15 | Hangzhou H3C Technologies Co., Ltd. | Link problem handling |
US9559989B2 (en) * | 2013-01-06 | 2017-01-31 | Hewlett Packard Enterprise Development Lp | Link problem handling |
US20190028299A1 (en) * | 2016-03-18 | 2019-01-24 | Hewlett Packard Enterprise Development Lp | Responses to loops in networks having a ring topology |
WO2017160306A1 (en) * | 2016-03-18 | 2017-09-21 | Hewlett Packard Enterprise Development Lp | Responses to loops in networks having a ring topology |
US10812289B2 (en) | 2016-03-18 | 2020-10-20 | Hewlett Packard Enterprise Development Lp | Responses to loops in networks having a ring topology |
US11551684B1 (en) * | 2017-12-08 | 2023-01-10 | Amazon Technologies, Inc. | State detection and responses for electronic devices |
US11012369B2 (en) * | 2019-07-05 | 2021-05-18 | Dell Products L.P. | Aggregated switch path optimization system |
CN114679348A (en) * | 2020-12-24 | 2022-06-28 | 华为技术有限公司 | Fault processing method and related product |
Also Published As
Publication number | Publication date |
---|---|
US8886831B2 (en) | 2014-11-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8886831B2 (en) | System and methodology for fast link failover based on remote upstream failures | |
US8117337B2 (en) | Method and device for implementing link pass through in point-to-multipoint network | |
US8804534B2 (en) | Interworking between MPLS/IP and Ethernet OAM mechanisms | |
EP2087712B1 (en) | Method and apparatus for computing alternate multicast/broadcast paths in a routed network | |
US9203732B2 (en) | Recovery of traffic in a connection-oriented network | |
EP1919138B1 (en) | A method for implementing backup of the uplink | |
JP4884184B2 (en) | Communication device and program for link aggregation | |
EP1895724A1 (en) | A method for implementing active/standby gateway device in the network and a system thereof | |
EP1958364B1 (en) | Vpls remote failure indication | |
US20080112333A1 (en) | Communicating an operational state of a transport service | |
US9385944B2 (en) | Communication system, path switching method and communication device | |
KR102088298B1 (en) | Method and appratus for protection switching in packet transport system | |
JP2007088949A (en) | Information processing apparatus, communication load diffusing method and communication load diffusion program | |
JP2008078893A (en) | Redundant method for network and medium switching equipment | |
JP2007525895A (en) | Recovery mechanism for network topology | |
EP3029883B1 (en) | Network protection method and apparatus, next-ring node, and system | |
EP2254289B1 (en) | Method, device, and system for establishing label switching path in fast rerouting switching | |
CN102857316B (en) | A kind of method and system realizing source looped network protection | |
US7174389B2 (en) | Tandem node system and a method therefor | |
CN102882779A (en) | VRRP (Virtual Router Redundancy Protocol) advertisement link protection method and system | |
WO2011011934A1 (en) | Method and apparatus for ethernet tunnel segmentation protection | |
CN113037622B (en) | System and method for preventing BFD from vibrating | |
JP5612560B2 (en) | Network relay device and network relay system | |
JP2003338831A (en) | Switching method for mpls network | |
JP4035820B2 (en) | Line switching apparatus, data transmission system, and line switching method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CISCO TECHNOLOGY, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TIRUMALAI, GOKUL S.;MITCHELL, ALBERT HYDE;REEL/FRAME:017762/0523 Effective date: 20060405 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20221111 |