Sei sulla pagina 1di 53

Design Guide

Cisco FabricPath Design Guide: Using FabricPath with an Aggregation and Access Topology
December, 2011

For further information, questions and comments please contact ccbu-pricing@cisco.com

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 1 of 54

Introduction.............................................................................................................................................................. 3 Hardware and Software Support ............................................................................................................................ 3 Next-Generation Data Center Architecture............................................................................................................ 4 FabricPath Technology Overview .......................................................................................................................... 7 Spine and Edge ................................................................................................................................................... 8 Switch-ids ............................................................................................................................................................ 9 Conversational Learning .................................................................................................................................. 11 VLAN Trunking .................................................................................................................................................. 11 Metrics ............................................................................................................................................................... 11 Multipath Load Balancing ................................................................................................................................ 12 Multidestination Trees ...................................................................................................................................... 12 Multicast Forwarding ........................................................................................................................................ 14 vPC+ ................................................................................................................................................................... 14 Design Considerations for the Spine/Aggregation Layer .................................................................................. 14 Topology Considerations ................................................................................................................................. 16 Declaring VLANs as FabricPath VLANs ..................................................................................................... 16 Declaring Ports as FabricPath Ports .......................................................................................................... 17 Topology 0 .................................................................................................................................................... 17 Unicast Layer 2 Multipathing ...................................................................................................................... 17 Building a Routed Spine .................................................................................................................................. 17 F1 and M1 Card Cooperation ...................................................................................................................... 17 M1 Card Routing Traffic............................................................................................................................... 18 F1 Card Architecture and MAC Optimization............................................................................................. 18 MAC Scalability Considerations ................................................................................................................. 20 Gateway Routing ............................................................................................................................................... 22 Active/Active Gateway Routing ................................................................................................................... 22 Benefits of vPC+ at the Spine ..................................................................................................................... 24 Dual-Active Exclude (Peer-Link Failure) .................................................................................................. 24 Connecting FabricPath Edge or Leaf Layer to the Spine Layer.................................................................... 25 Avoiding Flooding by Tuning ARP and Layer 2 Table ................................................................................... 25 Multicast Routing Considerations ................................................................................................................... 25 Multicast Forwarding in FabricPath ........................................................................................................... 26 Routed Multicast Configuration .................................................................................................................. 26 Routed Multicast with vPC+ ........................................................................................................................ 27 Summary Recommendations........................................................................................................................... 28 Design Considerations for the Edge and Access Layer .................................................................................... 28 FabricPath VLANs ............................................................................................................................................. 29 MAC Learning at the Access Layer ................................................................................................................. 29 Learning the HSRP MAC Address ................................................................................................................... 29 Direct Connectivity between Edge Switches.................................................................................................. 30 vPC+ at the Edge .............................................................................................................................................. 30 Integration with FEX ......................................................................................................................................... 32 Summary............................................................................................................................................................ 33 FabricPath Scalability Considerations ................................................................................................................ 33 FabricPath Convergence Times ........................................................................................................................... 37 Sample Configurations ......................................................................................................................................... 38 Cisco Nexus 7000 Series Spine 1 (DC1-Agg1) ............................................................................................... 38 Cisco Nexus 7000 Series Spine 2 (DC1-Agg2) ............................................................................................... 42 Edge Cisco Nexus 5500 Switch (DC1-5500-1) without vPC+ ........................................................................ 46

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 2 of 54

Edge Cisco Nexus 5500 Switch (DC1-5500-1) with vPC+ and FEX A/A ....................................................... 47 Edge Cisco Nexus 5500 Switch (DC1-5500-2) with vPC+ and FEX A/A ....................................................... 50

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 3 of 54

Introduction
The introduction of Cisco FabricPath technology in Cisco NX-OS Software Release 5.1(3) brings the benefit of routing protocols to Layer 2 network Ethernet environments. Cisco FabricPath technology is described in several documents available on the cisco.com page: http://www.cisco.com/en/US/prod/switches/ps9441/fabric_path_promo.html http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9402/at_a_glance_c45-605626.pdf http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9402/white_paper_c11-605488.html http://www.cisco.com/en/US/docs/switches/datacenter/sw/5_x/nxos/fabricpath/configuration/guide/fp_cli_Book.html The advantages of using FabricPath include:

MAC address scalability with conversational learning Spanning Tree Protocol independence: No reliance on Spanning Tree anymore. Each switch has a complete view of the Layer 2 topology, and it calculates the Layer 2 forwarding table based a shortest-pathfirst calculation.

Traffic distribution for unicast: Unicast Layer 2 traffic can take multiple equal-cost Layer 2 paths. Traffic distribution for multicast on multiple distribution trees: Multicast traffic can be distributed along two multidestination trees.

More direct communication paths: Any topology is possible, so cabling two access or edge switches directly to each other creates a direct communication path, unlike what happens with Spanning Tree.

Simplicity of configuration: The configuration of FabricPath is very simple. Tuning may still be advised but by default the switches negotiate switch-ids allocation with a protocol called Dynamic Resource Allocation Protocol (DRAP).

Loop mitigation with TTL in the frame field: Layer 2 loops, as they are known of today in SpanningTree-Protocol-based Layer 2 networks, are mitigated by dropping frames that have been propagated across too many hops. The Layer 2 FabricPath frames include a Time to Live (TTL) field that is decremented at each hop. The value of the TTL field is 32.

This document describes one of many possible designs that are enabled by Cisco FabricPath, and it includes tuning recommendations and sample configurations.

Hardware and Software Support


The solution described in this document requires the following hardware and software:

Cisco Nexus 5500 products as the edge (or access) device in a FabricPath topology with NX-OS Software Release 5.1(3)N1(1). Note that in the context of FabricPath domain, the edge device is also known as the leaf device.

Cisco Nexus 7000 Series products as the spine (or aggregation) device in a FabricPath topology with NXOS 5.1(1) Release or later. At the time of this writing, the Cisco Nexus 7000 Series configuration that enables FabricPath is based on the F1 series cards, which consist of 32 ports 1/10 Gbps. The Layer 3 functionality is provided by the use of M1 series cards.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 4 of 54

Cisco Nexus Fabric Extenders (FEX) platforms are supported with the FEX ports acting as FabricPath edge ports. FEX can be used to augment the number of edge ports or Classic Ethernet (CE) ports without adding configuration points.

Proper licensing: You need the ENHANCED_LAYER2_PKG license installed. For details, visit: http://www.cisco.com/en/US/docs/switches/datacenter/sw/nx-os/licensing/guide/Cisco_NXOS_Licensing_Guide_chapter1.html.

Note:

The F1 cards are complemented by M1 card for routing purposes. When using M1 cards in the same

virtual device context (VDC) as the F1 card, routing is offloaded to the M1 cards, and more routing capacity is added to the F1 card by putting more M1 ports into the same VDC as the F1 card. This design guide refers to the following FabricPath features:

FabricPath Shortest-Path-First-based forwarding FabricPath Layer 2 multipathing (Unicast traffic) FabricPath multidestination trees load balancing (unknown unicast, broadcast, and multicast traffic) Active/Active gateway routing using vPC+ configuration vPC+ at the edge switch for the purpose of connecting dual-homed servers or dual-homed Cisco Fabric Extenders

Next-Generation Data Center Architecture


The introduction of FabricPath enables a variety of Layer 2 topologies that were not possible before. Figure 1 illustrates some of them.
Figure 1. FabricPath enables new topologies

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 5 of 54

On the top-left of Figure 1 (topology a), you see a design where two spines connect to a set of edge devices, which looks more like an aggregation/access design. However, in contrast to designs based on Spanning Tree Protocol, any two endpoints can take all available paths via either spine. The second design, topology b at the top-right, illustrates a topology where servers connect to edge devices in vPC mode, which in the case of FabricPath is called vPC+. The left-bottom design, topology c, illustrates the ability for FabricPath to use multiple spines instead of just two. This gives a very granular control on the available bandwidth: the more spines you add, the lower oversubscription between the edge and the rest of the network. The bottom-right design, topology d, illustrates the ability for FabricPath to provide direct connectivity between edge devices in case direct communication between any edge devices is suitable. Most of these topologies would not make sense with Spanning Tree, but they are almost plug-and-play topologies in a FabricPath environment. Note: Direct connectivity between the spine devices is not strictly necessary. It depends on where the routing is

performed and on whether you consider it necessary to establish a direct communication path between spines, whether spines are just used for fabric switching or are used for end node attachment too, and so on. Figure 2 illustrates the traffic paths that any two given servers can take across a Cisco FabricPath topology:
Figure 2. Traffic paths in FabricPath designs

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 6 of 54

The Layer 3 demarcation point can be placed in several ways, but a very simple deployment model is to place the Layer 3 boundary at the aggregation point where all edge or access switches are aggregated. Figure 3 illustrates adding the routing function at the spine layer for the FabricPath cloud to connect to the rest of the network infrastructure:
Figure 3. Placing the Router at the Spine Layer

When placing the routing at the spine, the spine keeps switching Layer 2 traffic based on the switch-id, but it also becomes an edge for routed traffic. Yet another option to place the routing function is to connect the routers as an edge device to the spine. This approach ensures that the spine switches traffic exclusively based on switch-id, and it makes changes to the spine configuration very rare. Figure 14 illustrates this design. Analyzing this last design option is outside of the scope of this document. In this design guide, the focus is on a topology that uses two spines and multiple edges. The reason for this is that its an easy transition for customers who are deploying Spanning Tree today and want to benefit from FabricPath without having to rewire their data center. In addition, it is easy to add additional spines in case you want to lower the oversubscription ratio, without any substantial change to the design considerations that are the subject of this document.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 7 of 54

Figure 4 describes the topology used in this design guide:


Figure 4. Topology used in this design guide

If you want, you can add ports to this design by adding Cisco Fabric Extenders (FEX), as shown in Figure 5:
Figure 5. Adding FEX to a FabricPath topology

FabricPath Technology Overview


This section provides an overview of FabricPath but it is not meant to be a tutorial on the technology. Several other white papers already exist on the Cisco public webpage that cover this topic. The reason that FabricPath is almost a plug-and-play protocol, in contrast to several other technologies available in the market, is that it enables each device to build an overall view of the topology in a way that is similar to Open Shortest Path First (OSPF) protocol and other link state protocols.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 8 of 54

Each device in the topology is identified by a switch-id and all Layer 2 forwarding tables are built based on reachability to each switch-id (as opposed to reachability to a MAC address). The switch-id is dynamically assigned via the Dynamic Resource Allocation Protocol (DRAP), so no manual allocation of switch-ids is required. However, it is possible to manually assign a SW-ID while a dynamic check for misconfiguration is performed (as discussed in the Switch-ids section). FabricPath multipath forwarding relies on building a view of the network using a Link State Routing Protocol similar to the one used in Layer 3 networks. This is achieved with a Layer 2 protocol based on Intermediate-System-toIntermediate-System (IS-IS) routing, which doesnt require any Layer 3 capability on the FabricPath switching devices.

Spine and Edge


In FabricPath topologies, there are two types of functions (which can be performed by all FabricPath hardware):

Edge (or leaf) devices: These devices have ports connected to Classic Ethernet devices (servers, firewalls, router ports, and so on) and ports connected to the FabricPath cloud (or FabricPath ports). Edge devices are able to map a MAC address to the destination switch-id.

Spine devices: These devices exclusively interconnect edge devices. Spine devices switch exclusively based on the destination switch-id.

Figure 6.

Classic Ethernet ports in a FabricPath topology

Traffic from edge ports is looked up in the Layer 2 forwarding table, and encapsulated into the MAC-in-MAC frame whose destination switch-id is the switch that the destination host is attached to. This allows sending traffic along multiple paths with the same end-to-end total cost and provides the ability to construct practically any topology and be sure that traffic will take the most direct path (lowest cost) between any two devices.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 9 of 54

Switch-ids
In order to understand these concepts, lets consider the topology in Figure 7:
Figure 7. Example switch-id allocation

The following CLI capture from an edge device shows the output of the show fabric switch-id command. The output shows all the switch-ids present in the topology, and it highlights with a * the switch-id of the device that the user is on: DC1-5500-1# show fabricpath switch-id FABRICPATH SWITCH-ID TABLE Legend: * - this system ========================================================================= SWITCH-ID *111 112 113 121 122 SYSTEM-ID 0005.73d4.9141 0005.73fc.2001 0005.73fc.207c 0022.5579.d3c1 108c.cf18.0941 FLAGS Primary Primary Primary Primary Primary STATE Confirmed Confirmed Confirmed Confirmed Confirmed STATIC Yes Yes Yes Yes Yes EMULATED No No No No No ----------+----------------+------------+-----------+--------------------

The edge switch maintains a MAC address table with the mapping of MAC address to switch-ids as well as the local source MACs from its CE ports, as illustrated here: DC1-5500-1# show mac address-table vlan 101 Legend: * - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC age - seconds since last seen,+ - primary entry using vPC Peer-Link VLAN * 101 * 101 * 101 * 101 MAC Address 0000.0700.0000 0000.0700.0100 0022.5579.d3c1 108c.cf18.0941 Type dynamic dynamic dynamic dynamic age 0 10 10 0 Secure NTFY F F F F F F F F Ports Eth1/5 112 121 122 ---------+-----------------+--------+---------+------+----+------------------

As you can see from the output, the locally attached MAC addresses are associated with the local port. The remote MAC addresses are associated with the switch-id of the destination. On the spine devices, there is no MAC address table learning, so the forwarding table is purely switch-id-based. In the following example, you can see each switch-id and which port can be used to reach the destination: FabricPath Unicast Route Table for Topology-Default

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 10 of 54

0/121/0, number of next-hops: 0 via ----, [60/0], 3 day/s 08:11:51, local 1/111/0, number of next-hops: 1 via Eth8/3, [115/40], 0 day/s 06:15:55, isis_fabricpath-default 1/112/0, number of next-hops: 1 via Eth8/5, [115/40], 3 day/s 08:09:17, isis_fabricpath-default [] Figure 8 illustrates how traffic is forwarded from MAC address A to MAC address B in a Cisco FabricPath topology:
Figure 8. FabricPath forwarding

Notice how the switch-id values are assigned by default to each FabricPath device via the DRAP protocol. However, it may be good practice to manually assign them in order to provide a more meaningful numbering scheme. In this latter scenario, a specific check is performed every time a new switch-id value is configured on a given FabricPath node, in order to prevent you from configuring a duplicate switch-id within the same domain. However, if two FabricPath switches from two different and isolated FabricPath domains converge into a single FabricPath domain, a mechanism intervenes to resolve conflicts: DC1-5500-1# 2011 Aug 17 15:32:57 DC1-5500-1 %$ VDC-1 %$ fabricpath %FABRICPATH-2FABRICPATH_LINK_BRINGUP_STALLED_STATIC: Link bringup stalled due to conflicts DC1-5500-1# show fabricpath conflict all ==================================================== Fabricpath Port State Table Port port-channel2 State Suspended due to conflicts ---------------+-------------------------+---------============================================== Fabricpath Conflicts SYSTEM-ID 0005.73d4.9141 108c.cf18.0941 SWITCH-ID 3 3 STATIC Yes Yes ---------------+--------------+---------------

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 11 of 54

Conversational Learning
Edge switches learn MAC addresses only if they have a node connected on the edge port or if the switch is a remote MAC and is in active conversation with a known MAC address connected to an edge port. In other words, a simple flood of traffic (either unknown unicast or broadcast) from a remote switch is not going to populate the MAC address table of an edge switch.

VLAN Trunking
In spanning tree networks, the user has to specify which VLANs belong to which port by using the switchport trunk allowed vlan command. With FabricPath, the user doesnt need to explicitly specify which VLANs are carried on a FabricPath-enabled link (VLANs obviously must be defined on the switches). The configuration of a FabricPath core port is performed with the command switchport mode fabricpath. The command switchport mode trunk puts a port into Classic Ethernet mode instead of FabricPath mode.

Metrics
The preferred path to any switch-id is calculated based on the metric to any given destination. The metric is as follows:

1-Gbps Ethernet links have a cost of 400 10-Gigabit Ethernet links have a cost of 40 20-Gbps have a cost of 20

Figure 9 illustrates the use of the linkcost in a real topology:


Figure 9. Metrics calculation in FabricPath

The following code shows the cost of fabricpath links from the switch 112 from Figure 9. DC1-5500-2# show fabricpath isis interface br Fabricpath IS-IS domain: default Interface Ethernet1/1 Ethernet1/2 Type P2P P2P Idx State 1 2 3 Up/Ready Up/Ready Up/Ready Circuit 0x01/L1 0x01/L1 0x01/L1 MTU Metric Priority 64 64 64 Adjs/AdjsUp 1/1 1/1 1/1 -------------------------------------------------------------------------------1500 40 1500 40 1500 400

Ethernet1/11 P2P

Multipath Load Balancing


Figure 10 illustrates unicast load balancing in a FabricPath network:

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 12 of 54

Figure 10.

Unicast load balancing in FabricPath

In this example, each edge device has two equal-cost-path routes to each others spine: DC1-5500-2: show fabricpath route: 1/113/0, number of next-hops: 2 via Eth1/1, [115/80], 3 day/s 19:36:09, isis_fabricpath-default via Eth1/2, [115/80], 0 day/s 18:40:16, isis_fabricpath-default The load balancing is based on fields such as the IP address, Layer 4 port, and so on, and it is configured with the following command: DC1-5500-2(config)# fabricpath load-balance unicast layer-3, layer4, mixed, source, source-destination> < destination, include-vlan,

Multidestination Trees
In FabricPath multicast traffic, broadcast and flooded traffic is forwarded along a multidestination tree. FabricPath allows for multiple multidestination trees in order to achieve traffic load balancing for multidestination frames. This is achieved by using a special Tag present in FabricPath frames, which is called Forwarding Tag or FTag. The FTag, together with the type of the frames destination address (unknown unicast, broadcast, or multicast), selects a forwarding graph, which is the set of possible switches and interfaces along which the frame is forwarded. At each switch, the FTag carried in the frame header is used to forward the frame along the interfaces that are in the forwarding graph. FabricPath uses a link state protocol to determine the forwarding trees in the network. FabricPath technology allows the definition of multiple topologies. At the time of this writing, only one topology is supported, in which there are two multidestination trees. As Figure 11 illustrates, two FTags are used: FTag1 and FTag2. In this example, each one of these multidestination trees is rooted at one of the spines.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 13 of 54

Figure 11.

FTag trees assignment in FabricPath

FTag1 is used for the first multidestination tree and FTag2 is used for the second multidestination tree. FTag trees are used as follows:

FTag1 is used for unknown unicast, broadcast, and multicast. The highest priority switch in the FabricPath topology is chosen as the root for FTag1. As discussed later in the guide, it is possible to modify the default FabricPath root priority value in order to place the root for FTag1 or FTag2 on a given device in a deterministic fashion. If the default priority is maintained on all the switches as part of the FabricPath domain, the switch system-id value is used as a tiebreak to determine the switch with the highest root priority.

FTag2 is used only for multicast traffic. The second highest priority switch in the FabricPath topology is chosen as the root for FTag2.

Here are the factors, in order, that are used in order to elect the root for FTag1 and subsequently for FTag2:

Highest root priority Highest system-id Highest switch-id

Figure 12 illustrates the forwarding of flooded frames along the FTag1 tree:
Figure 12. Forwarding of unknown unicast and broadcast frames

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 14 of 54

Multicast Forwarding
Multicast traffic forwarding uses FTag trees based both on load balancing of multicast frames and on tree pruning that is governed by Internet Group Management Protocol (IGMP) snooping (platform-specific load-balancing methods are discussed in the design sections of this guide). The edge device performs IGMP snooping and the member reports are conveyed in FabricPath via the Group-Membership Link State Protocol (GM-LSP). IGMP snooping and IS-IS work in conjunction to build per-VLAN multicast group-based trees in the FabricPath network.

vPC+
A virtual PortChannel (vPC) allows links that are physically connected to two different Cisco Nexus switches to appear as a single PortChannel to a third device. This provides a loop free topology eliminating the spanning-treeblocked ports and maximizing the bandwidth usage. In a FabricPath network, a host or legacy (not FabricPathenabled) Ethernet switch can be connected through a port channel to two FabricPath edge switches by using a configuration construct called emulated switch. The emulated switch implementations in FabricPath, where two FabricPath edge switches provide a vPC to a third-party device, is called vPC+. Emulated switch is a construct in which two FabricPath switches emulate a single switch to the rest of the FabricPath network. The packets originated by the two emulated switches are sourced with the emulated switchid. The other FabricPath switches are not aware of this and simply see the emulated switch (identified by a dedicated switch-id value) as reachable through both switches. This means that the two emulated switches have to be directly connected via peer link, and there should be a peer-keepalive path between the two switches to form the vPC+. Other than the fact that a peer-link is used to synchronize MAC addresses between vPC+ peers, the peer-link is a regular FabricPath link, and as such it can be used by orphan ports for direct communication. The following list provides the reference terminology for vPC+ components:

vPC+: The functionality enabled on two independent upstream devices to make them appear as a single logical device to the downstream switches, allowing the establishment of PortChannel connections. All the MAC addresses learned from devices locally attached to downstream devices are advertised to the FabricPath domain as connected to the defined emulated switch.

vPC+ peer device: One of a pair of devices that are connected with the special port channel known as the vPC+ peer link.

vPC+ peer link: The link used to synchronize states between the vPC+ peer devices. Both ends must be on 10 Gigabit Ethernet interfaces.

vPC+ domain: This domain is formed by the two vPC+ peer link devices. It is also a configuration mode for configuring some of the vPC+ peer link parameters.

vPC+ peer-keepalive link: The peer-keepalive link, a Layer 3 link between the vPC+ peer devices, is used to ensure that both devices are up. The peer-keepalive link sends periodic keepalive messages between the vPC peers on an out-of-band path.

vPC+ member port: Interface that belongs to the vPC+.

Design Considerations for the Spine/Aggregation Layer


The specific topology discussed and validated in the context of this design guide consists of a pair of spine devices connected to multiple edge devices. Cisco Nexus 7000 Series platforms are used as spine devices, whereas Cisco Nexus 5500 Switches are connected at the edge.
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 15 of 54

The spine in this design is also used to perform the routing function between the FabricPath cloud and the rest of the network. The default gateway for the servers is located in the spine layer, which, being also the router in the topology, performs the function of an edge in FabricPath. At the time of this writing, the Cisco Nexus 7000 Series hardware combination that provides this capability consists of F1 cards (spine function) together with the M1 cards (routing function). In the topology shown in Figure 13, the spine layer, which is also an aggregation layer since it aggregates multiple edge switches, operates simultaneously in two modes:

It forwards Layer 2 traffic between the edge devices exclusively based on the destination switch-id, without the need to learn MAC addresses (east-to-west traffic).

It also learns MAC addresses exclusively for the purpose of encapsulating routed traffic into FabricPath frames (north-to-south traffic).

Figure 13.

Layer 3 Aggregation/Spine

These two behaviors are completely independent, in that even if, as a theoretical example, the 128-K adjacency table of the M1 card fully filled up, the E-W Layer 2 traffic would still be forwarded based on the switch-id and not flooded. Compared with a spanning-tree design or a vPC design, this design includes the following advantages:

Ease of configuration Multipath forwarding for unicast and multicast Layer 2 and Layer 3 traffic Faster convergence times Loop mitigation thanks to the use of a Time to Live (TTL) field in FabricPath frames

The focus of this section is on the design of this hybrid Layer 2/Layer 3 spine/aggregation layer. This is not the only possible design option. For instance, you can also add Layer 3 to an edge switch and use the spine as a pure Layer 2 spine, or connect a pair of Layer 2 edge switches to dedicated routing devices, as highlighted in Figure 14. It is outside of the scope for this document to discuss these specific design options further.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 16 of 54

Figure 14.

Fabricpath design with separate L3 edge

Topology Considerations
FabricPath technology allows building multiple logical topologies on the same physical topology. At the time of this writing, this functionality has not yet been enabled by the software, so you should not attempt to create additional topologies besides the default topology 0. Declaring VLANs as FabricPath VLANs All VLANs that are meant to be forwarded according to Cisco FabricPath rules must be created as FabricPath VLANs. A VLAN needs to be explicitly configured for Classic Ethernet (CE) mode or for FabricPath mode. By default, all VLANs are in CE mode. These CE VLANs are not brought up on FabricPath core links. They can be forwarded on the FabricPath edge ports locally or to other Classic Ethernet switches. The command required to configure a VLAN in FabricPath mode is: DC1-Agg1(config)# vlan 101 DC1-Agg1(config-vlan)# mode fabricpath

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 17 of 54

The FabricPath VLANs are forwarded on FabricPath core links using MAC-in-MAC header encapsulation and on FabricPath edge links without the MAC-in-MAC header. FabricPath VLANs can be grouped into topologies for traffic engineering. By default, all FabricPath VLANs belong to the base topology 0. CE VLANs and FabricPath VLANs cannot share a link unless the link is a FabricPath edge port. Declaring Ports as FabricPath Ports All ports that are carrying FabricPath traffic are defined as type fabricpath. With FabricPath, you dont configure trunking and pruning of VLANs on FabricPath ports. The required configuration to declare a port as a FabricPath port is very simple: DC1-Agg1(config)# interface ethernet 8/3 DC1-Agg1(config-if)# description FP Link to DC1-5500-1 DC1-Agg1(config-if)# switchport mode fabricpath DC1-Agg1(config-if)# no shutdown Topology 0 Topology 0 is the default topology and cannot be removed; it exists by default: DC1-Agg1(config)# fabricpath topology ? <1-63> <0-63> summary Fabricpath Topology ID 1-63 Specific topology information Display summary topology information DC1-Agg1# show fabricpath isis topology ?

The only configuration that you may want to modify in order to optimize traffic distribution is the root choice for the multidestination trees. Unicast Layer 2 Multipathing Unicast Layer 2 multipathing is on by default, but you can select the load balancing mechanism (by default it uses Layer 2, Layer 3, or Layer 4 source and destination IP and VLAN) for Layer 2 traffic with the following command: fabricpath load-balance unicast < destination, include-vlan, layer-3, layer4, mixed, source, source-destination>

Building a Routed Spine


In this design recommendation, the spine also performs routing functions, and as a result the default gateway must be configured. When the spine is built with the Cisco Nexus 7000 Series product family, you need to understand the integration of F1 and M1 cards. This integration provides the combined benefits of a FabricPath switch for Layer 2 traffic and decapsulation and encapsulation of Layer 3 traffic into the FabricPath cloud, without a need for any external wiring or additional equipment. F1 and M1 Card Cooperation When an F1 card is used in a Cisco Nexus 7000 Series that provides the spine function (pure spine deployment), no remote MAC address learning occurs on the spine, as previously described. The F1 card doesnt do routing either. M1 cards instead provide the ability to perform routing between VLANs and toward the routed domain of the data center network.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 18 of 54

If an M1 card is present in the same virtual device context (VDC) as the F1 card, the F1 card will perform learning, but this doesnt mean that the F1 card forwards FabricPath traffic based on MAC addresses. The F1 card forwards the east-to-west traffic using the remote switch-id. In this configuration, MAC addresses are learned on the F1 card exclusively with the purpose of being able to encapsulate north-to-south routed traffic into FabricPath. In addition to this, the learning is optimized in the sense that each component of the F1 card learns only the locally connected MAC addresses. This optimization requires the use of FabricPath on the links connecting to the edge. The reason is that the edge encapsulates the frames with the destination switch-id so even if learning is performed, it is only for the purpose of routed traffic, not for the east-to-west traffic. In order to fully understand, this concept you need to consider the architecture of the F1 card, which is highly optimized for MAC scalability. M1 Card Routing Traffic Figure 15 illustrates the traffic flow for FabricPath traffic that is routed via the M1 card. Traffic routed from one FabricPath VLAN to another FabricPath VLAN enters via F1 cards, exists via F1 cards, and is routed by the M1 card. If traffic needs to be routed from the FabricPath cloud to a Layer 3 infrastructure, it enters from an F1 card and it exits via an M1 card port.
Figure 15. Traffic forwarding with M1/F1 linecards in a Nexus 7000 chassis

F1 Card Architecture and MAC Optimization East-to-west FabricPath traffic (or in other words pure Layer 2 traffic) doesnt engage the M1 card at all, nor does it require it. Pure Layer 2 traffic is handled by F1 cards and switched exclusively based on the switch-id information, as illustrated in Figure 16:

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 19 of 54

Figure 16.

Forwarding with L2 traffic in a mixed M1/F1 chassis

Because the M1 card is present and routed traffic requires being encapsulated into a FabricPath header, the F1 card also stores MAC address mappings to destination switch-ids; hence some learning occurs, which is not used for Layer 2 switched traffic. In order to understand the scalability of this solution, you have to consider the architecture of the F1 card, which is based on the Switch-on-a-Chip concept (SoC). Each SoC performs switching as it if was an independent switch, and forwarding tables are not synchronized. Therefore, by distributing edge switches on multiple SoCs, you can take advantage of the aggregated capacity of all forwarding tables on the SoCs (Figure 17).
Figure 17. High level view of ASICs arrangement in the F1 linecard

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 20 of 54

MAC Scalability Considerations In an M1+F1 solution and assuming that the edge switches are connected via FabricPath to the F1 card, each SoC learns the MAC addresses that are specific to it. Layer 2 traffic that doesnt leverage the default gateway doesnt require space in the Layer 2 table as it is switched exclusively based on the switch-id information. The MAC learning is required exclusively for routing purposes and Layer 2 switched traffic source MACs arent learned on the F1 card unless a switch virtual interface (SVI) is present on the VLAN. Since each SoC provides 16K MAC address capacity for the purpose of encapsulating routed traffic, the hardware could provide 256K MAC addresses capacity on the F1 card and 128K MAC addresses on the M1 card so the solution could accommodate up to 128K end devices routed by this combination of cards. A feature called mac-proxy (currently planned for CY12) could unleash this scalability. At the time of this writing, in certain failure scenarios, the traffic routed to the fabricpath cloud may use up to a maximum of 16k MAC addresses, after which traffic routed to any additional MAC would be flooded on one of the FTAG trees. Notice that this behavior has no impact for the Layer 2 east-to-west traffic because this one is switch exclusively based on the remote switch- ids. Figure 18 illustrates how the Nexus 7k builds the adjacency table for traffic routed to the Fabricpath cloud: 16k source MAC addresses sourced by Cisco Nexus 5500-1 are learned only by the SoC associated with port 8/3. The 8k source MAC addresses sourced by the Cisco Nexus 5500-2 are learned only by the SoC associated with port 8/5.
Figure 18. MAC address learning with M1/F1 linecards

The following CLI capture shows the SoC ASIC utilization with the traffic pattern described in Figure 18. L2 Forwarding Resources ----------------------L2 entries: Module inst 8 8 8 8 8 8 1 2 3 4 5 6 total 16384 16384 16384 16384 16384 16384 used 4 16002 8001 1 0 0 mcast 0 0 0 0 0 0 ucast 4 16002 8001 1 0 0 lines 1024 1024 1024 1024 1024 1024 lines_full 0 642 0 0 0 0

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 21 of 54

8 8 8

7 8 9

16384 16384 16384

0 0 2001

0 0 0

0 0 2001

1024 1024 1024

0 0 0

So what is the scalability in terms of MAC addresses on a FabricPath design? For traffic that is purely Layer 2 that is, traffic whose VLAN doesnt include an SVI - the scalability is exclusively dependent on the MAC table sizes of the edge devices in the topology; thus, the more edge devices, the more MAC entries you can switch. As an example of such VLANs, consider VLANs that are used to interconnect virtual machines (VMs) to create a virtual data center in a multitier environment. These VLANs often do not require any routing. Figure 19 illustrates a typical virtual data center (vDC): there can be several isolated networks and fenced networks (that is, connected via a gateway), so not all networks require a default gateway on the spine.
Figure 19. Example of a Vitual datacenter network

How many edge switches you can have in the same Layer 2 domain is bound to the maximum number of switchids that have been validated by testing. These numbers are published in the release notes or in the configuration limits in the documentation on the Cisco Nexus 5000 Series and Cisco Nexus 7000 Series. Note: At the time of this writing, this number varies from 64 to 128 switch-ids, depending on the software

release and the platform. For traffic whose VLANs are routed, the scalability is 16k-128k MAC addresses for routed traffic, which is provided by the M1/F1 combination and is dependent on the traffic patterns. Note: As previously mentioned at the time of this writing and with this specific design, should the links

connecting edge to spine break, forcing traffic through the asymmetric path connecting the spines, the number of routed MAC addresses is reduced to 16k. A functionality named mac-proxy, is on the roadmap for a future NX-OS SW release that will eliminating this caveat. Once this functionality is available (currently planned for 1HCY12), it would truly be possible to consider the MAC address scalability off a FabricPath domain as only dictated by the M1 MAC table space (minimum 128k) even in presence of multiple link failures.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 22 of 54

Notice that even if the total MAC count is exceeded, only routed traffic will experience flooding, while Layer 2 traffic will be switched based on the switch-id information. For this reason, it wont be experiencing flooding. For more information about scalability, see the FabricPath Scalability Considerations section at the end of this guide.

Gateway Routing
The gateway protocols work in FabricPath the same way they work in regular Ethernet networks. The default Active/Standby behavior of Hot Standby Router Protocol (HSRP) doesnt negate the value of FabricPath. Even without any further configuration than the standard configuration of HSRP, the Layer 2 traffic (east-to-west) would benefit from multipath forwarding. Routed traffic (south-to-north) would instead be forwarded only to the active HSRP device, as shown in Figure 20:
Figure 20. The default HSRP configuration doesnt make use of both spines

Active/Active Gateway Routing To perform traffic forwarding from multiple spines, you can use the Gateway Load Balancing Protocol (GLBP), which hands out a different gateway MAC (up to four different MACs) in a round-robin fashion. A better idea is to use the concept of vPC+ at the spine. The spine devices are connected by an additional FabricPath link (which would be recommended anyway to optimize multidestination tree forwarding) and by defining it as a vPC+ peer-link. The vPC+ peer-link must be built using F1 ports and not M1 ports, since it must be configured as a FabricPath link (Figure 21).

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 23 of 54

Figure 21.

vPC+ makes active and standby HSRP devices active

Figure 21 illustrates how traffic originating from the server connected to the edge and destined to its defaultgateway can use two equal cost paths. Notice that the Active and Standby HSRP roles refer only to the control plane, as both aggregation devices are able to actively route traffic on the data plane. The configuration to achieve this behavior is as simple as follows: vpc domain 1 role priority 110 peer-keepalive destination 10.60.17.141 fabricpath switch-id 1000 interface port-channel10 switchport mode fabricpath vpc peer-link Even if the configuration requires the definition of a vPC domain, there is no real vPC port in the design described in this topology. Notice that you need to configure a switch-id under the vPC domain configuration. This identifies the emulated switch. Because of this switch-id configuration, HSRP, VRRP, and GLBP announce their vMAC as coming from the emulated switch-id instead of the switch-id of each individual spine. By doing this, the edge switches can forward HSRP frames to either spine. On both spines, the HSRP MAC address is programmed in the MAC tables with the special flag (G-flag) that indicates that this traffic is meant to be routed. DC1-Agg1# show mac address-table Legend: * - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC age - seconds since last seen, + - primary entry using vPC Peer-Link VLAN G 101 MAC Address 0000.0c07.ac01 Type static age Secure NTFY Ports/SWID.SSID.LID F F sup-eth1(R) ---------+-----------------+--------+---------+------+----+------------------

Benefits of vPC+ at the Spine In this design guide, the use of vPC+ at the spine is strongly recommended. vPC+ gives you the ability to forward routed traffic to multiple routing engines as well as to optimize failover times.
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 24 of 54

As a result of this configuration, the edge switches learn the association of HSRP Virtual MAC addresses with the emulated switch-id instead of the individual spine switch-id: DC1-5500-1# show mac address-table | include 0c07 * 101 * 101 0000.0a00.0c07 0000.0c07.ac01 dynamic dynamic 10 0 F F F F Eth1/5 1000

As Figure 22 illustrates, the use of vPC+ allows the routed traffic between any two given endpoints to benefit from both Layer 2 equal-cost multipathing and from the aggregated routing capacity of the spines.
Figure 22. Traffic load balancing for both Layer 2 and Layer 3 traffic

The existence of a dual-active path from each edge device to the emulated switch can be verified by looking at the FabricPath routing table, as highlighted here: DC1-5500-1# sh fabricpath route FabricPath Unicast Route Table a/b/c denotes ftag/switch-id/subswitch-id [x/y] denotes [admin distance/metric] ftag 0 is local ftag subswitch-id 0 is default subswitch-id FabricPath Unicast Route Table for Topology-Default 1/1000/0, number of next-hops: 2 via Eth1/1, [115/40], 0 day/s 03:39:23, isis_fabricpath-default via Eth1/3, [115/40], 0 day/s 03:39:23, isis_fabricpath-default

Dual-Active Exclude (Peer-Link Failure)


As a result of declaring the link that connects the spines as a vPC peer-link, the default behavior of vPC applies whereby if the peer-link goes down, the SVIs on the vPC secondary device are shut down. In the context of FabricPath designs, this behavior is not beneficial, because the FabricPath links are still available, and theres no good reason to shut down the SVIs on the secondary. It is thus recommended to complete the vpc domain configuration with the following command: vpc domain 1 dual-active exclude vlan <SVIs associated with Fabricpath VLANs>

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 25 of 54

Connecting FabricPath Edge or Leaf Layer to the Spine Layer


When connecting the FabricPath edge layer to the spine layer, you can use any ports on the spine layer F1 cards unless the scalability of MAC addresses in the topology is a key design factor. In this case, maximizing the use of the SoC in the F1 card is beneficial. In order to do this, you just have to distribute edge switches in a way that they use different SoCs. As an example, you could connect edge-1 to ports 1 and 2 of the F1 card, edge-2 to ports 3 and 4, and so on (Figure 23).
Figure 23. Best way to connect edge to spine F1 card

Avoiding Flooding by Tuning ARP and Layer 2 Table


The default aging time on the Layer 2 forwarding table on the Cisco Nexus 7000 Series is 1800 seconds. The default Address Resolution Protocol (ARP) timeout is 1500 seconds. In order to avoid flooding routed traffic, it is beneficial to make sure that on the spine, the ARP timeout is more aggressive than the Layer 2 forwarding table timeout, which is normally the default. Tuning the MAC address table on the edge device is not necessary because in this architecture, the edge doesnt perform learning based on the ARP. The edge device in a FabricPath topology learns MAC addresses only for active conversations. In the spine, the presence of the M1 card in the same VDC as the F1 card requires the F1 card to learn MAC addresses for the sole purpose of rewriting the routed traffic with the FabricPath information. In this specific case, the SoCs learn MAC addresses based on ARP, but they dont use this MAC information for Layer 2 traffic forwarding.

Multicast Routing Considerations


Multicast routing and FabricPath switching do not require particular configuration or tuning on the aggregation layer devices. Only PIM may require tuning if you want to reduce failover times, just like you would do in regular routed designs. Multicast Forwarding in FabricPath As previously mentioned, FabricPath builds two multidestination trees with two different roots: one for FTag1 and one for FTag2. At Layer 2, multicast traffic is hashed to either tree in order to utilize both of them. The hashing to either multidestination tree is platform-dependent; for instance, it can include the VLAN field or the IP address fields. In order to maximize the efficiency of the traffic distribution, it is advisable to manually set the priority for the root for the FTag1 and FTag2 at the spine:
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 26 of 54

Agg1: fabricpath domain default root-priority 66 Agg2: fabricpath domain default root-priority 65 Traffic load-balancing can be configured with the following command: DC1-Agg1(config)# fabricpath load-balance multicast ? destination [[ source-destination symmetric xor Include source and destination parameters Symmetric (default) Include ex-or of source and destination parameters Include destination parameters

IGMP snooping is used to build the Layer 2 multicast MAC forwarding table at the edge. The information is used to prune traffic on a per-VLAN basis on the two multidestination trees. You can verify the traffic path for a given multicast group using the following command: DC1-Agg1# show fabricpath mroute (ftag/1, vlan/101, 0.0.0.0, 239.1.1.5), uptime: 00:30:24, isis Outgoing interface list: (count: 2) Interface port-channel10, uptime: 02:15:06, isis Interface port-channel10, uptime: 00:45:54, isis (ftag/1, vlan/101, 0.0.0.0, 239.1.1.6), uptime: 00:25:27, isis Outgoing interface list: (count: 1) Interface port-channel10, uptime: 00:45:54, isis Routed Multicast Configuration Its outside of the scope of this document to describe all the details of the routed multicast configuration, since it follows the same rules as in a non-FabricPath design:

Enable PIM under the SVI interface. IGMP is automatically enabled under the SVI when PIM is enabled. Define the rendezvous points (RP) in the network and/or configure the spine device for auto-RP. NX-OS provides a specific functionality, based on RFC 4610, called Anycast-RP Using Protocol Independent Multicast (PIM). This functionality makes it possible to provide RP redundancy in a simpler fashion than when deploying Multicast Source Discovery Protocol (MSDP).

The following is a sample configuration: ip pim anycast-rp 1.1.1.100 1.1.1.1 ip pim anycast-rp 1.1.1.100 1.1.1.2 ip pim rp-address 1.1.1.100 group-list 224.0.0.0/4 interface loopback100 description anycast-RP ip address 1.1.1.100/32 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 27 of 54

From an operational perspective, it may be advisable to align the PIM designated router (DR) priority with the HSRP primary. Note: The FTag selection for routed multicast traffic may change after routing. Post-routed traffic may be

forwarded on a different FTag tree than the prerouted traffic. PIM times are by default very conservative, and they are the cause of the slower convergence times that are measured for multicast traffic. You can bring them down significantly by either tuning the PIM timers or by using Bidirectional Forwarding Detection (BFD) for PIM. Here is an example of the configuration: bfd interval 50 min_rx 100 multiplier 3 interface Vlan101 [...] ip pim sparse-mode ip pim bfd-instance ip pim dr-priority 10 ip pim hello-interval 1000 Note: PIM interval tuning should be considered with care because when deployed on too many interfaces, it

may increase the CPU utilization. BFD scales better. For information on scalability, please refer to the release notes and/or the configuration limits. Routed Multicast with vPC+ When using vPC+ in conjunction with the FTag priority settings that were previously described, the hashing of routed multicast traffic is optimized, as illustrated in Figure 24.
Figure 24. Multicast routing with FabricPath

This means there is a sort of affinity behavior, based on which each vPC+ peer device will always use the multicast destination tree rooted on itself to forward multicast traffic.

Summary Recommendations
This section summarizes the recommendation for building a Layer 3 spine in a FabricPath topology. Figure 25 illustrates these recommendations.

Configure VLANs in FabricPath mode.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 28 of 54

Consider using manually configured switch-ids. This can help with managing and operating the topology since you can define the numbering scheme.

Configure SVIs and HSRP as usual. Configure vPC+ by defining a vPC peer-link and an emulated switch-id in the vPC domain. This is to allow integration of legacy (non-FabricPath-enabled) edge layer devices and to provide Active/Active defaultgateway data-plane functionality.

Configure Layer 3 peering between routers over a FabricPath VLAN (for instance, on the same link that is used to interconnect the spine devices, such as the vPC+ peer-link). This is required to reroute traffic destined to the northbound Layer 3 domain if one of the aggregation devices loses physical connectivity to the upstream devices.

Use FabricPath-enabled links for vPC peer-link purposes. Configure dual-active exclude vlan for the list of SVIs (unless using vPC attached devices). Distribute edge switches onto different SoCs on the F1 card for maximum MAC address scalability. Even if this is the default configuration, at the spine, you may want to verify that the ARP aging timeout is faster than the Layer 2 forwarding table timeout.

Figure 25.

Summary recommendations for FabricPath designs

Design Considerations for the Edge and Access Layer


The edge layer in FabricPath provides Classic Ethernet connectivity to servers and interfaces with FabricPath core ports to the spine layer. The edge layer performs unicast Layer 2 multipathing load balancing on equal-cost-path Layer 2 routes. Flooded and unknown unicast traffic is sent on the multidestination tree associated with FTag1. Multicast traffic is load balanced on the two multidestination trees (identified as usual with FTag1 and FTag2).

FabricPath VLANs
As with the spine layer, the access port in a FabricPath topology can be configured by assigning a FabricPath VLAN to an access port. You should start the configuration by making a VLAN a FabricPath type of VLAN: DC1-5500-1# show vlan id 101

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 29 of 54

VLAN Name 101 FP_VLAN_101 Vlan-mode FABRICPATH

Status active

Ports Eth1/5

---- -------------------------------- --------- ------------------------------VLAN Type 101 enet

---- ----- ----------

The ports connecting the edge device to the spine or even to another edge device should be configured as FabricPath core ports: (config-if)# switchport mode fabricpath The ports connecting to servers are configured as switchport mode access or trunk access. They can forward the traffic with the regular Ethernet encapsulation. To distribute unicast traffic via equal-cost multipath (ECMP), use the following command: DC1-5500-2(config)# fabricpath load-balance unicast layer-3, layer4, mixed, source, source-destination> Note: < destination, include-vlan,

By default, the unicast hashing is configured to include the source and destination IP addresses.

MAC Learning at the Access Layer


The edge layer by default performs conversational learning. Conversational learning spares MAC address table consumption at the edge because remote MAC addresses are not learned in the Layer 2 forwarding table unless theres an active conversation between a local node and a remote node. If you configure an SVI on a FabricPath VLAN, conversational learning is disabled on this VLAN. It is therefore normally recommended to not configure management SVIs on FabricPath VLANs.

Learning the HSRP MAC Address


The HSRP and GLBP MAC addresses - and in general the gateway vMAC addresses - are of particular interest because theres never an active conversation between an end host and a vMAC. The reason is that when the server sends traffic to the gateway, it does use the vMAC as the destination MAC address, but when the router sends traffic to the end host, it never uses the vMAC as source MAC address; instead, it uses the burned-in address (BIA). If the edge switch was not learning the gateway vMAC address, all traffic destined to be routed would be continuously flooded. The mechanism that solves this problem consists in considering HSRP, GLBP, or VRRP hellos as part of the conversation with an end host. When a host sends traffic destined to the gateway vMAC address, the edge switch stores it temporarily in the Layer 2 forwarding table. As soon as an HSRP hello is received, it programs the switch-id associated with it. If the spine is configured for vPC+, the source switch-id is the emulated switch-id. As a result, the MAC address table of the edge switch will have the HSRP MAC associated with the emulated switch-id. DC1-5500-2# show mac address-table vlan 101 Legend: * - primary entry, G - Gateway MAC, (R) - Routed MAC, O - Overlay MAC age - seconds since last seen,+ - primary entry using vPC Peer-Link VLAN * 101 MAC Address 0000.0c07.ac01 Type dynamic age 0 Secure NTFY F F Ports 1000
Page 30 of 54

---------+-----------------+--------+---------+------+----+----------

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Direct Connectivity between Edge Switches


One interesting point to observe is that edge switches can be connected directly to each other. This can be used to provide a direct path between hosts. Consider Figure 26:
Figure 26. Direct connectivity between edge switches allows direct communication between servers

When DC1-5500-2 to DC1-5500-3 is connected, the two hosts can talk directly over this link without having to go via the spine switches. This is unique to FabricPath. Spanning Tree Protocol would have blocked this link.

vPC+ at the Edge


The edge switches can be configured to perform vPC-like functions, which in the particular case of a FabricPath infrastructure are referred to as vPC+ (as already discussed for providing Active/Active default-gateway functionalities on the spine/aggregation devices). Functions that are enabled by vPC+ at the edge include:

The ability to attach servers to edge switches with port-channel teaming The ability to attach additional Classic Ethernet switches in vPC mode The ability to attach Cisco Fabric Extenders in FEX Active/Active mode.

The key requirement for a vPC+ configuration to be operational in FabricPath is to define the emulated switch-id in the vPC domain configuration. The other requirement is that the vPC peer-link must be a FabricPath link. All VLANs on vPC+ links must be FabricPath VLANs. When the peer-link goes down vPC+ operates similarly to vPC

When all the links of the peer-link fail, the primary switch will take no action, and the secondary switch will bring down all its vPC links.

In addition, vPC+ also intervenes to prevent attracting traffic to the vPC secondary switch. This function is provided by IS-IS by ensuring that the secondary switch stops advertising reach ability information to the emulated switch. This is essential to prevent the secondary switch from attracting traffic destined for the emulated switch from the FabricPath cloud.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 31 of 54

Figure 27 illustrates a design with vPC+:


Figure 27. Using vPC+ on the edge switches

In such a topology, the MAC addresses associated with a pair of edge devices running vPC+ are advertised to the other switches that are part of the same FabricPath domain as belonging to the emulated switch-id, instead of the individual switch-ids of the edge Cisco Nexus 5500 Switches. Overall, with vPC+ most of the communications between devices that connect to the vPC+ peers is kept local and doesnt involve the spines. This is true for both vPC connected devices and for orphan ports as well. Orphan ports in vPC+ can communicate over the peer-link as if this link were a regular FabricPath link (Figure 28):
Figure 28. Communication between devices connected to a vPC+ edge

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 32 of 54

In addition, even if the vPC+ peer-link were to fail, Host1 and Host4 would not lose connectivity because they would still be able to communicate via the spines. In Figure 28, Host1 and Host4 send traffic to the switch-id of the respective edge switches and not to the emulated switch-id. As for the north-to-south traffic flows destined to the hosts supported by vPC links, they are sent to the vPC primary only as the vPC secondary device stops advertising edge MAC address information associated to the emulated switch-id (Figure 29).
Figure 29. vPC+ behavior upon loss of the peer-link

Integration with FEX


You can extend FabricPath topologies by adding Cisco Fabric Extenders. Fabric extender ports are edge ports, and if you are using vPC+, they allow creating vPCs between servers and FEX. Dual-layer vPC is also implemented.

Summary
In summary, at the edge layer there is no major tuning necessary. The configuration is very simple:

You should configure VLANs as FabricPath VLANs. Consider using manually configured switch-ids. This can help with managing and operating the topology since you can define the numbering scheme.

To avoid disabling conversational learning, make sure not to have an SVI (like the management SVI) on a FabricPath VLAN.

Consider the use of vPC+ if using port-channel teaming or FEX Active/Active or adapter-FEX/VM-FEX.

FabricPath Scalability Considerations


In terms of FabricPath scalability, there are three main elements you should consider in a given FabricPath domain:

Number of switches Number of VLANs Number of MAC addresses

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 33 of 54

The maximum number of switches and VLANs supported (that is, tested by QA) in the same FabricPath domain is continuously increasing at every Cisco NX-OS Software release. At the time of writing of this document, these are the expected supported values:

Cisco Nexus 7000 Series (NX-OS Release 6.0): 128 switch-ids and between 2000 and 4000 VLANs Cisco Nexus 5500 Switches (NX-OS Release 5.1(3)N1(1)): 128 switch-ids and between 2000 and 4000 VLANs

Note: release.

For updated information on the maximum number of switch-ids supported within the same FabricPath

domain, please refer to the release notes and the configuration limits documentation for each specific software

One of the benefits of FabricPath is the scalability both in terms of MAC addresses and in terms of throughput/oversubscription ratios. In terms of MAC address scalability, we should differentiate between edge and spine devices. The edge switches learn only MAC addresses for which there are active conversations established (conversational learning). The focus of this document is on using the Cisco Nexus 5500 Switches at the edge of the FabricPath network. The current maximum value of MAC addresses supported on these devices is 24k. For the spine devices, it is important to distinguish between east-to-west (switched) and north-to-south (routed) traffic flows:

East-west traffic: These flows are switched by the spine devices based on the outer MAC header information (FabricPath switch-id values) (see Figure 30). For the spine devices to perform this switching, there is technically no need for them to learn the MAC addresses of the endpoints connected to the FabricPath edge switches.

Figure 30.

FabricPath spines switch traffic based on the Destination Switch-id (DSID)

North-south traffic: Traffic coming from the Layer 3 network domain and directed to an endpoint connected to a FabricPath edge switch needs to be first handled by the M1 card (performing routing functionalities) and eventually sent out to the FabricPath core through an F1 interface.

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 34 of 54

Figure 31.

Traffic routing by M1 card and subsequent FabricPath encapsulation by F1 card

This essentially means that the MAC addresses scalability is mainly dictated by the CAM table available on the M1 card, which is capable of containing 128k addresses. The M1 card has all the information needed (MAC address, switch-id of the edge device) to internally forward the traffic to the F1 interface that is properly connected to the destination FabricPath edge switch. However, as previously discussed in the section MAC Scalability Considerations, the endpoint MAC addresses are also learned on the F1 card instead. This is because in the current implementation, the F1 card needs to associate to each remote MAC address a specific value (named local ID [LID]) and insert it into the FabricPath header before sending the frame toward the edge switch (Figure 32).
Figure 32. F1 learning MAC addresses for routed traffic purposes

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 35 of 54

The end result is that each SoC is currently learning all the remote MAC addresses for endpoints locally attached to each FabricPath edge switch. This happens in an independent fashion, whether or not traffic belongs to the same VLAN, which essentially brings the maximum MAC scalability values to 256k per F1 card (since it supports 16 SoCs, each capable of holding up to 16k MAC addresses). By properly assigning edge switches to spine ports, it is theoretically possible to use the aggregated capacity of the spine F1 cards switch-on-chips (SoC) component for up to 128k MAC addresses for routed traffic. However, at the time of this writing, the maximum number of MAC addresses that can be stored for routing purposes without incurring into flooding is 16k. The reason is described by the following examples. Lets assume that one of the uplinks connecting a FabricPath edge switches to the spine fails, as shown in Figure 33:
Figure 33. Failure scenarios for routed traffic

Traffic that originated from the upstream Layer 3 core devices can still be sent to the left aggregation switch, but it needs to use the transit connection (vPC+ peer-link) between the spines to be directed toward the FabricPath edge switch. As a consequence, the MAC addresses of the endpoints connected to that specific FabricPath edge switch will need to be installed on the MAC tables of the SoCs associated to these peer-link interfaces. This would basically limit to 16k the total amount of MAC addresses that can be learned on the FabricPath edge switch, in order to avoid flooding the routed traffic whose MACs have not been learned. Obviously, the situation would become worse if multiple uplinks belonging to separate FabricPath switches fail at the same time, leading to an extremely unlikely case where the maximum MAC scalability supported across the entire FabricPath domain is reduced to 16k MAC addresses (Figure 34).

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 36 of 54

Figure 34.

Impact of multiple links failure for routed traffic

If this unlikely scenario is a concern, it is possible to deploy port-channels (instead than single links) to interconnect each FabricPath edge devices to the spine switches. In that way, a single link failure event would not force the traffic on the peer-link. It is also worth noting that a specific functionality, named mac-proxy, is on the roadmap for a future NX-OS SW release that will help eliminating the issue described above, by removing the need for the F1 SoCs to learn endpoint MAC address information. Once this functionality is available (currently planned for 1HCY12), it would truly be possible to consider the MAC address scalability off a FabricPath domain as only dictated by the M1 MAC table space (minimum 128k). For Layer 2 traffic (that is, traffic that is not directed to the default gateway), the MAC address scalability is dependent only on the MAC scalability of the edge devices. This is even more relevant for VLANs that do not require routing, such as the isolated networks constituted by the virtual server building blocks of typical virtual data centers (vDCs), like the ones depicted in Figure 35:

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 37 of 54

Figure 35.

vDCs use many isolated VLANs

In such cases, the MAC address scalability for isolated VLANs is entirely decoupled from the F1 learning component. In all cases, east-to-west Layer 2 traffic doesnt use the MAC address table of the spine devices. Finally, the ability to do unicast multipathing on up to 16 different equal-cost paths makes it possible to place up to 16 spines in the core, thus lowering the oversubscription and providing an aggregated bandwidth that equals the sum of the bandwidth provided at the spine layer by each spine.

FabricPath Convergence Times


Cisco FabricPath improves significantly on the convergence times for any given failure. For most unicast failures, the traffic loss is within a few hundred milliseconds. For routed traffic, the worst convergence time is the time that it takes for routers to do an SPF calculation. Considering the various throttling mechanisms, this amounts to ~5 seconds, which can be further lowered by changing the default OSPF timers on the adjacent routing devices. The configuration required to achieve this is very simple: router ospf 10 timers throttle spf 10 100 5000 timers throttle lsa all 10 100 5000 timers lsa arrival 80 For Layer 2, multicast convergence times are also within a few hundred milliseconds, while for routed multicast, PIM must be tuned via BFD and/or more aggressive timers than the defaults. Routed multicast with proper tuning can converge within a couple of seconds. Notice that these slower convergence times are not due to FabricPath but to default timers in routing protocols. FabricPath has failure convergence times of a few hundred milliseconds. For better failover times, consider the following:

Use vPC+ at the spine in order for HSRP to be advertised with the emulated switch-id. Consider tuning the default OSPF SPF timers to avoid throttling SPF recalculations too much. Consider tuning the default PIM timers and or using BFD.
Page 38 of 54

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Sample Configurations
This section illustrates how to configure the Cisco Nexus 7000 Series Switches and Cisco Nexus 5500 Switches for a spine-edge design with the spine also providing the Layer 3 gateway function. The configuration samples refer to the network topology shown Figure 36.
Figure 36. Reference Topology for the configurations provided in this document

Note that in regard to topology shown in Figure 36 and following configuration samples, the F1 line card is inserted in slot 8, and the M1 line card is inserted in slot 10.

Cisco Nexus 7000 Series Spine 1 (DC1-Agg1)


install feature-set fabricpath hostname DC1-Agg1 vdc DC1-Agg1 id 1 allow feature-set fabricpath [] feature-set fabricpath [] feature ospf feature pim feature interface-vlan feature hsrp feature lacp feature glbp feature vpc feature pong feature bfd
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 39 of 54

bfd interval 50 min_rx 100 multiplier 3 vlan 101 mode fabricpath name FP_VLAN_101 vlan 102 mode fabricpath name FP_VLAN_102 vlan 103 mode fabricpath name FP_VLAN_103 vlan 333 mode fabricpath name Aggr_IGP_Peering vpc domain 1 role priority 110 peer-keepalive destination 10.60.17.141 delay restore 3600 dual-active exclude interface-vlan 101-104 fabricpath switch-id 1000 interface Vlan101 no shutdown ip address 10.20.0.2/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode ip pim bfd-instance ip pim dr-priority 10 ip pim hello-interval 1000 hsrp 1 preempt delay reload 300 priority 110 ip 10.20.0.1 interface Vlan102 no shutdown ip address 10.30.0.2/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode ip pim bfd-instance ip pim hello-interval 1000 hsrp 1 preempt delay reload 300
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 40 of 54

priority 110 ip 10.30.0.1 interface Vlan103 no shutdown ip address 10.40.0.2/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode hsrp 1 preempt delay reload 300 priority 110 ip 10.40.0.1 interface Vlan333 no shutdown description Aggr_IGP_Peering no ip redirects ip address 11.1.10.1/24 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode interface port-channel1 description FP Link to DC1-5500-1 switchport switchport mode fabricpath interface port-channel10 description vPC+ Peer-Link switchport switchport mode fabricpath vpc peer-link interface Ethernet8/1 description vPC+ Peer-Link Member 1 switchport mode fabricpath channel-group 10 mode active no shutdown interface Ethernet8/2 description vPC+ Peer-Link Member 2 switchport mode fabricpath channel-group 10 mode active no shutdown interface Ethernet8/3 description Port-channel to DC1-5500-1 Member 1
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 41 of 54

switchport mode fabricpath channel-group 1 mode active no shutdown interface Ethernet8/4 description Port-channel to DC1-5500-1 Member 2 switchport mode fabricpath channel-group 1 mode active no shutdown interface Ethernet8/5 description FP Link to DC1-5500-2 switchport mode fabricpath no shutdown interface Ethernet8/7 description FP Link to DC1-5500-3 switchport mode fabricpath no shutdown interface Ethernet10/5 description L3 Link to core ip address 12.1.5.1/24 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode no shutdown interface mgmt0 ip address 10.60.17.140/24 interface loopback0 description local-loopback ip address 1.1.1.1/32 ip router ospf 10 area 0.0.0.0 interface loopback100 description anycast-RP ip address 1.1.1.100/32 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode router ospf 10 router-id 1.1.1.1 timers throttle spf 10 100 5000 timers throttle lsa all 10 100 5000 timers lsa arrival 80

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 42 of 54

fabricpath domain default root-priority 66 topology 1 ip pim rp-address 3.3.3.3 group-list 224.0.0.0/4 ip pim ssm range 232.0.0.0/8 fabricpath switch-id 121

Cisco Nexus 7000 Series Spine 2 (DC1-Agg2)


install feature-set fabricpath hostname DC1-Agg2 vdc DC1-Agg2 id 1 allow feature-set fabricpath [] feature-set fabricpath feature vrrp cfs eth distribute feature ospf feature pim feature msdp feature interface-vlan feature hsrp feature lacp feature glbp feature vpc feature pong feature bfd bfd interval 50 min_rx 100 multiplier 3 vrf context management ip route 0.0.0.0/0 10.60.17.254 vlan 101 mode fabricpath name FP_VLAN_101 vlan 102 mode fabricpath name FP_VLAN_102 vlan 103 mode fabricpath name FP_VLAN_103 vlan 333 mode fabricpath name Aggr_IGP_Peering

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 43 of 54

vpc domain 1 peer-keepalive destination 10.60.17.140 delay restore 3600 dual-active exclude interface-vlan 101-104 fabricpath switch-id 1000 interface Vlan101 no shutdown no ip redirects ip address 10.20.0.3/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode ip pim bfd-instance ip pim hello-interval 1000 hsrp 1 ip 10.20.0.1 interface Vlan102 no shutdown no ip redirects ip address 10.30.0.3/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode ip pim bfd-instance ip pim hello-interval 1000 hsrp 1 ip 10.30.0.1 interface Vlan103 no shutdown no ip redirects ip address 10.40.0.3/16 ip ospf passive-interface ip router ospf 10 area 0.0.0.0 ip pim sparse-mode hsrp 1 ip 10.40.0.1 interface Vlan333 no shutdown description Aggr_IGP_Peering no ip redirects ip address 11.1.10.2/24 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 44 of 54

interface port-channel2 description FP Link to DC1-5500-1 switchport switchport mode fabricpath interface port-channel10 description vPC+ Peer-Link switchport switchport mode fabricpath vpc peer-link interface Ethernet8/1 description vPC+ Peer-Link Member 1 switchport mode fabricpath channel-group 10 mode active no shutdown interface Ethernet8/2 description vPC+ Peer-Link Member 2 switchport mode fabricpath channel-group 10 mode active no shutdown interface Ethernet8/3 description FP Link to DC1-5500-1 switchport mode fabricpath channel-group 2 mode active no shutdown interface Ethernet8/4 description FP Link to DC1-5500-1 switchport mode fabricpath channel-group 2 mode active no shutdown interface Ethernet8/5 description FP Link to DC1-5500-2 switchport mode fabricpath no shutdown interface Ethernet8/7 description FP Link to DC1-5500-3 switchport mode fabricpath no shutdown interface Ethernet10/5
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 45 of 54

description L3 Link to core ip address 12.1.6.1/24 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode no shutdown interface mgmt0 ip address 10.60.17.141/24 interface loopback0 description local-loopback ip address 1.1.1.2/32 ip router ospf 10 area 0.0.0.0 interface loopback100 description anycast-RP ip address 1.1.1.100/32 ip router ospf 10 area 0.0.0.0 ip pim sparse-mode router ospf 10 router-id 1.1.1.2 timers throttle spf 10 100 5000 timers throttle lsa all 10 100 5000 timers lsa arrival 80 fabricpath domain default root-priority 65 topology 1 ip pim rp-address 3.3.3.3 group-list 224.0.0.0/4 fabricpath switch-id 122

Edge Cisco Nexus 5500 Switch (DC1-5500-1) without vPC+


install feature-set fabricpath feature-set fabricpath hostname DC1-5500-1 feature telnet no feature ssh cfs eth distribute feature interface-vlan feature lacp feature lldp

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 46 of 54

vrf context management ip route 0.0.0.0/0 10.60.17.254 vlan 1 vlan 101 mode fabricpath name FP_VLAN_101 vlan 102 mode fabricpath name FP_VLAN_102 vlan 103 mode fabricpath name FP_VLAN_103 interface port-channel1 description FP Link to DC1-Agg1 switchport mode fabricpath interface port-channel2 description FP Link to DC1-Agg2 switchport mode fabricpath interface Ethernet1/1 description FP EC Member to DC1-Agg1 switchport mode fabricpath channel-group 1 mode active interface Ethernet1/2 description FP EC Member to DC1-Agg1 switchport mode fabricpath channel-group 1 mode active interface Ethernet1/3 description FP EC Member to DC1-Agg2 switchport mode fabricpath channel-group 2 mode active interface Ethernet1/4 description FP EC Member to DC1-Agg2 switchport mode fabricpath channel-group 2 mode active interface Ethernet1/23 description to_Ixia2_port_1 switchport access vlan 101 spanning-tree port type edge interface mgmt0
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 47 of 54

ip address 10.60.17.106/24 fabricpath domain default fabricpath switch-id 111

Edge Cisco Nexus 5500 Switch (DC1-5500-1) with vPC+ and FEX A/A
install feature-set fabricpath feature-set fabricpath hostname DC1-5500-1 feature telnet no feature ssh cfs eth distribute feature interface-vlan feature lacp feature vpc feature lldp feature fex fex 101 pinning max-links 1 description FEX0101 fex 102 pinning max-links 1 description FEX0102 slot 101 provision model N2K-C2232P slot 102 provision model N2K-C2232P vrf context management ip route 0.0.0.0/0 10.60.17.254 vlan 101 mode fabricpath name FP_VLAN_101 vlan 102 mode fabricpath name FP_VLAN_102 vlan 103 mode fabricpath name FP_VLAN_103 vpc domain 2 role priority 110 peer-keepalive destination 10.60.17.107 fabricpath switch-id 1122
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 48 of 54

interface port-channel1 description FP Link to DC1-Agg1 switchport mode fabricpath speed 10000 interface port-channel2 description FP Link to DC1-Agg2 switchport mode fabricpath speed 10000 interface port-channel3 description vPC+ Peer-Link switchport mode fabricpath vpc peer-link interface port-channel101 description vPC to FEX101 switchport mode fex-fabric fex associate 101 vpc 101 interface port-channel102 description vPC to FEX102 switchport mode fex-fabric fex associate 102 vpc 102 interface Ethernet1/1 description FP EC Member to DC1-Agg1 switchport mode fabricpath interface Ethernet1/2 description FP EC Member to DC1-Agg1 switchport mode fabricpath interface Ethernet1/3 description FP Link to DC1-Agg2 switchport mode fabricpath interface Ethernet1/4 description FP Link to DC1-Agg2 shutdown switchport mode fabricpath interface Ethernet1/5 switchport access vlan 101
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 49 of 54

speed 1000 interface Ethernet1/15 switchport mode fabricpath switchport trunk allowed vlan 101-103 channel-group 3 interface Ethernet1/16 switchport mode fabricpath switchport trunk allowed vlan 101-103 channel-group 3

interface Ethernet1/23 switchport access vlan 101 spanning-tree port type edge interface Ethernet1/31 switchport mode fex-fabric fex associate 101 channel-group 101 interface Ethernet1/32 switchport mode fex-fabric fex associate 102 channel-group 102 interface mgmt0 ip address 10.60.17.106/24 interface Ethernet101/1/1 switchport access vlan 101 interface Ethernet102/1/1 switchport access vlan 102 fabricpath domain default fabricpath switch-id 111

Edge Cisco Nexus 5500 Switch (DC1-5500-2) with vPC+ and FEX A/A
install feature-set fabricpath feature-set fabricpath hostname DC1-5500-2 feature telnet
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 50 of 54

no feature ssh cfs eth distribute feature interface-vlan feature vpc feature lldp feature fex fex 101 pinning max-links 1 description FEX0101 fex 102 pinning max-links 1 description FEX0102 vrf context management ip route 0.0.0.0/0 10.60.17.254 vlan 101 mode fabricpath name FP_VLAN_101 vlan 102 mode fabricpath name FP_VLAN_102 vlan 103 mode fabricpath name FP_VLAN_103 vpc domain 2 peer-keepalive destination 10.60.17.106 fabricpath switch-id 1122 interface port-channel3 switchport mode fabricpath vpc peer-link interface port-channel101 switchport mode fex-fabric fex associate 101 vpc 101 interface port-channel102 switchport mode fex-fabric fex associate 102 vpc 102 interface Ethernet1/1 description FP Link to DC1-Agg1 switchport mode fabricpath channel-group 1 mode active
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 51 of 54

interface Ethernet1/2 description FP Link to DC1-Agg1 switchport mode fabricpath channel-group 1 mode active interface Ethernet1/3 description FP Link to DC1-Agg2 switchport mode fabricpath channel-group 2 mode active interface Ethernet1/4 description FP Link to DC1-Agg2 switchport mode fabricpath channel-group 2 mode active interface Ethernet1/8 switchport access vlan 101 interface Ethernet1/11 switchport mode fabricpath speed 1000 interface Ethernet1/15 description vPC+ Peer-Link Member switchport mode fabricpath channel-group 3 interface Ethernet1/16 description vPC+ Peer-Link Member switchport mode fabricpath channel-group 3 interface Ethernet1/23 switchport access vlan 101 spanning-tree port type edge interface Ethernet1/24 switchport access vlan 101 spanning-tree port type edge interface Ethernet1/31 switchport mode fex-fabric fex associate 102 channel-group 102 interface Ethernet1/32
2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information. Page 52 of 54

switchport mode fex-fabric fex associate 101 channel-group 101 interface mgmt0 ip address 10.60.17.107/24 interface Ethernet101/1/1 switchport access vlan 101 interface Ethernet102/1/1 switchport access vlan 102 fabricpath domain default ip route 0.0.0.0/0 10.1.101.254 fabricpath switch-id 112

Printed in USA

C07-690079-00

12/11

2011 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public Information.

Page 53 of 54

Potrebbero piacerti anche