Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Platform Architecture
ASR 9000 “At-a-Glance”
Next-Generation SP Edge & Aggregation
Optimized for
high dense10G &
100G aggregation
IOS-XR
Video DNA Non-Stopping
service
400G/Slot
Longevity
Green
ANA service
management and
Fully converged provisioning
L2 & L3 service
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 2
ASR 9000 System Architecture “At-a-Glance”
Line Card
Fully Distributed
Architecture for High RSP
Performance and High Multi-
dimensional Control Plane
Scale CPU
CPU BITS/DTI
Data forwarding is fully FIA
distributed on the line cards
FIC
RSP (0-1)
System fan trays
Air draw
DC Supplies
A
B 2.1/1.5 kW 6 & 10 slot use same power supplies
A Single power zone
B 2.1 kW
All power supplies run in active mode
Power draw shared evenly
AC Supplies
Power Supply 50 Amp DC Input or 16 Amp AC
for Easy CO Install
A 3 kW
B 3 kW
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 5
Power Check and Rules
Available power is checked when:
–An LC card is inserted
–An LC card is powered up via the CLI
–An LC card is reset via “hw-mod reload”
If the system does not have enough available power to
accommodate the LC, then the LC becomes “UNPOWERED”
Installing new power supplies will not automatically power up any
UNPOWERED line cards. The user can force a recheck using:
“hw-mod reload loc <>”
RSP and Fan Tray cards are given priority allocation of power
budget
LC power budget is checked in numeric order until it is exhausted.
The actual power up of the LC's is in parallel, and asynchronous
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 6
RSP Engine
Performs control plane and management functions
Dual Core CPU processor with 4GB or 8GB DRAM
2MB NVRAM, 4GB internal bootdisk, 2 external compact flash
slots
Dual Out-of-band 10/100/1000 management interface
Console & auxiliary serial ports
Hard Drive: 70G HDD
Status light
Console Port BITS ALARM
Compact Flash Status LED
Management AUX Port Slots
Ethernet
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 7
RSP Operations Impact Fabric ?
Guarantee “0” packet loss for RSP failover or OIR
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 8
Ethernet Line Card Family Common HW ASIC and HW architecture
Identical SW features
Line card Memory options for QoS scale
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 9
Line Card Architecture – Hardware Components
CPU (same as RSP)
Program HW forwarding tables
NP: Network Processor Distributed Control planes
Main forwarding ASIC SW switched packets
L2 & L3 forwarding, Inline Netflow
features (QoS, ACL, etc), CPU
control plane policing, PHY NP0
mcast replication, etc
PHY NP1
10Gbps bi-directional with B0
features applied FIA0 FIA: Fabric Interface ASIC
PHY NP2 Provide non-blocking data
B1 connection to switch fabric
Internal system queues/VoQ
PHY NP3 Intelligent mcast replication
PHY NP4
A9K-8T/4-E/B/L A9K-40G-E/B/L
PHY 3
PHY 7
NP0 NP0
PHY 2
PHY 6
NP1 NP1
FIA FIA
PHY 1
PHY 5
NP2 NP2
PHY 0
PHY 4 NP3 NP3
A9K-8T-E/B/L A9K-16T/8-B
PHY
PHY NP0 PHY NP0
PHY
PHY NP1 PHY NP1
FIA0 PHY
FIA0
PHY NP2 PHY NP2
PHY
PHY NP3 PHY NP3
PHY
PHY NP4 PHY NP4
PHY
PHY NP5 PHY NP5
FIA1 PHY
FIA1
PHY NP6 PHY NP6
PHY
PHY NP7 PHY NP7
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 12
Line Card Architecture
A9K-8T/4
60Gbps raw bandwidth
A9K-16T/8-B
PHY CPU 120Gbps raw bandwidth
Crossbar
NP0 Fabric
ASIC
PHY
NP1 Crossbar
B0 Fabric
PHY FIA0 ASIC
NP2
B1
PHY Arbiter
NP3
RSP0
PHY
NP4 Crossbar
Fabric
PHY ASIC
NP5 B0
PHY
FIA1 Crossbar
NP6 B1 Fabric
ASIC
PHY NP7 Arbiter
RSP1
B0 NP1 PHY
FIA
FGID – Fabric Group ID B1 NP2 PHY
MGID – Multicast Group ID
MFIB – Multicast Forwarding Information Base NP3 PHY
LC3
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 14
Multicast Packet Replication (2)
Switch Fabric and Egress LC Replication
2 FIA Replication replicate
single copy to NPs which
1 Fabric Replication receive IGMP join, based on NP Replication replicate
4
replicate single copy to MGID table in FIA copy per receiver based on
LCs which receive IGMP multicast FIB table
3 Bridge Replication similar
join, based on FGID table as FIA replication, single copy
in switch fabric to NP MFIB
1 0000000000 CPU
IGMP joins
4
2 0000000001 NP0 PHY
FGID=3 MGID
MGID
3 0000000011
hit 3 NP1
2 B0 PHY
IGMP joins CPU …
FIA
PHY NP0 N 1111001111 B1 NP2 PHY
1 00 Replicate to the first
3 FGID
PHY NP1 FPOE
B0
Table2 MGID 2 hit 01 NP3
Bridge only PHY
1 LC2
IGMP joins FIA 3 10
PHY NP2
B1 Switch …
Multicast Fabric CPU
Source PHY NP3 N 10 NP0 PHY
MGID Table in FIA NP1
B0 PHY
FIA
FGID – Fabric Group ID B1 NP2 PHY
MGID – Multicast Group ID
MFIB – Multicast Forwarding Information Base NP3 PHY
LC3
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 15
Multicast Packet Replication (3)
Inside NP
To egress
port Second lookup 14Gbps
replication
TM capacity per NP
NP
Initial Lookup
– Original packet is lookup in the MFIB/L2FIB returning the number of copies
to be replicated on the given NPU, each copy per output logical interface
– Packet is sent to a TM (traffic manager). TM replicates the packets and send
the copy to the processing engine
Second Lookup
–Each copy is processed and sent out of interface. Egress feature like QoS,
ACL is applied at this pass
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 16
L3 Multicast Control Plane Overview
4 MFIB
PHY NP0 PIM IGMP
CPU 3 2
5
PHY NP1
B0 MRIB CPU
FIA 1
PHY NP2
B1 Crossbar Crossbar
Fabric Fabric
PHY NP3 ASIC
Switch Fabric
ASIC
Incoming IGMP and PIM packets are punted to RP directly bypassing LC CPU
Protocols (PIM/IGMP) send their Route/OLIST Information to MRIB process to build multicast
route/olist table
MRIB sends the multicast state information to MFIB process on all LCs
MFIB program HW forwarding tables in NP, Bridge FPGA and Fabric interface ASIC
Software switched multicast packet or data packet for protocol signaling is sent to local line
card CPU
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 17
L2 Multicast Control Plane
4 L2FIB IGMP snooping
PHY NP0 CPU 3 2
5
PHY NP1
B0 L2FIB CPU
FIA 1
PHY NP2
B1 Crossbar Crossbar
Fabric Fabric
PHY NP3 ASIC
Switch Fabric
ASIC
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 18
Line Card Memory Options – Queue Scale
High
3 memory options for each line card: Extended (or high queue),
Base (medium queue), Low (low queue)* Medium
Different memory option has different QoS queue scale and L2 Low
sub-interface scale. All other system wide scale is the same
across different type of the line cards, including FIB, MAC
address, Bridge-domain, L3 sub-interface, VRF, etc
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 19
L/B/E Line Cards – What’s the Difference?
STATS MEMORY
FIB MAC
LOOKUP FRAME MEMORY
Forwarding ASIC
MEMORY
-
TCAM
NP complex
Each NPU has Four Main Associated memories TCAM , Search/Lookup memory , Frame/buffer
memory and statistics memory
–TCAM is used for VLAN tag, QoS and ACL classification
–Lookup Memory is used for storing FIB tables, Mac address table and Adjacencies
–Stats memory is used for all interface statistics, forwarding statistics etc
–Frame memory is buffer memory for Queues
E/B/L line card have different TCAM , Stats and Frame Memory size, which give different scale number
of the QoS queues and L2 sub-interfaces per line card
Lookup Memory is the same across line card s why?
–To support mix of the line cards without impacting the system wide scale including routing,
multicast, MAC address, L3 interface, MPLS label space scale
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 20
Optics
Standard, CWDM, & DWDM XFPs/SFPs/SFP+ available
IPoDWDM G.709 FEC/EFEC
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public *Not Commtted 22
Switch Fabric Overview
Active-active load balancing: Unicast: per-packet load balancing, Multicast: per (S,G) load balancing
Arbiter for fabric access control. Arbiter is in active/standby mode, which is controlled by low level
hardware signalling
Frame format over fabric: super-frame, it can aggregate multiple small packet into a big sup-frame to
improve the fabric throughput
Crossbar FIA0
Fabric
ASIC
Arbiter FIA1
FIA RSP0
Dual-FIA
8xNPs
Crossbar Linecard
Single-FIA Fabric
4xNPs ASIC
Linecard
Crossbar
Fabric
4x23Gbps =92Gbps with dual RSP ASIC 8x23Gbps =184Gbps with dual RSP
2x23Gbps=46Gbps with single RSP 4x23Gbps=92Gbps with single RSP
Arbiter
RSP1
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 23
Fabric Redundancy
Guarantee “0” packet loss during RSP failover and OIR
Separate data and arbitration(control) paths
All fabric data channels run in active mode for extra fabric bandwidth and instant fabric
switch over
Both Arbiters work in parallel – both answer all requests, FIAs follow active Arbiter, FIAs
switch to backup if needed instant control path switchover
Arbiter switchover is controlled by low level hardware signalling
Crossbar
Fabric
ASIC
Crossbar FIA0
Fabric
ASIC
Arbiter FIA1
FIA RSP0
Dual-FIA
8xNPs
Crossbar Linecard
Single-FIA Fabric
4xNPs ASIC
Linecard
Crossbar
Fabric
ASIC
Arbiter
Presentation_ID RSP1
© 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 24
L3 Control Plane Overview
BGP OSPF
LDP RSVP-TE
Static
ISIS EIGRP
LSD RIB RP
ARP
FIB Adjacency
SW FIB
AIB LC NPU
0 1 2 3 . . . … n
BGP PIC
VRF table
TE-FRR PIC
Link bundle PIC
OutI/F
adj
LAG OutI/F
Recursive
NR Prefix adj
Leaf Table Prefix Leaf
Leaf Table
Table
Protected
LAG OutI/F
TE adj
NR LDI 0
Recursive
LDI/ADJ LDI (4-way) NR LDI 1
… Backup
OutI/F
NR LDI N TE adj
Presentation_ID © 2010 Cisco and/or its affiliates. All rights reserved. Cisco Public 26
Two-Stage Packet Forwarding
Fully Distributed Forwarding on Line Cards
Each line card has independent AIB only
for local interfaces
Packet is forwarded to the egress Each line card has independent Interface
NP based on the information in DB for local interfaces
the NP/fabric header
Both Ingress and Egress FIB – allows
forwarding features to be independently
applied on LCs
CPU CPU
PHY NP0 NP0 PHY
2
PHY NP1 NP1 PHY
FIA FIA
PHY NP2 NP2 PHY
Switch
Fabric
PHY NP3 Ingress LC Egress LC NP3 PHY
1