0% found this document useful (0 votes)
1K views

Data Center Design Power Session

Cisco Public Infrastructure Design LAN Switching Analysis Recap on Current Trends New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling. Cisco Public Data Center Layout Better to move the Horizontal Distribution closer to the servers to reduce the cable length.

Uploaded by

Jarod Ahlgren
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
1K views

Data Center Design Power Session

Cisco Public Infrastructure Design LAN Switching Analysis Recap on Current Trends New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling. Cisco Public Data Center Layout Better to move the Horizontal Distribution closer to the servers to reduce the cable length.

Uploaded by

Jarod Ahlgren
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 192

Data Center Power Session

TECDCT-3873

Presentation_ID

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Agenda
Infrastructure Design LAN Switching Analysis
Recap on Current Trends New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling

Break Demos: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server Break Demo: Nexus1kv Blade Servers
Blade Switching LAN Blade Switching SAN Unified Compute System

Break Demo: UCS SAN Switching Analysis

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Infrastructure Design

Presentation_ID

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Data Center Layout


Better to move the Horizontal distribution closer to the servers to reduce the cable length

Main Distribution Area

Horizontal Distribution Area

Equipment Distribution Area

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Horizontal Distribution at Each Row (aka End of the Row Design)

From Direct connect to End of the Row

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Datacenter Building Block: the POD


HDA

Physical Pod

Defines a discrete amount of physical infrastructure


Racks + Power Distribution + CRAC

Pay-as-you-grow modularity - Predictable, Scalable & Flexible Pod server density affected by power & cooling, cabling & server connectivity
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Overall DC Layout

HDA MDA

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Mapping Between Physical Layout and Network Topology: HDA

Equipment Distribution Area (EDA) Single POD HDA


Acc1 Acc2

336 Servers

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Mapping Between Physical Layout and Network Topology: MDA


Core 1 Core 2

Additional Equipment: Core Routing\Firewalls LAN Appliances SAN Directors

Agg1

Agg2

Agg3

Agg4

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

10 Gigabit Ethernet
for Server Connectivity
Mid 1980s 10Mb
UTP Cat 3

Mid 1990s 100Mb


UTP Cat 5

Early 2000s 1Gb


UTP Cat 5 MMF, SMF

Late 2000s 10Gb


UTP Cat6a MMF, SMF TwinAx, CX4

10G Options
Connector (Media)
SFP+ CU*
copper

Cable
Twinax Twinax MM OM2 MM OM3 MM OM2 MM OM3
Cat6 Cat6a/7 Cat6a/7

Distance
<10m 15m 10m 100m 82m 300m
55m 100m 30m
** Draft 3.0, not final

In-rack X-rack Power


(each side)
~ 0.1W 4W 1W

Transceiver Latency (link)


~ 0.1s ~ 0.1s ~0 ~0
2.5s 2.5s 1.5s

Standard
SFF 8431** IEEE 802.3ak

X2 CX4
copper

SFP+ USR
MMF, ultra short reach

SFP+ SR
MMF,short reach

Across racks 1W
~ 6W*** ~ 6W*** ~ 4W***

~50% power none savings with EEE IEEE 802.3ae

RJ45 10GBASE-T
copper
* Terminated cable

IEEE 802.3an

*** As of 2008; expected to decrease over time

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

10

Twisted Pair Cabling For 10GBASE-T (IEEE 802.3an)

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

11

10G Copper Infiniband 10GBase-CX4


IEEE 802.3ak Supports 10G up to 15 meters Quad 100 ohm twinax, Infiniband cable and connector Primarily for rack-to-rack links Low Latency

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

12

10G SPF+ Cu
SFF 8431 Supports 10GE passive direct attached up to 10 meters Active cable options to be available Twinax with direct attached SFP+ Primarily for in rack and rack-to-rack links Low Latency, low cost, low power

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

13

10GBase-*X (IEEE 802.3ae)


The 802.3ae 10GbE standard defines 3 MM and 1 SM fiber category based on the maximum transmission reach as shown below (ISO 11801 Standard defines the following MM and SM ( g fiber types):
REACH SPEED
300m 500m 200m

100Mb/s 1,000Mb/s 10Gb/s

OM1 OM1 OM3

OM1 OM2 OS1

OM1 OS1 OS1

OM1 is equivalent to standard 62.5/125m MM fiber OM2 is equivalent to standard 50/125m fiber. OM3 is laser enhanced 50/125m fiber 10gig OS1 is equivalent to SM 8/125m fiber.

Not all laser optimized 10Gig fiber cable is the same.


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

10Gig

150M OM2 Plus

300M OM3

550M OM3 Plus


14

Cisco Public

Optics Positioning for Data Centers


1G Optics Type
1000 BASE-LX 1000 BASE-SX 1000 BASE T BASE-T

Max PMD Distance (m) 10G Optics Type


10GBASE-LR 10GBASE-LRM

10

100

500

~10000

Require OM3 MMF

10GBASE-SR 10GBASE-T 10GBASE-USR 10GBASE-CX4 10GBASE-CX1


30M/100M OM3 MMF Only

Max PMD Distance (m)

10

26-82

100

220

300

~10000

In Rack X-rack
<10M
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

Mid to End of Rack


<100 M
Cisco Public

Across Aisles
<300 M

Across Sites
<10 KM
15

Cost Effective 10G Server Connectivity Today

SFP+ USR Ultra Short Reach


100M on O OM3 fiber, 30M on OM2 fiber f O f Support on all Cisco Catalyst and Nexus switches Low Cost Target FCS: Q1 CY09

SFP+ Direct Attach


1, 3, 1 3 5 and 7M on Twinax 0.1W Power Support across all Nexus Switches Low Cost

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

16

Cabling Infrastructure Patch Panels for End of the Row or Middle of the Row
Category 6A (Blue) with OM3 MM (Orange) per Rack, terminating in patch rack at EoR Cable count varies based on design requirement

Fiber for SAN or for TOR switches Copper for EoR server connectivity
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

17

HDA Photos for End or Middle of the Row


cables on the back go to the TOR patch panels

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

18

End of the Row Connectivity Details


End of Row End of Row
Traditionally used Copper from server to access switches
Patch P t h panel l X-connect Patch P t h panel l X-connect Patch panel server Patch panel server

Common Characteristics
Typically used for modular access Cabling is done at DC build-out Model evolving from EoR to MoR Lower cabling distances (lower cost) Allows denser access (better flexibility) 6-12 multi-RU servers per Rack 4-6 kW per server rack, 10Kw-20Kw per network rack Subnets and VLANs: one or many per switch. Subnets tend to be medium and large
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Network Access Point A-B

Network Access Point C-D server server

Fiber Copper

Middle of Row

Patch panel server Patch panel X-connect Patch panel X-connect

Patch panel server

Network Access Point A-B server

Network Access Point C-D server

19

Top of the Rack Connectivity Details


ToR
Used in conjunction with dense access racks(1U servers) Typically one access switch per rack Some customers are considering two + cluster Typically: ~10-15 server per rack (enterprises) ~15-30 server per rack (SP) Use of either side of rack is gaining traction Cabling: Within rack: Copper for server to access switch Outside rack (uplink): Copper (GE): needs a MoR model for fiber aggregation Fiber (GE or 10GE):is more flexible and also requires aggregation model (MoR)
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Patch panel Top of Rack p server Patch panel X-connect Patch panel X-connect

Patch panel Top of Rack server

server

Network Aggregation Point A-B

Network Aggregation Point A-B server

To network core

Patch panel Top of Rack Top of Rack server Patch panel X-connect Patch panel X-connect

Patch panel Top of Rack Top of Rack server

Network Access Point A-B server

Network Access Point C-D server

20

Blade Chassis Connectivity Details


End of Row (Switch to Switch)
Scales well for blade server racks ( (~3 blade chassis per rack) p ) Most current uplinks are copper but the NG switches will offer fiber
Patch panel Patch panel X-connect Patch panel X-connect Patch panel

sw1

sw2

sw1

sw2

Blade Chassis sw1 sw2 Network Aggregation Point ABC-D

Blade Chassis sw1 Network Aggregation Point AB-C-D sw2

Blade Chassis

Blade Chassis sw1 sw2

End of Row (Pass-through)


Scales well for pass-through blade racks Copper from servers to access switches

sw1

sw2

Blade Chassis

Blade Chassis

Patch panel

Patch panel Patch panel X-connect Top of Rack Patch panel X-connect Pass-through Blade Chassis Pass-through Network Aggregation Point ABC-D Network Aggregation Point AB-C-D Blade Chassis Pass-through Blade Chassis

ToR
Pass-through

Viiable option on pass-through environments is the access port count is right

Blade Chassis Pass-through Blade Chassis Pass-through Blade Chassis

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

21

Final Result
12 Server PODs Consists of the following: 4 Switch Cabinets for LAN & SAN 32 S Server C bi t Cabinets 12 Servers per Server Cabinet
Core 1 Core 2

Servers: 4032 6509 Switches: 30 Server/Switch Cabinets: 399 Midrange/SAN Cabinets Allotted For: 124 Mid /SAN C bi t All tt d F

Agg1

Agg2

Agg3

Agg4

Acc1

Acc2
6 Pair Switches

Acc11

Acc12

Acc13

Acc14
6 Pair Switches

Acc23

Acc24

336 Servers
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

336 Servers
Cisco Public

336 Servers

336 Servers 22

LAN Switching in the Datacenter

Presentation_ID

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

23

Icons and Associated Product


=
Nexus 7000 Nexus 5000 Catalyst 4948-10GE Catalyst 4900M

with Service Modules

=
Nexus 1000v Catalyst 6500 with VSS = Nexus 2148T CBS 3100 Blade Switches

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

24

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server Break Demo: Nexus1kv
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

25

Data Center Architecture


Existing Layer 2 STP Design Best Practices
Rapid PVST+ UDLD Global Spanning Tree Pathcost Method=Long

L3+ L4 CEF Hash

LACP+L4 Port Hash Dist EtherChannel for FT and Data VLANs FT Data

Agg1: STP Primary Root HSRP Primary HSRP Preempt and Delay Dual Sup with NSF+SSO
LACP+L4 Hash Dist EtherChannel Min-Links

Agg2: STP Secondary Root HSRP Secondary HSRP Preempt and Delay Single Sup

Rootguard LoopGuard Portfast + BPDUguard

Blade Chassis with Integrated Switch

Rapid PVST+: Maximum Number of STP Active Logical Ports- 8000 and Virtual Ports Per Linecard-1500
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

26

Migration from Inline Services


The Need: Higher performance/ scalability required in aggregation and/or core The Migration: Move Catalyst 6500 chassis with service modules to an on-the-stick configuration and re-use high speed links to connect to the aggregation Layer

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

27

VSS Allows a Migration to a L2 Topology Based on Etherchannels


IETF NSF 10 Gig uplinks

STP Root

6500 with VSS IETF NSF-capable 10 Gig uplinks

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

28

VSS Is Currently Being Used Also for Data Center Interconnect


nPE nPE

VSS system

VSS system

Main benefits Loop Avoidance Load balancing Failover


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

29

Nexus-Based Datacenters High Density 10 Gigabit and Unified IO readiness


10G Core Performance
WAN
Nexus Core Layer Nexus Agg Layer Nexus Agg Layer

10G Aggregation Density


Core

Access 1G/10G to the Host


Core

Blade Servers Top of Rack

Nexus

Access Layer

High performance, Hi h f highly available 10GE core connectivity

Full F F ll Featured 10G d Density for aggregating 10G Top of Rack and 10G Blade Servers

As i A virtualization li i drives host I/O utilization, 10G to the host requirements are becoming reality

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

30

New Aggregation Layer Nexus 7k in the Aggregation


4*10GbE

8*10GbE 8 10GbE

4*10GbE

Optional dedicated links in case 6k is deployed as VSS

Nexus-based Aggregation Layer with VDCs, CTS and vPCs d PC Catalyst 6500 services chassis with Firewall Services and ACE Module provides Advanced Service delivery Possibility of converting the Catalyst 6500 in VSS mode

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

31

New Access Layer Options


New Options Include: Nexus 7k, FEX, 10 GigE TOR
1GbE End-of-Row

New Options highlighted in red 10 Gigabit Top of the Rack Connectivity with the Nexus 5k Fabric Extender (Nexus2k) Server Virtual Switching (Nexus1kv)
Nexus 7018 Catalyst 6500 1GbE Top-of-Rack Nexus 2148T Nexus 5000

10GbE End-of-Row

Nexus 7000 Nexus 1000v 10GbE Top-of-Rack Nexus 5000

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

32

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server Break Demo: Nexus1kv
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

33

New Layer 2 Technologies


Applicability Link Layer Encryption Virtual Domain Contexts vPC Nexus 7k and future on other Nexus platforms Nexus 7k Nexus 7k Nexus 5k Catalyst 6500 (as VSS) MAC pinning L2MP VNTAG Datacenter Ethernet OTV
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

Fabric Extender (Nexus2148T) Nexus1kv Future on Nexus products Nexus5k, Nexus2k, SR-IOV 10 Gigabit Adapters Nexus 5k and future linecards, Converged Network Adapters Layer 2 extension
Cisco Public

34

Cisco TrustSec
TrustSec Linksec (802.1ae) Frame Format
The encryption used by TrustSec follows IEEE Standards-based LinkSec (802.1ae) encryption, where the upper layers are unaware of the L2 header/encryption.

CMD E_TYPE

Version

Length

SGT Option Length & Type

SGT Value

Variable

DMAC

SMAC

802.1ae Header (16 Octets)

.1Q (4)

CMD (8 Octets)

ETH_TYPE ETH TYPE

Payload P l d

ICV (16 Octets)

CRC

Encrypted Authenticated

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

35

Nexus 7000 TrustSec


Sample Config Manual 802.1AE Symmetric Configuration
DC1

Encrypted Traffic Public Soil

DC2

Nexus-7000-1(config)# interface ethernet 2/45 Nexus-7000-1(config-if)# cts manual Nexus-7000-1(config-if-cts-manual)# sap pmk 12344219 Nexus-7000-1(config-if-cts-manual)# exit Nexus-7000-1# show cts CTS Global Configuration ============================== CTS support : enabled CTS device identity : test1 CTS caching support : disabled Number of CTS interfaces in DOT1X mode : 0 Manual mode : 1

Nexus-7000-2(config)# interface ethernet 2/3 Nexus-7000-2(config-if)# cts manual Nexus-7000-2(config-if-cts-manual)# sap pmk 12344219 Nexus-7000-2(config-if-cts-manual)# exit Nexus-7000-2# show cts CTS Global Configuration ============================== CTS support : enabled CTS device identity : test2 CTS caching support : disabled Number of CTS interfaces in DOT1X mode : 0 Manual mode : 1

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

36

Nexus 7000 TrustSec


Interface Verification

Nexus-7000-1# show cts interface e 2/3 CTS Information for Interface Ethernet2/3: CTS i enabled, mode: is bl d d CTS_MODE_MANUAL CTS MODE MANUAL IFC state: CTS_IFC_ST_CTS_OPEN_STATE Authentication Status: CTS_AUTHC_SKIPPED_CONFIG Peer Identity: Peer is: Not CTS Capable 802.1X role: CTS_ROLE_UNKNOWN Last Re-Authentication: Authorization Status: CTS_AUTHZ_SKIPPED_CONFIG PEER SGT: 0 Peer SGT assignment: Not Trusted Global policy fallback access list: SAP Status: CTS_SAP_SUCCESS g p p Configured pairwise ciphers: GCM_ENCRYPT Replay protection: Disabled Replay protection mode: Strict Selected cipher: GCM_ENCRYPT Current receive SPI: sci:225577f0860000 an:1 Current transmit SPI: sci:1b54c1a7a20000 an:1

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

37

Virtual Domain Contexts


Horizontal Consolidation
Objective: Consolidate lateral infrastructure that delivers similar roles for separate operational or administrative domains. Benefits: R d B fit Reduced power and space requirements, can maximize d d i t i i density of the platform, easy migration to physical separation for future growth. Considerations: Number of VDCs (4), Four VDCs != Four CPU Complexes, does not significantly reduce cabling or interfaces needed.
core 1 core 2 core core

Core Devices

Core

Aggregation Devices

agg1

agg2

agg3

agg4

agg VDC 1 agg VDC 2

agg VDC 1 agg VDC 2

Aggregation VDCs

acc1

acc2

accN

accY

acc1

acc2

accN

accY

Admin Group 1 TECDCT-3873_c2

Admin Group 2 Cisco Public

agg VDC 1 Admin Group 1

agg VDC 2 Admin Group 2

2009 Cisco Systems, Inc. All rights reserved.

38

Default VDC
The default VDC (VDC_1) is different from other configured VDCs.

Default VDC
vrf
VDC Admin

Can create/delete VDCs Can allocate/de-allocate resources to/ from VDCs Can intercept control plane and potentially some data-plane traffic from all VDCs (using wireshark) Has control over all global resources and p parameters such as managment0 g interface, console, CoPP, etc.

Network Admin

Can have the network-admin role which network admin has super-user priviledges over all VDCs

VDC2
VDC Admin

vrf

VDC3
VDC Admin

vrf

VDC4

vrf

With this in mind for high-security or critical environments the default VDC should be treated differently. It needs to be secured.

mgmt port (mgmt0)

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

39

VDC Best Practices


For high-security environments the Default VDC is really the Master VDC: Reserve the master VDC for VDC and resource administration when deploying a multi VDC environment Avoid multi-VDC environment. running data-plane traffic via the master VDC. Protect the Master VDC: Restrict access to the master VDC to the absolute minimum required to support VDC and overall global system administration. Default HA policy (2-Sups) is switchover: For enhanced VDC independence in dual supervisor configurations, explicitly set the HA polic for VDCs to restart or bringdo n policy bringdown. CoPP is global: Review CoPP policies to ensure that limits are inline with collective requirements of all VDCs. In multi-administrative environments make sure co-ordinate potential service or outage windows with administrative groups across VDCs.
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

40

Resource Scalability Limits


Some resource scalability is limited per system, others are per VDC
16,000 maximum Logical Interfaces (RPVST+) TOTAL for all configured VDCs* 75,000 maximum Logical Interfaces (MST) TOTAL for all configured VDCs* 256 per configured VDC* 4096 VLANs per configured VDC*

FIB TCAM can be scaled by planning interface allocations


FIB is per I/O module and is only populated with entries for VDCs assigned on a module You can optionally maximize this by using the following rule: Assign 1 VDC per module (slot), with 2 modules minimum per VDC on a single system (to preserve redundancy)
*

for 4.0(3)
41

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

VDC Granularity for Current 10 GigE Ports


VDC A Ports are assigned on a per VDC basis and cannot be shared across VDC s VDCs VDC C

32 port 10GE module Once a port has been assigned to a VDC, ll b t fi ti done f from all subsequent configuration iis d within that VDC On 32-port 10GE module ports must be assigned to a VDC by 4-block groups.

VDC B

VDC C

http://www.cisco.com/en/US/docs/switches/datacenter/sw/4_1/nx-os/virtual_device_context/configuration/guide/ vdc_overview.html#wp1073104
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

42

VDC Granularity for 10/100/1000 Ports


On the 10/100/1000 card each port can be on a different VDC regardless of the adjacent ports (limited of course by the total of 4 VDCs) Using VDC it is possible to move servers seamlessly from a staging environment for example, to a production environment in the topology without having to re-cable the servers

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

43

Virtual Device Contexts


VDC Resource Utilization (Layer 3)
VDC 10
Linecard 1
FIB TCAM

VDC 20
Linecard 2
FIB TCAM

VDC 30
Linecard 4
FIB TCAM

FIB and ACL TCAM resources are more effectively utilized

Linecard 3
FIB TCAM

Linecard 5
FIB TCAM

Linecard 6
FIB TCAM

Linecard 7
FIB TCAM

Linecard 8
FIB TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

128K ACL TCAM

64K

64K

64K

64K

64K

64K

64K

64K

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

44

Layer 2 Links, All Forwarding and No Loops


vPC
LAN
Virtual Switch

MAC Pinning
LAN

L2MP
LAN

vPC/MEC

MAC A

MAC B Active-Active

L2 ECMP

MAC A

MAC B

L2 ECMP

Virtual Switch (VSS on C6K, ( , vPC on Nexus 7K) Virtual port channel mechanism is transparent to hosts or switches connected to the virtual switch STP as fail-safe mechanism to prevent loops even in the case of control plane failure

Host Mode Eliminates STP on Uplink Bridge Ports Allows Multiple Active Uplinks Switch to Network Prevents Loops by Pinning a MAC Address to Only One Port Completely Transparent to Next Hop Switch

Uses ISIS based topology Up to 16 way ECMP Eliminates STP from L2 domain Preferred path selection

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

45

vPC Terminology
STP Root vPC FT link

vPC peer a vPC switch, one of a pair vPC member port one of a set of ports (port channels) that form a vPC vPC the combined port channel between the vPC peers and the downstream device vPC peer-link Link used to synchronize state between vPC peer devices, must be 10GbE vPC ft-link the fault tolerant link p , , between vPC peer devices, i.e., backup to the vPC peer-link

vPC Peer link STP Secondary Root vPC peer

vPC member Ports

10 Gig uplinks

CFS Cisco Fabric Services protocol, used for state synchronization and configuration validation between vPC peer devices

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

46

vPC Layer 2 Processing (i.e. Etherchannel)


Notice that the Peer-link is almost unutilized

Etherchanneling modified to keep traffic local Downstream Switch runs LACP

hashing enhanced to keep traffic local

LACP Unmodified Portchanneling

10 Gig uplinks

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

47

vPC: Layer 3 Traffic Processing


Notice that the Peer-link is almost unutilized
HSRP active process communicates the active MAC to its neighbor. Only the HSRP active process responds to ARP requests

HSRP MAC populated in the Layer 2 table with the R flag

HSRP primary HSRP standby y

10 Gig uplinks

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

48

vPC Versus VSS


vPC Control Plane SSO HSRP Etherchannel to prefer local links Failover time Configuration synchroniziation Split Brain detection
TECDCT-3873_c2

VSS Unified Yes 1 single IP address, i.e. NO HSRP yes subsecond Yes, automatically done because of the unified CP Yes via BFD and PagP+
49

Separated Yes (2 sups per chassis) 2 entities yes In the order of seconds in the current release CFS to verify configurations and warn about mismatches Yes via the Fault Tolerant link
Cisco Public

2009 Cisco Systems, Inc. All rights reserved.

Pinning
1 Border interface 2 3 4

Server interface (SIF)

A
TECDCT-3873_c2

B
Cisco Public

F
50

2009 Cisco Systems, Inc. All rights reserved.

Outgoing Traffic Known Unicast


1 2 3 4

Traffic sourced by a station y connected to a SIF can go to one of the locally connected servers Or, if no local match is found, goes out of its pinned border interface

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

51

Outgoing Traffic Multicast/Broadcast


1 2 3 4

Local replication to all SIFs is done by the End Host Virtualizer switch One copy of the packet is sent out of the source SIFs pinned border interface

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

52

Incoming Traffic Reverse Path Forwarding


1 2 3 4

Reverse Path Forwarding protects from Loops Packets destined to a station behind a SIF are accepted only by the SIF pinned border interface

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

53

Incoming Traffic Multicast/ Broadcast Portal


1 2 3 4

Multicast/Broadcast Portal protects from Loops One border interface is elected to receive broadcast, multicast and unknown unicast traffic for all the SIFs

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

54

Incoming Traffic Deja-vu Check


1 2 3 4

The Deja-vu check prevents Loops If the source MAC belongs to a local station
The multicast/broadcast portal drops the packet The pinned port accepts the packet, but no replication is done This is regardless of the destination MAC (known/unknown unicast, multicast or broadcast)

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

55

Pinning Configurations (1)


correct configuration

Border interface

Server interface (SIF)


incorrect configuration

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

56

Pinning Configurations (2)


all Border Interfaces of the same subnet must be in the same L2 domain

Border interface

Server interface (SIF)


Virtual Switching can be connected to End Host Virtualizer

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

57

Layer 2 Multipathing
Clos Networks

L2

L2

Layer 2 MultiPathing enables designs that up until today were only possible with Infiniband
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

58

Layer 2 Multipathing
Edge switches
Determine which Edge id can reach a given MAC address Set the destination id IS-IS computes shortest path to id

Core switches
Forward from Edge switch to Edge switch based on destination id IS-IS computes shortest path to id

Source MAC sends to Destination MAC Edge switch does lookup for id attached to Destination MAC
If found, forward based on id If not found, flood on broadcast tree

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

59

Core Forwarding Table

FORWARDING TABLE on 3

L2

Core l1

3 l2

Destination Switch 1

Link L1 L2 N/A L1,L2 L1,L2

L2

Edge

Switch 2 Switch 3 Switch 4

M A C
TECDCT-3873_c2

Switch 5

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

60

Edge Forwarding Table

FORWARDING TABLE on 1

L2

Core l1

3 l2

Destination
l3 2

Link Directly Switch 2

L2

MAC A, B, C MAC D, E, F

Edge

M A C
TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

61

Server Connectivity Evolution Present


Management Challenges Shift towards server virtualization Multiple VMs inside each physical server, connected by virtual switches Rapid proliferation of logical elements that need to be managed Feature parity issues between virtual and physical elements

VSwitch

VSwitch

VSwitch

VSwitch

vNICs

vNICs

vNICs

vNICs

VMs

VMs

VMs

VMs

Separate management of physical (


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

) and logical (

) elements
62

Server Connectivity Evolution Future


Future with Network Interface Virtualization and VNTAG: Consolidated Management
Virtual Interfaces within VMs are now visible to the switch Both network configuration and policy enforcement for these interfaces can now be driven from the switch This allows consolidated management of physical and virtual elements

VSwitch

VSwitch

VSwitch

VSwitch

vNICs

vNICs

vNICs

vNICs

VMs

VMs

VMs

VMs

Consolidated management of physical (


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

) and logical elements


63

Interface Virtualizer (IV) Architecture

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

64

VNTAG

VNTAG Format
VNTAG Ethertype l source virtual interface
SA[6]

d p

destination virtual interface

DA[6]

direction indicates to/from adapter source virtual interface indicates frame source
looped indicates frame came back to source adapter

VNTAG[6] 802.1Q[4]

destination virtual interface dictates forwarding


pointer helps pick specific destination vNIC or vNIC list

Frame Payload

Link local scope


Rooted at Virtual Interface Switch 4096 virtual interfaces 16,384 Virtual interface lists

CRC[4]

Coexists with VLAN (802.1Q) tag


802.1Q tag is mandatory to signal data path priority
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

65

VNTAG Processing (1)


SAN LAN

Interface Virtualizer adds VNTAG


Unique source virtual interface for each vNIC d (direction) = 0 p (pointer), l (looped), and destination virtual interface are undefined (0) Frame is unconditionally sent to the Switch

Virtual Interface Switch

Interface Virtualizer v v OS v v OS v v OS

Application Payload P l d TCP VNTAG Ethertype l source virtual interface d p destination virtual interface IP VNTAG Ethernet

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

66

VNTAG Processing (2)


SAN LAN

Virtual Interface Switch ingress processing


Extract VNTAG Ingress policy based on port and source virtual interface Access control and forwarding based on frame fields and virtual interface policy Forwarding selects destination p ( ) and g port(s) destination virtual interface(s) VIS adds a new VNTAG
access control & forwarding policy

Virtual Interface Switch

Interface Virtualizer v v OS v v OS v v OS

Application Payload P l d TCP IP Ethernet

VNTAG

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

67

VNTAG Processing (3)


SAN LAN

Virtual Interface Switch egress processing


Features from port and destination virtual interface Insert VNTAG(2)

Virtual Interface Switch

Interface Virtualizer

direction is set to 1 destination virtual interface and pointer select a single vNIC or list source virtual interface and l (looped) filter a single vNIC if sending frame to source adapter

v v OS

v v OS

v v OS

Application Payload P l d TCP VNTAG Ethertype l source virtual interface d p destination virtual interface IP VNTAG(2) Ethernet

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

68

VNTAG Processing (4)


SAN LAN

Interface Virtualizer (IV) forwards based on VNTAG


Extract VNTAG Upper layer protocol features from frame fields destination virtual interface and pointer select vNIC(s) source virtual interface and looped filter a single vNIC if source and destination are same IV
ULP features v v OS v v OS vNIC forwarding

Virtual Interface Switch

Interface Virtualizer v v OS v v OS v v OS

Application Payload P l d TCP IP Ethernet

x
v v OS v v OS v v OS

x
v v OS

VNTAG(2)

Unicast (single vNIC)


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

Multicast (vNIC list)


Cisco Public

69

VNTAG Processing (5)


SAN LAN

OS stack formulates frames traditionally Interface Virtualizer adds VNTAG Virtual Interface Switch ingress processing Virtual Interface Switch egress processing Interface Virtualizer forwards based on VNTAG OS stack receives frame as if directly connected to Switch

Virtual Interface Switch

Interface Virtualizer v v OS v v OS v OS

Application Payload P l d TCP IP Ethernet

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

70

VNTAG + MAC Pinning


Interface Virtualizers connect to the network in a redundant fashion Redundancy can be addressed using MAC pinning: each downlink port is associated with an uplink port Forwarding is based on a VIF forwarding table which is made of 1024 entries For multicast traffic, a VIF_LIST table is indexed by a VIF_LIST_ID and the result is a bitmask indicating which SIF ports should the frames be sent to.

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

71

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server Break Demo: Nexus1kv
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

72

Nexus 2000 Fabric Extender


Network Topology Physical vs. Logical Physical Topology
Core Layer

Logical Topology
Core Layer

VSS

VSS

L3 L2
4x10G uplinks from each rack
FE

L3 L2

Nexus 5020

Nexus 5020 Nexus 5020 Nexus 5020


12 FEX

FEX

FEX

FEX

FEX

FEX

FEX

12 FEX

Servers
Servers

Rack-1

Rack-N

Rack-1

Rack-N

Rack-1

Rack-2

Rack-3

Rack-4

Rack-5

Rack-12

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

73

Data Center Access Architecture


Distributed Access Fabric
De-Coupling of the Layer 1 and Layer 2 Topologies Optimization of both Layer 1 (Cabling) and Layer 2 (Spanning Tree) Designs Mixed cabling environment (optimized as required) Flexible support for Future Requirements

...
Nexus 5000/2000 Mixed ToR & EoR
Combination of EoR and ToR cabling
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

74

Cabling Design for FEX


Copper Connectivity

Top of Rack Fabric Extenders provide 1G server connectivity Nexus 5000 in Middle of Row connects to Fabric Extenders with CX1 copper 10G pp between racks Suitable for small server rows where each FEX is no longer than 5 meters from the 5Ks CX1 copper between racks is not patched Middle of Row Nexus 5000 can also provide 10G server connectivity within their rack

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

75

FEX Inner Functioning


Inband Management Model Fabric extender is discovered by switch using an L2 Satellite Discover Protocol (SDP) that is run on the uplink port of fabric extender NX5K checks software image compatibility, assign an IP address and upgrade the fabric extender if necessary N5K pushes programming data to Fabric Extender Fabric Extender updates the N5K with its operational status and statistic. Extension to existing CLI on N5K is used for Fabric Extender CLI information
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

76

FEX Design Considerations


Uplink Dual Homing
Without vPC support N5K-A N5K-B With vPC support N5K A N5K-A N5K B N5K-B

SDP exchange

Err-disable

Static pinning is not supported in a redundant supervisor mode Server ports appear on both N5K Currently configuration for all ports must be kept in sync manually on both N5Ks
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

77

FEX Design Considerations


Server Dual Homing
vPC provides two redundancy designs for the virtualized access switch Option 1 - MCEC connectivity from the server Two virtualized access switches bundled into a vPC pair Logically a similar HA model to that currently provided by VSS
Two Virtualized access switches Each with a Single Supervisor

vPC peers

MCEC from server to the access switch


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

78

FEX Design Considerations


NIC Teaming with 802.3ad Across Two FEX Devices

N5KA N5K

N5KB

By leveraging vPC it is possible to create 802.3ad configurations with dual-homed servers


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

79

FEX Design Considerations


MAC Pinning on Fabric Extender (FEX)
Static Pinning Fabric Extender associates (pins) a server side (1GE) port with an uplink (10GE) port Server ports are either individually pinned to specific uplinks (static pinning) or all interfaces pinned to a single logical port channel Behavior on FEX uplink failure depends on the configuration g p pinned to Static Pinning Server ports p the specific uplink are brought down with the failure of the pinned uplink Port Channel Server traffic is shifted to remaining uplinks based on port channel hash
Server Interface stays active
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

NIC teaming required

Port Channel

80

FEX Design Considerations


N2K/N5K Spanning Tree Design Considerations
Root Bridge HSRP Active Secondary Root Bridge HSRP Standby y

BPDU Guard

Bridge Assurance

Global BPDU Filter reduces the spanning tree load (BPDUs generated on a Host Port) VMWare S VMW Server Trunk T k Needs to Carry Multiple VLANs which can increase the STP load
VSwitch
VM #1 VM #2 VM VM #3 #4

UDLD

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

81

FEX Design Considerations


vPC - Spanning Tree Design Considerations
Enabling vPC on the access to aggregation links improves y y layer 2 scalability Removing physical loops out of the layer 2 topology Reducing the STP state on the access and aggregation layer The use of vPC does result in a reduction of logical port count on the aggregation but does involve CFS synchronization of state between the two aggregation nodes
Fabric Links (No Spanning p g Tree) vPC
Single Logical Link to STP Both vPC Peers Act as the default GW

Server Ports BPDU Guard

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

82

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server Break Demo: Nexus1kv
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

83

vPC Configuration Commands


Configure vPC, and start the ft-link on both peers:
(config)# feature vpc (config)# vpc domain 1 (config-vpc-domain)# peer-keepalive destination x.x.x.x source x.x.x.y (conifg)# int port-channel 10 (config-int)# vpc peer-link

Move any port-channels into appropriate vPC groups


(config)# int port-channel 20 (config-int)# vpc 20

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

84

vPC Control Plane vPC Role


vPC domain is identified by a configured ID, and after successful establishment of peer-link peer link adjacency, vPC domain is operationally enabled.
MAC-address derived from domainID is used for link-specific protocol operations (LACP lag-id for vPC, designated bridge-id for STP)
vPC primary vPC secondary

vPC domain

vPC election generates vPC role (primary/secondary) for each switch. vPC role is used only when dual-active topology is detected.

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

85

vPC Control Plane FT Link


FT link (can be routed)

vPC FT (fault-tolerant) link is an additional mechanism to detect liveness of the peer. can peer use any L3 port. By default, will use management network.
used only when peer-link is down does NOT carry any state information

VRF FT

VDC A (e.g. 2)

Peer-link

Rare lik lih d of d l R likelihood f dualactive topology vPC is within the context of a VDC

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

86

vPC Deployment
Recommended Configurations
vPC is a Layer 2 feature Port has to be in switchport mode before configuring vPC vPC/vPC peer link support following port/ peer-link port-channel modes Port Modes: Access or Trunk Port-channel Modes: On mode or LACP (active/passive) mode Recommended port mode Trunk vPC peer-link should support multiple VLANs and should trunk the access VLANs Recommended port-channel mode is Link Aggregation Control Protocol (LACP). Dynamically react to runtime changes and failures Lossless membership change Detection of mis-configuration Maximum 8 ports in a port-channel in on-mode and 16 ports with 8 operational ports in a LACP port-channel
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

VRF FT

VDC A (e.g. 2)

Peer-link

LACP

87

vPC Control Plane


CFSoE
l2fm igmp stp vpcm { opcode, payload} vpcm stp igmp l2fm vpc-transport-api vpc-transport-api

cfs

cfsoe
netstack

sw-1

sw-2

cfs

cfsoe
netstack

CFS (Cisco Fabric Service), over Ethernet (CFSoE), provides a reliable transport layer to all applications that need to co-operate with peer vPC switch. CFSoE
uses retransmissions & acknowledgements per segment transmitted. supports fragmentation and re-assembly for payloads more than MTU uses BPDU class address, and is treated with highest QoS/drop-thresholds.

Each component has (one or more) request-response handshakes (over CFSoE) with its peer. Protocols (STP/IGMP/FHRP) continue to exchange regular protocol BPDUs. In addition, theyll use CFS for state synchronization

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

88

CFS Distribution
CFS only checks that the VLANs assigned to a vPC are the same on both devices that are on the same vPC This warns the person on the other 7k that he has to make configuration changes to include the same exact VLANs Distribution is automatically enabled b enabling vPC by PC (config)#cfs distribute enable ( (config)#cfs ethernet g) distribute enable tc-nexus7k01-vdc3# show cfs status Distribution: Enabled Distribution over IP: Disabled IPv4 multicast address: 239.255.70.83 IPv6 multicast address: ff15::efff:4653 Distribution over Ethernet: Enabled
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

89

CFSoIP vs CFSoE
vPC uses CFSoE, Roles Leverage CFSoIP
vPC domain (CFSoE) CFSoIP Cloud

Role Defintion

The user creates new Role User commits the changes Role get automatically propagated to the other switches
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

90

Type-1 Compatibility Parameters


Port Channel is disabled if one of the following parameters is mismatched
Port-channel Speed (10M, 100M, 1000M or 10G) Port-channel Duplex (half or full) Port Mode (access or trunk) Port-channel MTU Port-channel Native VLAN Port-channel mode (on, active or passive)

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

91

Detecting Mis-Configuration
Sw1 (config)# show vpc brief VPC domain id : Peer status : VPC keep-alive status : Configuration consistency status: 1 peer adjacency formed ok Disabled success

VPC status --------------------------------------------------id Port Consistency Reason ---- -------------- ----------- ---------------1 Po2 success success 2 Po3 failed vpc port channel mis-config due to vpc links in the 2 switches connected to different partners

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

92

vPC Failure Scenarios Peer-Link Failure (Link Loss)


vPC primary vPC secondary

In case vPC peer-link fails Check active status of remote vPC peer via vPC ft-link (heartbeat) If both peers are active, then Secondary will disable all vPC ports to prevent loops Data will automatically forward down remaining active port channel ports Failover gated on CFS message F il t d failure, or UDLD/Link state detection

CFSoE

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

93

vPC Failure State Diagram


Start CFS message delivery failure? vPC secondary peer? Yes Suspend vPC member ports Recover vPC member ports No

No vPC Peer link failed? (UDLD/Link state) Yes vPC ft-link heartbeat detect? No Other processes take over based on priority (STP root, HSRP active, PIM DR) Yes

Yes

No

vPC peer recovered?

Yes

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

94

vPC Between Sites and Within Each DC


CFSoE Region 1 DC1 N7kA-DC1 Eth2/3 access Eth2/26 Eth2/9 Po60 Eth2/25 Eth8/40 Eth8/4 vPC between sites DC2 N7kC-DC2 N7kC DC2 Eth7/3 Eth7/9 Eth8/5 Eth7/25 Po50 CFSoE Region 2

Peer link
Eth2/9

Peer link
Eth7/9 Eth7/25 Eth8/5

Eth2/25

Eth2/26

Eth2/3 Po30

Eth7/3 N7kD-DC2

N7kB-DC1

Links Protected by IEEE 802.1ae FT link


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

95

Routing Design
for the Extended VLANs
DC1 gw 1.1.1.1 DC2 gw 1.1.1.2 Failover direction SRP Group 2 (e.g. 1.1.1.2) for HS

150, 150 120


Failover di irection for HSRP Group 1 (e.g. 1.1.1.1) G

120, 120 150

HSRP Group 1

HSRP Group 2

140, 140 130

130, 130 140

G 60 0000.0c07.ac3c static << group that is active or standby * 60 0000.0c07.ac3d static << group that is listen mode G 60 0000.0c07.ac3d static << group that is active or standby * 60 0000.0c07.ac3c static << group that is listen mode
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

96

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

97

Why Is a Virtual Switch Needed in the First Place


Forwarding Table

Destination MAC MAC1


DMAC = MAC2 Ethernet1/1 DMAC = MAC2

Port 1/1 1/1

MAC2

MAC1

MAC2

?
VM2
Cisco Public

VM1

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

98

Virtual Switching Virtualized Servers Need VN-link Technology


Forwarding Table

Destination MAC MAC1


Ethernet1/1

Port 1/1 1/1

MAC2

vSwitch or Nexus 1000v

=
MAC1 MAC2 Nexus1kv

VM1

VM2
99

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

ESX Server Components


VMware ESX is a bare-metal hypervisor that partitions physical servers in multiple virtual machines Virtual Machine
App App App

OS

OS

OS

vnics S Software virtual switch

vmnics

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

100

Nexus 1000v
Distributed Virtual Switch
N1k-VSM# sh module

Linecards Equivalent

Mod Ports Module-Type Model Status 1 1 Supervisor Module Cisco Nexus 1000V active * 2 1 Supervisor Module Cisco Nexus 1000V standby 3 48 Virtual Ethernet Module ok 4 48 Virtual Ethernet Module ok

App

App

App

App

App

App

App

App

App

App

App

App

App

App

App

App

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

Hypervisor

Hypervisor

Hypervisor

Hypervisor

vCenter

Virtual Ethernet Module

Fabric Function
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Virtual Supervisor Module


101

Nexus 1000V
Virtual Interface
veth = Virtual Machine port (vnic)
Hypervisor
App App App App OS OS OS OS

veth3

veth7 veth68

N1k-VSM# sh interface virtual Port Adapter Owner Veth3 Veth7 Veth68

Mod Host

Net Adapter 1 Ubuntu VM 1 pe-esx1 Net Adapter 1 Ubuntu VM 2 pe-esx1 Net Adapter 1 Ubuntu VM 3 pe-esx1

Cisco VSMs

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

102

Nexus 1000v
Ethernet Interface

App

App

App

App

Eth = uplink port on the ESX Server

eth3/1 th3/1

OS

OS

OS

OS

eth3/2

Hypervisor

WS-C6504E-VSS#sh cdp neighbors Device ID Local Intrfce Platform N1k-VSM N1k-VSM N1k-VSM N1k-VSM Gig Gig Gig Gig 1/1/1 2/1/2 1/8/1 2/8/2 Nexus1000 Nexus1000 Nexus1000 Nexus1000

Port ID Eth Eth Eth Eth 3/1 3/2 4/1 4/2

eth4/1

App

App

App

App

OS

OS

OS

OS

eth4/2
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Hypervisor
103

What Is a Port Profile?


n1000v# show port-profile name WebProfile port profile port-profile WebProfile description: status: enabled capability uplink: no system vlans: port-group: WebProfile config attributes: switchport mode access switchport access vlan 110 no shutdown evaluated config attributes: switchport mode access it h t d switchport access vlan 110 no shutdown assigned interfaces: Veth10

Support Commands Include: Port management VLAN PVLAN Port-channel ACL Netflow Port Security QoS

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

104

Port-Profile as Viewed from the Network Administrator and Server Administrator


Network Administrator view
N1k-VSM# sh port-profile name Ubuntu-VM port-profile Ubuntu-VM description: status: enabled capability uplink: no capability l3control: no system vlans: none port-group: Ubuntu-VM max-ports: 32 inherit: config attributes: switchport mode access switchport access vlan 95 no shutdown assigned interfaces: Vethernet2 Vethernet4
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Server admin view

105

What Makes the Virtual Switch Distributed?


ESX servers that are under the same Nexus 1kv VSM share the same PortProfile Configuration When a new Port-Profile is defined it gets automatically propagated to all the ESX servers (VEMs) that are the VSM In this example ESX1 and ESX2 are under VSM1 and share the green and red Port-Profile ESX3 and ESX4 are under VSM2 and share the Blue and Yellow Port Profile
1 2 3 4

VSM1
Cisco VSMs

VSM2
Cisco VSMs

Port Profiles

Port Profiles

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

106

Prior to DVS Ensuring Port-Group Consistency Was a Manual Process

Each ESX host is configured individually for Networking


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

107

VMotion Requires the Destination vSwitch to Have the Same Port Groups/Port-Profiles as the Originating ESX Host
Rack1 Rack10

c0 vmnic0 ESX Host 1 vSwitch

vmnic1

Prior to DVS you had to manually ensure that the same Port-Group existed on ESX Host 1 as ESX Host 2

vmnic0 ESX Host 2

vmnic1

vSwitch
App OS

App OS

App OS

App OS

App OS

App OS

VM1
TECDCT-3873_c2

VM2

VM3
Cisco Public

VM4

VM5

VM6
108

2009 Cisco Systems, Inc. All rights reserved.

Distributed Virtual Switching Facilitates VMotion Migration


Port Profiles

Server 1
VM #1 VM #2 VM #3 VM #4 VM #1

Server 2
VM #2 VM #3 VM #4

VEM VMW ESX

VEM VMW ESX

VMs Need to Move


VMotion DRS SW Upgrade/Patch Hardware Failure
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

109

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

110

Configuring Access-Lists, Port Security, SPAN, etcWithout Nexus1kv Is Complicated


Is VM#1 on Server 1? Or on which server, on which switch do I put the ACL?
VM #4

Server 1
VM #1 VM #2 VM #3

vSwitch
VMW ESX

ACL need to be specify the IP address of the VM else you risk to drop both VM1 and VM3 traffic SPAN will get all traffic from VM1, VM2, VM3, VM4!! You need to filter that!! Port Security CANT be used

ACLs (complicated)

SPAN (realistically cant be used)

Port Security needs to be disabled

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

111

You Can Use Access-Lists, Port Security, SPAN, etcWITH Nexus1kv


ACLs specific to a Port-Group Server 1
VM #1 VM #2 VM #3 VM #4

Is VM#1 on Server 1? It doesnt matter ACL follows the VM SPAN will get only the traffic from the virtual Ethernet Port Port Security ensures that VMs wont generate fake make addresses

VEM VMW ESX Port Security

SPAN on a virtual ethernet port

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

112

vNIC Security
Server

VMs can be secured in multiple ways:


VLANs ACLs Private VLANs Port-Security

VM #1

VM #2

VM #3

VM #4

vnics i
Nexus 1000 DVS

vmnic IEEE 802.1q trunk

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

113

Private VLANs Can Be Extended Across ESX Servers by Using the Nexus1kv
Promiscuous ports receive and transmit to all hosts Communities allow communications between groups Isolated ports talk to promiscuous promisc o s ports only
App App App App App App App

Promiscuous Port

Only One Subnet

Promiscuous Port

x x
App

Primary VLAN
OS OS OS OS OS OS OS OS

Community VLAN Community VLAN Isolated VLAN


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved.

.11

.12

.13

.14

.15

.16

.17

.18

Community A
Cisco Public

Community B

Isolated Ports
114

Tracing Virtual Ethernet Ports


show interface VEthernet Vethernet2 is up Hardware is Virtual, address is 0050.5675.26c5 Owner is VMware VM1, adapter is vethernet1 Active on module 8, host tc-esx05.cisco.com VMware DVS port 16777215 Port-Profile is MyApplication Port mode is access Rx 444385 Input Packets 444384 Unicast Packets 0 Multicast Packets 1 Broadcast Packets 572675241 Bytes Tx 687655 Output Packets 687654 Unicast Packets 0 Multicast Packets 1 Broadcast Packets 1 Flood Packets 592295257 Bytes 0 Input Packet Drops 0 Output Packet Drops
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

115

SPAN Traffic to a Catalyst 6500 or a Nexus 7k Where You Have a Sniffer Attached
Capture here
App App App App App App App App App App App App

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

OS

Hypervisor Virtual Ethernet Module

Hypervisor Virtual Ethernet Module

Hypervisor Virtual Ethernet Module

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

116

Ease of Provisioning
Plug-and-play designs with VBS
1 Add or replace a VBS Switch to the Cluster 2 Switch config and code automatically propagated Virtual Ethernet Module

3 Add a blade Server 4 Its always booted from the same LUN

Virtual Ethernet Module

Virtual Ethernet Module

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

117

Ease of Provisioning
Making Blade Servers Deployment Faster
1 Physically Add a new blade (or replace an old one)

2 Go to vCenter, add host to cluster

Virtual Ethernet Module


Nexus 1000v

3 Done: the new blade is in production All port-groups appear

Virtual Ethernet Virtual Ethernet Module Module

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

118

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

119

Cisco Nexus 1000V Switch Interfaces


Ethernet Port (eth)
1 per physical NIC interface Specific to each module vmnic0 = ethx/1 Up to 32 per host
Eth3/1 Po1

Eth3/2

Port Channel (po)


Aggregation of Eth ports Up to 8 Port Channels per host
VM1 VM2

Veth1

Veth2

Virtual Ethernet Port (veth)


1 per VNIC (including SC and VMK) Notation is Veth(port number). 216 per host

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

120

Loop Prevention without STP

Eth4/1 Cisco VEM Cisco VEM

Eth4/2

X
Cisco VEM

X
VM1 VM2 VM3 VM4 VM5 VM6 VM7 VM7 VM9 VM10 VM11 VM12

BPDU are dropped

No Switching From Physical NIC to NIC

Local MAC Address Packets Dropped on Ingress (L2)


121

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

MAC Learning
Each VEM learns independently and maintains a separate MAC table VM MACs are statically mapped
Other vEths are learned this way (vmknics and vswifs) No aging while the interface is up
Eth3/1 Cisco VEM Eth4/1 Cisco VEM

Devices external to the VEM are learned dynamically

VM1

VM2

VM3

VM4

VEM 3 MAC Table


VM1 VM2 VM3 VM4 Veth12 Veth23 Eth3/1 Eth3/1 Static Static Dynamic Dynamic VM1 VM2 VM3 VM4

VEM 4 MAC Table


Eth4/1 Eth4/1 Veth8 Veth7 Dynamic Dynamic Static Static

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

122

Port Channels

Standard Ci St d d Cisco Port Channels P t Ch l


Behaves like EtherChannel

Link Aggregation Control Protocol (LACP) Support 17 hashing algorithms available


Selected either system wide or per module Default is source MAC
VM1 VM2 VM3 VM4

Po1

Cisco VEM

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

123

Virtual Port Channel Host Mode


Allows a VEM to span multiple upstream switches using subgroups Forms up to two subgroups based on Cisco Discovery Protocol (CDP) or manual configuration Does not support LACP veths are associated in a round robin to a subgroup and then hashed within a subgroup Does not require a port channel p g g upstream when using single link in each sub-group Required when connecting a port channel to multiple switches unless MCEC is configured on the access side
CDP received from the same switch creates the sub-group bundle
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

N5K View

SG0

Po1

SG1

VEM View

VM #2

VM #3

VM #4

124

Automated Port Channel Configuration


Port channels can be automatically formed using port profile Interfaces belonging to different modules cannot be added t same I t f b l i t diff t d l t b dd d to channel-group. E.g. Eth2/3 and Eth3/3 auto keyword indicates that interfaces inheriting the same uplink port-profile will be automatically assigned a channel-group.
n1000v(config)# port-profile Uplink n1000v(config-port-prof)# channel-group auto

Each interface in the channel must have consistent speed/duplex Channel-group does not need to exit and will automatically be created

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

125

Uplink Port Profiles


Special profiles that define physical NIC properties Usually configured as a trunk Defined by adding capability uplink to a port profile Uplink profiles cannot be applied to vEths Non-uplink profiles cannot be applied to NICs Only l t bl in C t O l selectable i vCenter when adding a h ddi host or additional NICs
n1000v(config)# port-profile DataUplink n1000v(config-port-prof)# switchport mode trunk n1000v(config-port-prof)# switchport trunk allowed vlan 10-15 n1000v(config-port-prof)# system vlan 51, 52 n1000v(config-port-prof)# channel-group mode auto sub-group cdp n1000v(config-port-prof)# capability uplink n1000v(config-port-prof)# no shut
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Cisco VEM

VM1

VM2

VM3

VM4

126

System VLANs
System VLANs enable interface connectivity before an interface is programmed
i.E VEM cant communicate with VSM during boot

Cisco VSM

Required System VLANs


Control
L2 Cloud

Packet

Recommended System VLANs y


IP Storage Service Console VMKernel Management Networks
Cisco VEM C P

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

127

Four NIC Configuration


Access Layer Configuration
Trunk port

No EtherChannel

N1KV Port Channel 1


vPC-HM VM Data
SG0 SG1 Po1

N1KV Port Channel 2


vPC-HM Service Console, VM Kernel, Control and Packet

SG0 SG1 Po2

Cisco VEM Ci
C P

VEM Configuration
Source Based Hashing

Use Case
Medium 1Gb servers (rack or blade) Need to separate VMotion from Data
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

SC
VM Data

VMK

128

Four NIC Configuration with N2k w/o vPC


In a Four NIC implementation Access switch configured with Trunk g ports (no Etherchannel) VEM Configured with SRC based hashing N1KV Port Channel 1 (vPC-HM)
VM Data
Trunk Edge Port supporting only the VM VLANs

N1KV Port Channel 2 (vPC-HM) ( )


Service Console, VM Kernel, VEM Control and Packet

SG0

SG1

SG0

SG1

VM traffic carried on a second vPCHM uplink bundle

VM

VMK

SC

SC and VMK traffic carried on one upstream vPC-HM uplink bundle

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

129

Four NIC Configuration with vPC


Using 2 Separate Regular Port-Channels
Access switch configured with two server vPC MCEC trunk ports VEM C fi Configured with L3/L4 b d ith based d hashing N1KV Port Channel 1
VM Data

N1KV Port Channel 2


Service Console, VM Kernel, VEM Control and Packet

vPC MCEC Bundles

VM traffic carried on a second uplink bundle

VM

VMK

SC

SC and VMK traffic carried on one upstream uplink bundle

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

130

Four NIC Configuration with vPC


Using a Single vPC-HM of Four Ports
Combine vPC-HM and MCEC vPC to load share traffic across four NICs Access switch configured with t A it h fi d ith two server vPC MCEC trunk ports VEM Configured with SRC based hashing N1KV Port Channel 1 (vPC-HM)
VM Data
vPC MCEC Bundles

Do not use CDP to create the sub-groups in this type of topology (manually configure the sub-groups)

SG0

SG1

VM 1

VM 2

VM 3

Single shared upstream vPCHM comprised of four links

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

131

Cisco Nexus 1000V Scalability


A single Nexus 1000V supports:
2 Virtual Supervisor modules (HA) 64* Virtual Ethernet modules 512 Active VLANs 2048 Ports (Eth + Veth) 256 Port Channels
Nexus 1000V

A single Virtual Ethernet module supports:


216 Ports Veths 32 Physical NICs 8 Port Channels
* 64 VEMs pending final VMware/Cisco scalability testing ** Overall system limits are lower than VEM limit x 64
132 Cisco VEM

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

133

Virtual Machine Considerations


Hardware MAC learning Large HW-based MAC address Tables Control plane policing Layer 2 trace Broadcast and Storm Control Private VLAN integration Unified I/O ready
Virtual Servers

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

134

10 Gigabit Server Connectivity


VNTAG / Nexus 1000v

FCoE

10 Gigabit Ethernet

Class-Based Bandwidth Allocation

DCE

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

135

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels Scalability Considerations LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

136

With Nexus1kv the Switch Just a Plug-and-Play Fabric


With the Nexus1kv the Profiles are defined on the Nexus1kv The Mapping is p pp g performed on the Virtual Center The Switch provides simply the Switching Fabric and trunks all necessary VLANs.

Nexus1kv Mapping of servers t M i f to VLANs/Port Profiles Profile Definition vCenter C t Nexus1kv CLI

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

137

Switching Fabric With Virtualized Servers


You have Virtualized Servers on the Blades You are better off using clustered Cisco VBS g Cisco VBS Network Management Model Stacking Capability Etherchanneling Server Identity Equivalent to a 3750 stackable: plug-andplay Up to 8 Blade Switches Switches, i.e. single config point Across switches in the stack Flexattach

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

138

Nexus 1000v With Blade Enclosures


Port-Profile Definition

App App App

App

App App App App

App App App

App

App App App App

OS

OS

OS

OS

OS OS OS OS

OS

OS

OS

OS

OS OS OS OS

Fabric Function 10 Gigabit Uplinks


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

139

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and design with virtual Port Channeling Break Demo: vPC Designs with Server Virtualization
Nexus1kv Components Operational benefits VEM Forwarding: NIC Teaming and Etherchannels Scalability Considerations LAN switching infrastructure requirements Designs with Blade Servers

10 Gigabit Ethernet to the Server Break Demo: Nexus1kv

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

140

Todays Data Center Networks

LAN HPC

SAN A

SAN B

Ethernet FC High Perf. Comp. (HPC)

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

141

Consolidation Vision
Why? VM integration Cable Reduction Power Consumption reduction Foundation for Unified Fabrics IPC
(*) RDMA = Remote Direct Memory Access (**) iWARP = Internet Wide Area RDMA Protocol
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

FCoE

142

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Designs with Server Virtualization Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server
10 Gigabit Ethernet Performance Considerations 10 Gigabit Performance in Virtualized Environments Datacenter Ethernet

Break Demo: Nexus1kv


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

143

10 Gigabit Adapters Typical Features


MSI-X (Message Signaled Interrupt) Support PCIe 8x for 10 Gigabit Performance g TCP Offload (TOE) in Hardware
Configurable TCP SACK (Selective Acknowledgement) (not really configurable) Checksum offload

Large Send Offload (LSO): allows the TCP layer to build a TCP message up to 64KB and send it in one call down the stack through the device driver. Segmentation is handled by the Network Adapter Receive Side Scaling queues: 2 4 or disabled. Allows distributing incoming traffic to the available cores. VLAN offload in Hardware NetDMA support
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

144

OS Enablers
TCP Chimney Offload Receive Side Scaling ( RSS g (+ capable NIC) In Windows 2003 this requires the Scalable Networking Pack (SNP). (SNP) In Windows 2008 this is already part of the OS.

Do not enable TSO in HW And disable TCP Chimney Or vice-versa!

Make sure to apply changes in:


DRIVER ADVANCED CONFIGURATIONS (which controls the 10 GigE Adapter HW) REGISTRY EDITOR

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

145

Evaluating 10 GigE Performance


The following distinctions need to be made to evaluate the 10 GigE adapter impact on the applications
TSO cards without proper OS support dont yield more than 3-4Gbps Throughput tests stress vs Transaction/s tests use different HW features You must distinguish TX performance vs RX performance TCP and UDP traffic are handled very differently in the HW y y TCP Checksum Offload and Large Segment Offload provide different functionalities.

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

146

Preliminary Tests
Maximum Throughput Is ~3.2 Gbps

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

147

Why?
Only 1 core is dealing with TCP/IP processing The OS doesnt know that the Adapter is TOE capable so it doesnt really use it A lot of memory copies between user space and kernel space Is the card plugged in the p gg PCIe x8? Solution:
Make sure that the OS uses TCP offloading in Hardware Enable Large Segment Offload Enable TCP/IP distribution to all available cores

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

148

Engaging More Than 1 Core: Receive Side Scaling (RSS)


Core 1 Core 2 Core 3 Core 4

CPU 1

CPU 2

RSS Capable NIC p Incoming Packets


Hash Interrupt Logic

Receive FIFOs

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

149

Processing W/O Large Segment Offload (LSO)


Core 1

application user I/O library


MSS

Data record
100%
MSS Data MSS MSS record MSS MSS

% CORE overhead

OS kernel device driver


MSS

TCP/IP
MSS

transport processing Intermediate buffer copies

40% 20%

I/O Adapter

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

150

Large Send Offload


V1 (Scalable Networking Pack):
allows the TCP layer to build a TCP message up to 64KB and send it in one call down the stack through the device driver. Segmentation is handled by the Network Adapter

V2 (Windows 2008):
allows the TCP layer to build a TCP message up to 256KB and send it in one call down the stack through the device driver. Segmentation is handled by the Network Adapter Supports IPv4/IPv6

Main Benefit: Reduces CPU utilization Key Use Cases: Large I/O applications such as Storage, backup, and ERP.
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

151

Processing With Large Segment Offload (LSO)


Core 1

application user I/O library

Data record
100%

% CORE overhead

OS kernel device driver


MSS MSS MSS MSS

Intermediate buffer copies

20%

I/O Adapter

TCP/IP

MSS

MSS

MSS

MSS

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

152

Registry Configuration (for Windows 2003)


In Windows 2008 Just Use netsh cmd

Set to 1

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

153

W/o LSO Checksum Offload Alone Doesnt Do Much

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

154

LSO Combined With TCP Offload Is Better

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

155

But the RX Side Cannot Keep Up With the TX Hence You Need to Enable SACK in HW

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

156

Enabling Jumbo Frames

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

157

Network Design Considerations for HPC Parallel Applications


Latency has an important effect on messaging between the nodes What matters is end-to-end application messaging, as opposed to network latency There is a big difference between regular TCP/IP stack, TCP/IP with TCP offloading (TOE), and RDMA (Remote Direct Memory Access) accelerated
Speedup

Key measurement factor: speedup Relevant protocols:


Message Passing Interface (MPI) MPI over Ethernet uses TCP

10 GigE with iWARP RDMA GigE Number of Nodes

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

158

Sources of Overhead in Datacenter Servers


Sources of Overhead in Server Networking Transport Processing Intermediate Buffer Copying Application Context Switches CPU Overhead 40% 20% 40% User Solutions for Overhead in Server Networking Transport Offload Engine (TOE) Moves Transport processor cycles to the NIC o es C / protocol stac buffer copies Moves TCP/IP p otoco stack bu e cop es from system memory to the NIC memory RDMA Eliminates intermediate and application buffer copies (memory bandwidth consumption) Kernel
TCP/IP

CPU

System Memory App Buffer

TCP Buffer Pool

s/w h/w
NIC

Kernel Bypass direct user-level access to hardware Dramatically reduces application context switches

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

159

iWARP
The Internet Wide Area RDMA Protocol (iWARP) is an Internet Engineering Task Force (IETF) update of the RDMA Consortium's RDMA over TCP standard. iWARP is a superset of the Virtual Interface Architecture that permits zero-copy transmission over legacy TCP. It may be thought of as the features of InfiniBand (IB) applied to Ethernet. http://www.openfabrics.org/ http://www openfabrics org/ runs on top of iWARP

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

160

Latency on the Switch


Latency of modular 1GbE sw can be quite high (>20us)
store & fwd many hops line serialization

Nexus 5k TOR fixes this


Cut through implementation 3.2 us latency

A single frame dropped in a sw or adapter causes significant impact on performance:


TCP NACK delayed by up to 125us with NIC with interrupt throttling enabled TCP window shortened (burst of traffic, lost of frame, slowdown most of traffic and brings burst again..etc.. for financial customer such as trading companies may suffer)
TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

161

Latency Fundamentals
What matters is the application-to-application latency and jitter
Driver/Kernel software Adapter Network components Kernel NIC NIC Kernel Application
Data Packet

Application

Latencies of 1GbE switches can be quite high (>20ms)


Store and forward Multiple hops Line serialization delay

N5000 Switch 3.2 s 3 2 s

Nexus 5000 Solution


Cut through implementation 3.2 ms latency (port to port with features turned on) End to End latency

Protocol processing dominates latency

Nexus 5000 in latency optimized application


162

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Latency

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

163

What Helps Where


Checksum Offload RSS LSO iWARP

TX RX CPU % TCP workload Transactions/s TCP workload Throughput UDP throughput Latency

+ + ++ + + + ++ ++
(4)

+++ +++ + +
(1,2) +++

+
Cisco Public

+++
164

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Designs with Server Virtualization Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server
10 Gigabit Ethernet Performance Considerations 10 Gigabit Performance in Virtualized Environments Datacenter Ethernet

Break Demo: Nexus1kv


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

165

How Much Traffic Can a Single VM Generate? (TX, aka Virtual-to-Physical)


A single VM can drive alone more than 1Gbps worth of bandwidth (in the tested configuration a single VM can drive up to 3.8 Gbps of traffic) Even if the Guest OS displays Network Adapter of 1Gbps, the performance is not gated at 1 Gbps!

ESX 3.5 U2 CPU 2 x dual core Xeon5140 Guest OS Windows 2003 R2 SP2 Memory 8 GB

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

166

Traffic Flow VM-to-Physical (V2P) With Quad-GigE Cards

Catalyst C t l t 6500

GigE 2/13 - 16

vNIC vNIC vNIC vNIC


TECDCT-3873_c2 TECDCT-3873_c2

vmnic0 vmnic1 vmnic2 vmnic3

1 GigE 1 GigE 1 GigE 1 GigE Te4/3

10 GigE
client 1

Traffic Flow VM-to-Physical (V2P) With 10 GigE Cards

ESX 1 ESX 1

Te4/4 GigE 2/17 - 20 10 GigE

client 2

1 GigE

1 GigE 1 GigE 1 GigE


vmnic1 vmnic2 vmnic3

vmnic0

vNIC

vNIC

vNIC

vNIC

ESX 2
2009 Cisco Systems, Inc. All rights reserved. Cisco Public

167

Catalyst C t l t 6500

vNIC vNIC vNIC vNIC

2009 Cisco Systems, Inc. All rights reserved.

vmnic0 vmnic1 vmnic2 vmnic3

10 GigE Te4/3

10 GigE
client 1

Te4/4 GigE 2/17 - 20 1 GigE 1 GigE 1 GigE 1 GigE


vmnic1 vmnic2 vmnic3

client 2

10 GigE

vmnic0

vNIC

vNIC

vNIC

vNIC

ESX 2
Cisco Public

168

How Much Traffic Can 4 VMs Generate? (TX aka V2P)


A typical configuration made of 4 VMs could drive up to ~8-9 Gbps worth of traffic, which means that an ESX server equipped with a Quad-GigE adapter throttles the VMs performance of a typical ESX implementation

ESX 3.5 U2 CPU 2 x dual core Xeon5140 Guest OS Windows 2003 R2 SP2 Memory 8 GB

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

169

P2V (RX) vs V2P (RX) Throughput With 10 GigE NICs to 4 VMs


RX: ~4.3 Gbps TX: ~ 8Gbps

ESX 3.5 U2 CPU 2 x dual core Xeon5140 Guest OS Windows 2003 R2 SP2 Memory 8 GB

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

170

How to Improve VMWARE Performance in RX?


VMWARE Solution: Netqueue What is Netqueue? Netqueue is the equivalent of Receive Side Scaling in VMWARE, i.e. it helps distributing incoming traffic to the available cores.

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

171

P2V With Netqueue Disabled


Maximum Throughput is ~3.9 Gbps CPU goes all the way to 100%

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

172

P2V With Netqueue Enabled


Maximum Throughput is ~4.2Gbps All cores are below 100%

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

173

Tips for Tuning VMWARE With 10 GigE (Courtesy of Intel)


Set CPU affinity for virtual machines:
In the vCenter (VC) console select a Virtual Machine (VM), right click and select Edit Settings. In the VM Properties dialog box select the Resources tab. Click on the Advanced g p g CPU object and in the right pane of the window click on the Run on processor(s) radio button. Select a processor core for the VM to run on and click OK to close the window. Repeat for all VMs.

Turn on NetQueue support in ESX


On the vCenter management console select the host to be configured and click the configuration tab. In the Software box select and open Advanced Settings. Find the parameter labeled VMkernel.Boot.netNetqueue and check the box to enable it. Reboot the system.

Load the driver with multiple queue support:


After the driver rpm has been installed and the machine has rebooted, the driver will have initialized in its default single queue mode. Unload the driver with the command vmkload_mod u ixgbe. Reload the driver and set it in multiple queue mode with the command vmkload_mod ixgbe VMDQ=X,X InterruptType=2,2 (where the comma separated parameter value is repeated for each physical port installed in the machine which uses the ixgbe driver and the value X is the desired number of queues. For a configuration with 8 VMs I use VMDQ=9. This gives 8 dedicated Rx queues to assign to the VMs plus the default TxRx queue.
174

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

LAN Switching
Evolution of Data Center Architectures New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling Designs with Server Virtualization Break Demo: vPC Designs with Server Virtualization 10 Gigabit Ethernet to the Server
10 Gigabit Ethernet Performance Considerations 10 Gigabit Performance in Virtualized Environments Datacenter Ethernet

Break Demo: Nexus1kv


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

175

I/O Consolidation
I/O consolidation supports all three types of traffic onto a single network Servers have a common interface adapter that supports all three types of traffic

IPC: Inter Process Communication


TECDCT-3873_c2 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

176

Data Center Ethernet Summary


Feature
Priority-based Flow Control (PFC) CoS Based BW Management Congestion Notification (BCN/QCN) Data Center Bridging g g Exchange L2 Multi-path for Unicast & Multicast

Benefit
Provides class of service flow control Ability to support control. storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

Auto negotiation Auto-negotiation for Enhanced Ethernet capabilities DCBX (Switch to NIC) Eliminate Spanning Tree for L2 topologies Utilize full Bi-Sectional bandwidth with ECMP

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

177

SAN Switching

Presentation_ID

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

178

Complete Your Online Session Evaluation


Give us your feedback and you could win fabulous prizes. Winners announced daily. Receive 20 Passport points for each session evaluation you complete. Complete your session evaluation online now (open a browser through our wireless network to access our portal) or visit one of the Internet stations throughout the Convention Center.

Dont forget to activate your Cisco Live Virtual account for access to all session material, communities, and on-demand and live activities throughout the year. Activate your account at the Cisco booth in the World of Solutions or visit www.ciscolive.com.
179

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Recommended Readings

www.datacenteruniversity.com

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

180

Recommended Readings

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

181

TECDCT-3873_c2

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

182

Data Center Power Session


TECDCT-3873

Presentation_ID

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Agenda
Infrastructure Design (Mauricio Arregoces) LAN Switching Analysis (Maurizio Portolani)
Recap on Current Trends and Past Best Practices New Layer 2 Technologies Fabric Extender Deep dive and Design with virtual Port Channeling

Break Demos: vPC, OTV (Maurizio Portolani)


Designs with Server Virtualization 10 Gigabit Ethernet to the Server

Break Demo: Nexus1kv (Maurizio Portolani) Blade Servers (Carlos Pereira)


Blade Switching LAN Blade Switching SAN Storage Networking with VMware ESX / vSphere

Break
Unified IO Unified Compute System

Demo: UCS (Carlos Pereira)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Blade Switching - LAN

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

What Are Going to Talk About ?


Cisco Catalyst Virtual Blade Switches (VBS)

Cisco Bl d S it h Ci Blade Switch Entry Level GE switch GE VBS

Cisco Part Number CBS30x0

OEM

CBS31x0X

10G VBS

CBS31x0G

x = 1 for IBM, 2 for HP and 3 for Dell


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Setting the Stage


On this session of the Data Center techtorial the maximum number of enclosures per rack will be considered for the SAN design calculations.

Nevertheless, power and cooling constraints needs to be considered on a case by case basis when implementing blade servers.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Design with Pass-Thru Module and Modular Access Switch


Cable density Rack example:
Four Enclosures Per Rack Up to 16 servers per enclosure 32 1GE LOMs + 2 Management interfaces per enclosure. 136 available 1GE access ports Requires structured cabling to support 136 1GE connections/rack

Modular Access Switches

Blade Server Rack

Pair of Cat 6513 Nexus 7010

Supports up to pp p 28 enclosures (7 racks) 10 x 6748 cards per each switch


19 enclosures (5 racks) 7 x 48 1GE cards + 1 x 10GE card per each switch 42 enclosures (11 racks) 15 x 48 1GE cards + 1 x 10GE card per each switch
Cisco Public

Nexus 7018
TECDCT-3873

Gigabit Ethernet Connections


6

2009 Cisco Systems, Inc. All rights reserved.

Design with Pass-Thru Module and Modular Access Switch

Does this look Manageable?

How to you find and replace a bad cable?

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Design with Pass-Thru Module and Top of the Rack (TOR) Switches
High Cable density within the rack High capacity uplinks p provide g p y p aggregation layer connectivity Rack example:
Up to Four blade enclosures/rack Up to 128 cables for server traffic Up to 8 cables for Server management p pp Up to four rack switches support local blade servers Additional switch for server management Requires up to 136 cables within the rack 10 GigE Uplinks Aggregation Layer

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Design with Blade Switches


Reduces cables within the rack High capacity uplinks p provide g p y p aggregation layer connectivity Rack example:
Up to four enclosures per rack Two switches per enclosure Either 8 GE or one 10GE uplink per switch Between 8 and 64 cables/fibers per rack Reduces number of cables within the rack but increases the number of uplinks compared to ToR solution Based on cable cost 10GE from Blade Switch is a better option.
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Aggregation Layer

10 GigE Or GE Uplinks

Design with Virtual Blade Switches (VBS)


Removes Cables from Rack High capacity uplinks p provide g p y p aggregation layer connectivity Rack example:
Up to Four blade enclosures/rack Up to 64 Servers per rack Two switches per enclosure p One/Two Virtual Blade Switch per rack Two or Four 10GE uplinks per Rack Reduces number of Access Layer switches by factor of 8 Allows for local Rack traffic to stay within the Rack
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Aggregation Layer

10 GigE Or GE Uplinks

10

Cisco Virtual Blade Switch (VBS)


Physical Connections

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

11

Cisco Virtual Blade Switch (VBS)


Logical View

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

12

Multiple Deployment Options for Customers


Caters to Different Customer Needs
Benefits Common Scenario Single Virtual Blade switch per rack Entire rack can be deployed with as little as two 10 GE uplinks or two GE Etherchannels Allows for Active/Active NIC teams Creates a single router for entire rack if deploying L3 on the edge d l i h d Keeps Rack traffic in the Rack Design Considerations Ring is limited to 64 Gbps May cause Oversubscription
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

13

Multiple Deployment Options for Customers


Caters to Different Customer Needs
Benefits Sepa ate Separate VBS d de Left/Right s tc es S divide e t/ g t switches More resilient Provides more Ring capacity since two rings per Rack Design Considerations Requires more Uplinks per Rack Servers can not form A/A NIC teams

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

14

Multiple Deployment Options for Customers


Caters to Different Customer Needs
Benefits Allows for 4 NICs per server p Can Active/Active Team all 4 NICs More Server Bandwidth Useful for highly virtualized environments Design Considerations Creates smaller Rings Requires more Uplinks May Increase Traffic on each Ring

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

15

Additional Options
By combining above three scenarios, the user can:
Deploy up to 8 switches per enclosure Build smaller Rings with fewer Switches Split VBS between LAN on Motherboard (LOM) and Daughter Card Ethernet NICs Split VBS across racks Connect unused uplinks to other Devices such as additional Rack Servers or Appliances such as storage

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

16

Plug-and-Play Designs with VBS and Nexus1000v


1 Add or replace a VBS Switch to the Cluster 2 Switch config and code automatically propagated Virtual Ethernet Module

3 Add a blade Server 4 Its always booted from the same LUN

Virtual Ethernet Module

Virtual Ethernet Module

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

17

Cisco Virtual Blade Switch (VBS)


Scalability

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

18

Proper VBS Ring Configuration


Each offer a full ring, could be built with 1 meter cables, and looks similar But: Certain designs could lead to a split ring if an entire enclosure is powered down
For No example, in the 4 enclosure example, if enclosure 3 had p power removed y would end up with two rings, one made up of you p g , p the switches in enclosures 1 and 2, and one made up of the switches in enclosure 4. This, at a minimum would leave each VBS contending for the same IP address, and remote switch No Yes management would become difficult The Yes examples also have a better chance of maintaining connectivity for the servers in the event a ring does get completely split due to multiple faults
Cable Lengths are 0.5, 1.0 and 3.0 Meter. The 1.0 Meter cable ships standard
No Yes

ENC 4

ENC 3 No Yes

ENC 2

ENC 1

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

19

Virtual Blade Switch Across Racks


VBS cables are limited to max of 3 meters Insure that switches are not isolated in case of failure of switch or enclosure May require cutting holes through side walls of Cabinets/Racks

~2 FT

~2 FT
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

20

Deployment Scenario without vPC / VSS


Straight forward configuration
Ensure uplinks are spread across switches and enclosures If using EtherChannel (EC), make sure members are not in same enclosure By using RSTP and EC, recovery time on failure is minimized Make Master Switch (and Alternate) are not Uplink switches Use FlexLinks if STP is not desired

Aggregation Layer

Core Layer

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

21

Deployment Scenario without vPC / VSS


Access Layer (Virtual Blade Switch) Aggregation Layer

Single Switch / Node (for Spanning Tree or Layer 3 or Management)


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Spanning-Tree Blocking

22

Deployment Scenario without vPC / VSS


Aggregation Layer Access Layer (Virtual Blade Switch)

Single Switch / Node (for Spanning Tree or Layer 3 or Management)


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Spanning-Tree Blocking

23

Deployment Example
Switch Numbering 1 to 8, left to Right, Top to Bottom Master Switch is Member 1 Alternate Masters will be 3,5,7 Uplink Switches will be Members 2,4,6,8
1 2

10 GE ECs from 2,4 and 6,8 will be used


3 4

RSTP will be used User Data VLANs will be interleaved


7 5 6

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

24

Configuration Commands
switch 1 priority 15 switch 3 priority 14 switch 5 priority 13 switch 7 priority 12 spanning-tree mode rapid-pvst vlan 1-10 state active g g g g g interface range gig1/0/1 gig1/0/16 switchport access vlan xx Assign ports to VLANs Sets Sw 1 to pri master Sets Sw 3 to sec master Sets Sw 5 to 3rd master Sets Sw 7 to 4th Master Enables Rapid STP Configures VLANs

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

25

Configuration Commands
interface range ten2/0/1, ten4/0/1 switchport mode trunk switchport trunk allowed vlans 1 10 1-10 channel group 1 mode active interface range ten6/0/1, ten8/0/1 switchport mode trunk switchport trunk allowed vlans 1-10 channel group 2 mode active interface po1 spanning-tree vlan 1 3 5 7 9 port-priority 0 i t l 1,3,5,7,9 t i it spanning-tree vlan 2,4,6,8,10 port-priority 16 interface po2 spanning-tree vlan 1,3,5,7,9 port-priority 16 spanning-tree vlan 2,4,6,8,10 port-priority 0

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

26

Deployment Scenario with vPC / VSS


Access Layer (Virtual Blade Switch) Aggregation Layer Nexus vPC, Cat6k VSS

Single Switch / Node (for Spanning Tree or Layer 3 or Management)


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

All Links Forwarding

27

Deployment Scenario with vPC / VSS


Aggregation Layer (Nexus vPC) Access Layer (Virtual Blade Switch)

Single Switch / Node (for Spanning Tree or Layer 3 or Management) All Links Forwarding
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

28

Deployment Scenario with vPC / VSS


Physical View

VBS 1

VBS 2

VBS 3

VBS 4

VSS vPC

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

29

Deployment Scenario with vPC / VSS


Logical View

VBS 1

VBS 2

VSS vPC

VBS 3
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

VBS 4
30

Rules to Live by for EC / MCEC


1. Split links across line cards on Catalyst 6500 / Nexus 7000 side prevents against Line Card Outage 2. Split across pair of Catalyst 6500 or across pair of Nexus 7000 prevents against aggregation switch outage 3. Split links across members on blade side if using VBS - prevents against blade switch outage 4. Split links across Blade Enclosures if possible prevents against enclosure outage 5. Split VLANs 5 S lit VLAN across f l d b l for load balancing prevents idle EC i t idl ECs. 6. Chose appropriate EC load balancing algorithm example: Blade servers generally have even number MAC addresses. Consider the hashing algorithms enhancements with MCEC 7. Last but Not least, monitor your ECs - Only way to know if you need more BW or Better MCEC load balance
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

31

Further Points to Consider on Layer 2:


When is Layer 2 Adjacency Required?
Clustering: applications often execute on multiple servers clustered to appear as a single device. Common for HA, Load Balancing and High Performance computing requirements. MS-Windows Advanced Server Clustering Linux Beowulf or proprietary clustering (HPC)

NIC teaming software typically req ires la er t picall requires layer 2 adjacency

AFT SFT ALB


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

32

Blade NIC Teaming Configurations


Network Fault Tolerance (NFT)
Typical referred to as Active/Standby Used when server sees two or more upstream switches NIC connectivity is PREDEFINED with built-in switches and may limit NIC configuration options

Transmit Load Balancing (TLB)


Primary adapter transmit and receives Secondary adapters transmit only Rarely used

Switch Assisted Load Balancing (SLB)


Often referred to as Active/Active Server must see same switch on all member NICs GEC/802.3ad Increased throughput Available with VBS switches

Active Standby
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

33

Blade Server Access Topologies


Different Uplinks Possibilities
V-Topology U-Topology Trunk-Failover Topology

Very Popular Topology Some Bandwidth not available


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved.

Not as Popular

Maximum Bandwidth available Needs NIC Teaming

Cisco Public

34

Layer 2 Trunk Failover


Typical Blade Network Topologies
L3 Switches

Link State Group 1

Cisco Blade Switches

Link State Group 1

Blade Server Chassis


FEATURE Map Uplink EtherChannel to downlink ports (Link State Group) If all uplinks fail, instantly shutdown downlink ports Server gets notified and starts using backup NIC/switch
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Blade Server Chassis


CUSTOMER BENEFIT Higher Resiliency / Availability Reduce STP Complexity
35

Flexlink Overview
Achieve Layer 2 resiliency without using STP Access switches have backup links to Aggregation switches p gg g Target of sub-100msec convergence upon forwarding link failover Convergence time independent of #vlans and #mac-addresses Interrupt based link-detection for Flexlink ports. Link-Down detected at a 24msec poll. No STP instance for Flexlink ports. Forwarding on all vlans on the <up> flexlink port occurs with a single update operation low cost.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

36

(Mac address Move Notification) MMN Overview


Achieve near sub-100 msec downtime for the downstream traffic too, upon flexlink switchover. Lightweight protocol : Send a MMN packet to [(Vlan1, Mac1, Mac2..) (Vlan2, Mac1, Mac2..) ..] distribution network. Receiver parses the MMN packet and learns or moves the contained mac-addresses. Alternatively, it can flush the mac-address table for the vlans vlans. Receiver forwards packet to other switches.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

37

Flexlink Preemption
Flexlink enhanced to :
provide flexibility in choosing FWD link, optimizing available bandwidth utilization

User can configure Fl li k pair when previous FWD li k comes b k up : U fi Flexlink i h i link back
Current FWD link continues Preemption mode Off Previous FWD link preempts the current and begins FWD instead Preemption mode Forced Higher bandwidth interface preempts the other and goes FWD Preemption mode Bandwidth

Note: By default, flexlink preemption mode is OFF default When configuring preemption delay:
user can specify a preemption delay time (0 to 300 sec) default preemption delay is 35 secs

Preemption Delay Time :


Once the switch identifies a Flexlink preemption case, it waits an amount of <preemption delay> seconds before preempting the currently FWD Flexlink interface.
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

38

Flexlink Configuration Commands


CBS3120-VBS-TOP#config t Enter configuration commands, one per line. CBS3120-VBS-TOP(config)#int po1 CBS3120-VBS-TOP(config-if)#switchport backup int po 2 CBS3120-VBS-TOP(config-if)# CBS3120-VBS-TOP#show interface switchport backup detail End with CNTL/Z.

Switch Backup Interface Pairs:

Active Interface

Backup Interface

State

-----------------------------------------------------------------------Port-channel1 Preemption Mode Port-channel2 : off p/ p Active Up/Backup Down

Bandwidth : 20000000 Kbit (Po1), 10000000 Kbit (Po2) Mac Address Move Update Vlan : auto

CBS3120-VBS-TOP#

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

39

Management Screenshot
Topology View

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

40

Management Screenshot
Front Panel View

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

41

Blade Switching - SAN

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

42

What are Going to Talk About ?


Cisco MDS 4Gb Fibre Channel Blade Switches

16 internal copper 1/2/4-Gbps Fibre Channel connecting to blade servers through blade chassis backplane Up to 8 SFP uplinks Offered in 12-port and 24-port configurations via port licensing

14 internal copper 1/2/4-Gbps Fibre Channel connecting to blade servers through blade chassis backplane Up to 6 SFP uplinks Offered in 10-port and 20-port configurations via port licensing

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

43

Virtual Storage Area Network Deployment


Consolidation of SAN islands
Increased utilization of fabric ports with JustIn-Time provisioning p g
Department A

SAN Islands

Deployment of large fabrics


Dividing a large fabric in smaller VSANs Disruptive events isolated per VSAN RBAC for administrative tasks Zoning is independent per VSAN
Department B Department C

Advanced traffic management


Defining the paths for each VSAN VSANs may share the same EISL Cost effective on WAN links
Department A Department B Department C

Virtual SANs (VSANs)

Resilient SAN Extension Standard solution (ANSI T11 FC-FS-2 section 10)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

44

Understanding VSANs (or Virtual Fabrics)


Production SAN
FC FC FC
FC FC

Tape SAN
FC
FC

Test SAN
FC
FC

FC

SAN A DomainID=1 DomainID=7

SAN B DomainID=2 DomainID=8

SAN C DomainID=3

SAN D DomainID=4

SAN E DomainID=5

SAN F Domain ID=6

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

45

VSAN Technology
The Virtual SANs Feature Consists of Two Primary Functions:
Hardware-based isolation of tagged traffic belonging to different VSANs Create independent instance of Fibre Channel services for each newly created VSAN VSAN services include:
Fibre Channel Services for Blue VSAN Fibre Channel Services for Red VSAN

VSAN Header Is Removed at Egress Point Cisco MDS 9000 Family with VSAN Service Enhanced ISL (EISL) Trunk Carries Tagged Traffic from Multiple VSANs VSAN Header Is Added at Ingress Point Indicating Membership No Special Support Required by End Nodes

Trunking E_Port (TE_Port)

Trunking E_Port (TE_Port)


Fibre Channel Services for Blue VSAN Fibre Channel Services for Red VSAN

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

46

Enhanced vs. Basic Zoning


Basic Zoning Enhanced Zoning Enhanced Advantages

Administrators can make simultaneous configuration changes

All configuration changes fi ti h are made within a single session. Switch locks entire fabric to implement change

One configuration session for entire fabric to ensure consistency within fabric

If a zone is a member References to the zone are Reduced payload size as the zone is referenced. of multiple zonesets , used by the zonesets as The size is more required once you define the an instance is pronounced with bigger zone. created per zoneset. database Default zone policy is defined per switch.
Enforces and exchanges default zone setting throughout the fabric Fabric-wide policy enforcement reduces troubleshooting time.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

47

Enhanced vs. Basic Zoning


Basic Zoning Enhanced Zoning Enhanced Advantages

Managing switch provides combined status about activation. Will not identify a failure switch. To distribute zoneset must re-activate the same zoneset. During D i a merge MDS specific types can be misunderstood by noncisco switches.

Retrieves th activation R ti the ti ti results and the nature of the problem from each remote switch.

Enhanced error E h d reporting reduces troubleshooting process.

Implements changes to the This avoids hardware zoning database and changes for hard distributes it without zoning in the switches. activation. Provides a vendor ID along with a vendor-specific type value to uniquely identify a member type Unique Vendor type

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

48

Inter VSAN Routing


Similar to L3 interconnection between VLAN Allows sharing of centralized storage services such as tape libraries and disks across VSANswithout merging separate fabrics (VSANs) Network address translation allow interconnection of VSANs without a predefined itho t addressing schema
Engineering VSAN_1 VSAN-Specific Disk

IVR

IVR Tape VSAN_4 (Access via IVR) HR VSAN_3

Marketing VSAN_2

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

49

Quick Review 1
VSANs enable creation of multiple virtual fabrics on top of a consolidated physical SAN infrastructure; Enhanced Zoning recommended and helpful from both scalability and troubleshooting standpoints; Inter VSAN Routing (IVR) required when selective communication between shared devices on distinct fabrics is needed.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

50

N-Port ID Virtualization (NPIV)


Mechanism to assign multiple N_Port_IDs to a single N_Port Allows all the Access control, Zoning, Port Security ( g y (PSM) be ) implemented on application level So far, multiple N_Port_IDs are allocated in the same VSAN

Application Server

FC Switch

Email

Email I/O N_Port_ID 1 Web I/O N_Port_ID 2 File Services I/O N_Port_ID 3 F_Port

Web

File Services

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

51

NPIV Configuration Example


NPIV Is Enabled Switchwide with the C th Command: d
npiv enable Notice that a F-port supports multiple logins

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

52

NPIV Usage Examples


Virtual Machine Aggregation Intelligent Pass-thru

FC

FC

FC

FC

FC

FC

FC

FC

NPV Edge Switch


FC

NP_Port

NPIV enabled HBA F_Port F_Port

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

53

N-Port Virtualizer (NPV)


Enabling Large-Scale Blade Server Deployments
Deployment Model - FC Switch Mode
Blade System
Blade N Blade 2 Blade 1 E-Port
TECDCT-3873

Deployment Model HBA Mode


Blade System Blade System
Blade N Blade 2 Blade 1

Blade System

Blade Switch configured as NPV (i.e. HBA mode)

Blade N

Blade N

Blade 2

Blade 2

Blade 1

Blade 1

FC Switch

FC Switch

NPV

NPV

NPV enables large scale Blade Server deployments by:


E-Port

N-Port

SAN

SAN

- Reducing Domain ID usage - Addressing switch interop issues - Simplifying management

F-Port

Storage

Storage

Blade Switch Attribute FC Switch Mode (E-Port) One per FC Blade Switch Yes Medium Deployment Model # of Domain IDs Used Interoperability issues with multi-vendor Core SAN switch Level of management coordination between Server and SAN Administrators
Cisco Public

HBA Mode (N-Port) None (uses Domain ID of core switch) No Low

NPV is also available on the MDS 9124 & 9134 Fabric Switches
2009 Cisco Systems, Inc. All rights reserved.

54

N-Port Virtualizater (NPV): An Overview


NPV-Core Switch (MDS or 3rd party switch with NPIV support)
FC FC

Solves the domain-id explosion problem l i bl

10.1.1 10 1 1

20.2.1 20 2 1

F-port

NP-port
MDS 9124 MDS 9134

Can have multiple uplinks, on different VSANs (port channel and trunking in a later release)

Up to 100 NPV switches

Cisco MDS in a Blade Chassis


Blade Server 1 Blade Server 2 Blade Server n

server port (F)

10.5.2
FC

10.5.7 20.5.1
Initiator (no FL ports)

NPV Device
Uses the same domain(s) as the NPV-core switch(es)

Target

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

55

NPV FLOGI/FDISC Login Process


NPV Core Switch

When NP port comes up on a NPV edge switch, it first FLOGI and PLOGI into the core to register into the FC Name Server End Devices connected on NPV edge switch does FLOGI but NPV switch converts FLOGI to FDISC command, creating a virtual PWWN for the end device and allowing to login using the physical NP port. All I/O of end device will always flow through same NP port
F

NP P1 NP P2

NPV Edge Switch


F

P4 = vP2

P5 = vP3

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

56

FlexAttach
Because Even Physical Devices Move How it works ?
Based on WWN NAT of Servers Server s WWN
Bl lade 1

Blade Server

.
NPV

Re eplaced Blade B

Bl lade N

Key Benefit:
Flexibility for Server Mobility - Adds, Moves and Changes Eliminates need for SAN and server team to coordinate changes Two modes: Lock identity to port Identity follows physical PWWN

No Blade Switch Config Change

Flex Attach

No Switch Zoning Change g

SAN

No Array Configuration Change

Storage

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

57

Flex Attach
Example
Creation of virtual PWWN (vPWWN) on NPV switch F-port Zone vPWWN to storage LUN masking is done on vPWWN Can swap Server or replace physical HBA
No need for zoning modification No LUN masking change required

Automatic link to new PWWN


no manual re-linking to new PWWN is needed

Before: switch 1

After: switch 2

1
FC1/1 vPWWN1 FC1/6 vPWWN1

PWWN 1

Server 1
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Server 1 58

PWWN 2

Whats Coming:
Enhanced Blade Switch Resiliency
F-Port Port Channel
F-Port Port Channel Blade N Blade 2 Blade 1

Core Director Storage g

Blade System

SAN

N-Port

F-Port

F-Port Trunking
Core Director Storage Blade System
Blade N VSAN 1

F-Port Trunking

Blade 2 Blade 1

VSAN 2

SAN

VSAN 3 N-Port F-Port

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

59

Whats Coming:
F-Port Trunking for the End-Host / Storage
Hardware-based isolation of tagged traffic belonging to different VSANs up to Servers or Storage Devices
Non VSANTrunking capable end node
Fibre Channel Services for Blue VSAN Fibre Channel Services for Red VSAN

VSAN-trunking-enabled drivers required for end nodes (for example, Hosts)

VSAN Header removed at egress point

Trunking E_Port

Implementation example: traffic tagged in Host depending on the VM


VSAN-trunking support required by end nodes VSAN Hdader added by the HBA driver indicating Virtual Machine membership
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Enhanced ISL (EISL) Trunk carries tagged traffic from multiple VSANs Trunking E_Port
Fibre Channel Services for Blue VSAN Fibre Channel Services for Red VSAN

Trunking F_Port
60

Quick Review 2
NPIV standard mechanism enabling F-port (switches and HBAs) virtualization NPV allows a FC switch to work on HBA mode. The switch behaves like a proxy of WWN and doesnt consume a Domain ID, enhancing SAN scalability (mainly on blade environments) Flex-Attach adds flexibility to server mobility allowing the server FC identity to follow the physical pWWN (for blades and rack mount servers) F-port port-channel on NPV scenarios, the ability to bundle p p , y multiple physical ports in to 1 logical link F-port trunking extend VSAN tagging to the N_Port to F_Port connection. Works between switches together with NPV. For host, needs VSAN support on the HBA and allows per-VM VSAN allocation.
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

61

SAN Design: Initial Considerations


Requirements
-F Fan-out maintenance t i t - Dual physical fabrics

SAN Design

Factors to Consider
- Topologies - Bandwidth reservation - Networking / gear capacity
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Parameters
- Number of end devices - Speed variation

62

SAN Design: Initial Considerations


Requirements:
1. Fan-out ratio needs to be maintained to have a predictable and scalable SAN. 2. Dual physical fabric (Fabric A, Fabric B) are identical

Parameters:
1. Number of end-devices (servers, storage and tape) 2. Speed: Majority of end device connection speeds will be primarily 1G, 2G or 4G

Factors to consider:
1. Required topology (core-edge, colapsed core-edge, edge-coreedge, etc.) 2. Bandwidth reservation versus Oversubscription 3. Networking capacity needed (VSANs, ISL, fabric logins, zones, NPIV instances, etc.)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

63

SAN Design: Initial Considerations


Requirements:
1. Fan-out ratio needs to be maintained to have a predictable and scalable SAN. 2. Dual physical fabric (Fabric A, Fabric B) are identical

Parameters:
1. Number of end-devices (servers, storage and tape) 2. Speed: Majority of end device connection speeds will be primarily 1G, 2G or 4G

Factors to consider:
1. Required topology (core-edge, colapsed core-edge, edge-coreedge, etc.) 2. Bandwidth reservation versus Oversubscription 3. Networking capacity needed (VSANs, ISL, fabric logins, zones, NPIV instances, etc.)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

64

SAN FAN-OUT Ratio: What is That ?


Fan-out ratio represents the number of hosts that are connected to a single port g p of a storage array
SAN Fan-out needs to be maintained on the whole SAN design SAN Fan-out defines the SAN oversubscription. Its fixed on blades! Oversubscription is introduced at multiple points Switches are rarely the bottleneck in SAN implementations Must consider oversubscription during a network failure event Remember, all traffic flows towards targets main bottlenecks

Disk Oversubscription Disk do not sustain wire-rate I/O with realistic I/O mixtures. A major vendor promotes 12:1 host:disk fan-out.

Tape O Oversubscription Low sustained I/O rates. All technologies currently have max theoretical native transfer rate << wire-speed FC (LTO, SDLT, etc)

ISL Oversubscription Typical oversubscription in two-tier design can approach 8:1, some even higher

8:1 O.S. (common)

Host Oversubscription Most hosts suffer from PCI bus limitations, OS, and application limitations thereby limiting maximum I/O and bandwidth rate

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

65

SAN FAN-OUT How to Calculate ?


Simple math with physical hosts only. Clusters, VMs and LUN/server ratio should be considered too. Three variables not to be exceeded:
Port queue depth: both storage and HBA; IOPS: to avoid port saturation Throughput: port speed versus sustained traffic.

Design by the maximum values leads to over engineered and underutilized SANs. Oversubscription helps to achieve best cost / performance ratio. Rule of thumb: limit the number of hosts per storage port based on the array fan-out. For instance, 10:1 or 12:1.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

66

SAN Design: Initial Considerations


Premises:
1. Fan-out ratio needs to be maintained to have a predictable and scalable SAN. 2. Dual physical fabric (Fabric A, Fabric B) are identical

Parameters:
1. Number of end-devices (servers, storage and tape) 2. Speed: Majority of end device connection speeds will be primarily 1G, 2G or 4G

Factors to consider:
1. Required topology (core-edge, colapsed core-edge, edge-coreedge, etc.) 2. Bandwidth reservation versus Oversubscription 3. Networking capacity needed (VSANs, ISL, fabric logins, zones, NPIV instances, etc.)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

67

Cisco MDS 9000 Line Cards Detailed

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

68

SAN Design: Initial Considerations


Premises:
1. Fan-out ratio needs to be maintained to have a predictable and scalable SAN. 2. Dual physical fabric (Fabric A, Fabric B) are identical

Parameters:
1. Number of end-devices (servers, storage and tape) 2. Speed: Majority of end device connection speeds will be primarily 1G, 2G or 4G

Factors to consider:
1. Required topology (core-edge, colapsed core-edge, edge-coreedge, etc.) 2. Bandwidth reservation versus Oversubscription 3. Networking capacity needed (VSANs, ISL, fabric logins, zones, NPIV instances, etc.)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

69

Core-Edge
Traditional SAN design for growing SANs High density di t Hi h d it directors i in core and, on the edge:
Unified IO (FCoE) switches [1]; Directors [2] , Fabric Switches [3] or Blade switches [ 4 ]
A A B B A B A B A B

Predictable P di bl performance f Scalable growth up to core and ISL capacity

[1]

[2]

[3]

[4]
70

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

SAN Design: Initial Considerations


Premises:
1. Fan-out ratio needs to be maintained to have a predictable and scalable SAN. 2. Dual physical fabric (Fabric A, Fabric B) are identical

Parameters:
1. Number of end-devices (servers, storage and tape) 2. Speed: Majority of end device connection speeds will be primarily 1G, 2G or 4G

Factors to consider:
1. Required topology (core-edge, colapsed core-edge, edge-coreedge, etc.) 2. Bandwidth reservation versus Oversubscription 3. Networking capacity needed (VSANs, ISL, fabric logins, zones, NPIV instances, etc.)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

71

Cisco MDS 9000 Capacity

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

72

Blade Servers Fibre Channel Integration Challenges


Domain ID scalability limits the maximum number of FC switches to 239 devices per VSAN as per the Standard Standard.
Resellers today do not support more than ~40-75 devices EMC: 40 domains HP: 40 domains

Being able to remove and reinsert a new blade without having to change Zoning Configurations VMWare Integration (discussed later on
this Techtorial)

Up to 8 FC switches per rack (4 Blade Servers x 2) 8 bits 8 bits


Area

8 bits
Device
73

Switch Topology Model


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved.

Switch Domain
Cisco Public

IBM BladeCenter H Core-Edge Design:


Fibre Channel with Cisco MDS FC Switch Module
MDS as FC blade switch
(1100+ usable ports per fabric, all VSAN enabled) BladeCenter Design i Bl d C t H D i using 2 x 4G ISL per blade switch. Oversubscription can be reduced for individual blade centers by adding additional ISLs as needed. VSAN supported.
[A] Storage Ports 240 (2G dedicated): or [B] Storage Ports (4G dedicated): 120 Host Ports (4G HBAs): 1152 ISL Oversubscription (ports): 7.5 : 1 Disk Oversubscription (ports): 10 : 1 Core-Edge Design 8.4 : 1 Oversubscription:
9 racks 56 dual attached servers/rack Each Cisco MDS FC blade switch: 2 ISL to core @ 4G 14 host ports @ 4G 7.5:1 oversubscription 504 total servers 1008 HBAs Storage Array 10:1 oversubscription (fan-out)

[A] 120 storage ports @ 2G or [B] 60 storage ports @ 4G 72 ISL to edge @ 4G

NPIV

NPV + Flex Attach

Cisco MDS 9513 as SAN Aggregation Directors

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

74

Blade Server FC Attached Storage:


Fibre Channel with Cisco MDS FC Switch Module HP cClass
MDS as FC blade switch
(1200+ usable ports per fabric, all VSAN enabled)
Storage Array 10:1 oversubscription (fan-out)

Blade S Bl d Server Design using 2 D i i x 4G ISL per blade switch. Oversubscription can be reduced for individual blade centers by adding additional ISLs as needed.
[A] Storage Ports 240 (2G dedicated): or [B] Storage Ports (4G dedicated): 120 Host Ports (4G HBAs): 1152 ISL Oversubscription (ports): 8 : 1 Disk Oversubscription (ports): 10 : 1 Core-Edge Design 9.6 : 1 Oversubscription:

[A] 120 storage ports @ 2G or [B] 60 storage ports @ 4G 72 ISL to edge @ 4G

NPIV

NPV + Flex Attach

Cisco MDS 9513 as SAN Aggregation Directors

9 racks 64 dual attached servers/rack

Each Cisco MDS FC blade switch (02 switches per HP c-Class enclosure): 2 ISL to core @ 4G 16 host ports per HP c-Class enclosure @ 4G 8:1 oversubscription

576 total servers 1152 HBAs

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

75

Storage Networking with VMWare ESX

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

76

Virtual Machines (VM) @


Storage Networking with Blade Servers
Virtual Machines pose new requirements for SANs Switching Performance Support complex, unpredictable, dynamically changing traffic patterns d i ll h i ffi Provide fabric scalability for higher workload Differentiate Quality of Service on a per VM basis Deployment, Management, Security Create flexible and isolated SAN sections, support management Access ti t tA Control Support performance monitoring, trending, and capacity planning up to each VM Allow VM mobility without compromising security
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public
Virtualized Servers Virtualized Servers Virtualized Servers Virtualized Servers

Virtual Machines

Fabric

Storage Array Storage Array

Tier 1

Tier 2

Tier 3

77

VMware ESX Storage Options


FC iSCSI/NFS DAS

VM

VM

VM

VM

VM

VM

FC

FC

SCSI

iSCSI is popular in SMB market DAS is not popular because it prohibits VMotion

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

79

Virtual Servers Share a Physical HBA


A zone includes the physical hba and the storage array
Virtual Serve ers

Access control is demanded to storage array LUN masking and mapping, it is based on the physical HBA pWWN and it is the same for all VMs The hypervisor is in charge of the mapping, errors may be disastrous

Hypervisor

MDS9124e Mapping

Storage Array (SAN A or B) (LUN Mapping and Masking)

FC

HW

pWWN-P

FC

pWWN-P

Zone
TECDCT-3873

Single Login on a Single Point-to-Point Connection


2009 Cisco Systems, Inc. All rights reserved. Cisco Public

FC Name Server
80

Virtual Server Using NPIV and Storage Device Mapping


Virtual HBAs can be zoned individually LUN masking and mapping is based on the virtual HBA pWWN of each VMs Very safe with respect to configuration errors Only supports RDM Available since ESX 3.5
MDS9124e
Mapping Mapping Mapping Mapping

Virtual Servers s

Hypervisor

Storage Array (SAN A or B)

FC

FC

FC

FC

FC

To pWWN-1
pWWN-1 pWWN-2 pWWN-3 pWWN-4

To pWWN-2 pWWN-P pWWN-1 pWWN-2 pWWN-3 pWWN-4 To pWWN-3 To pWWN-4

HW

pWWN-P

FC

Multiple Logins on a Single Point-to-Point Connection


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

FC Name Server
81

QoS for Individual Virtual Machines


Zone-Based QoS: VM-1 has Priority; VM-2 and any Additional Traffic has Lower Priority VM-1 Reports Better Performances than VM-2
Virtual Machines s

VM-1

VM-2

Congested Link Cisco MDS 9124e Multilayer Fabric Switch Cisco MDS 9000 Multilayer Fabric Switch Storage Array (SAN A or B)
Storage Array

FC

H Hypervisor

pWWN-V2 Low Priority

QoS
FC

Q QoS IVR
pWWN-T

pWWN-V1 High Priority

HW

pWWN-P

FC

Low-Priority Traffic

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

82

Routing Virtual Machines Across VSANs Using NPIV and IVR


Targets are in different VSANs Inter VSAN Routing Zoning:
Virtual Machines s

VM-1

VM-2

IVR-Zone-P includes the physical devices pWWN-P and pWWN-T IVR-Zone-Vx includes the virtual machine x and the physical target only p y g y

LUN Mapping and Masking


Each LUN x is exposed to the physical initiator pWWN-P and to virtual machine x pWWN-Vx only
Raw Device Mapping Raw Device Mapping

FC

IVR-Zone-V2 MDS9124e

MDS9000 VSAN-20
FC

ESX Hypervisor X

pWWN-V2

pWWN-T2 WWN T2
FC

IVR-Zone-V1
pWWN-V1

VSAN-1
HW
pWWN-P
FC

IVR VSAN-1 VSAN-10 VSAN-20

VSAN-10

FC

IVR

pWWN-T1

IVR-Zone-P
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

83

VMotion LUN Migration without NPIV


VM1 VM2 VM3 VM1 VM2 VM3 VM1 VM2 VM3

Standard HBAs
WWPN

W S-X9 01 6

1 STAT S U

10

11

12

13

14

15

16

1/2 Gbp s FC Mod ule

All configuration parameters are based on the World Wide Port Name (WWPN) of the physical HBA

FC

All LUNs must be exposed to p every server to ensure disk access during live migration (single zone)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

84

VMotion LUN Migration with NPIV


VM1 VM2 VM3

HBAs with NPIV


WWPN1 WWPN2 WWPN3
W S-X9 01 6 1 STAT S U 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16

1/2 Gbp s FC Mod ule

No need to reconfigure zoning or LUN masking Dynamically reprovision VMs without impact to existing infrastructure
FC

Centralized management of VMs and resources Redeploy VMs and support live migration

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

85

VMotion: Switch Name Server - Before

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

86

VMotion: Switch Name Server - After

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

87

Virtualization Infrastructure and Management


Example: Mapping vCenter Data Centers to VSAN
Data Center Red
Storage Array

Frame Tagged on Trunk Cisco MDS 9124e Blade Data CenterGreen VSAN-10 VSAN-20 VSAN-30 Cisco MDS 9000 Family

VSAN-10 VSAN 10

Storage Array

VSAN-20

Storage Array

VSAN VSAN30

Data Center Yellow Administrator Privileges Admininistrative Team Red Green Yellow
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Virtual Machines Data Center Red Data Center Green Data Center Yellow

Storage Network VSAN-10 VSAN-20 VSAN-30

Storage Array Red Array Green Array Yellow


88

In Summary: Blade Servers w/ Cisco LAN & SAN

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

89

Unified IO (FCoE)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

90

What Is Data Center Ethernet (DCE)?

Data Center Ethernet is an architectural collection of Ethernet extensions designed to improve Ethernet networking and management in the Data Center.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

91

Whats the Difference Between DCE, CEE and DCB ?

Nothing! All 03 acronyms describe the same thing, meaning the architectural collection of Ethernet extensions (based on open standards) Cisco has co-authored many of the standards associated and is focused on providing a standards-based solution for a Unified Fabric in the data center The IEEE has decided to use the term DCB (Data Center Bridging) to DCB describe these extensions to the industry. http://www.ieee802.org/1/pages/dcbridges.html

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

92

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz Congestion Notification (BCN/QCN) - 802.1Qau Data Center Bridging p y g Capability Exchange Protocol - 802.1AB (LLDP) L2 Multi-path for Unicast & Multicast Lossless Service

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

Auto-negotiation for Enhanced Ethernet capabilities DCBX Eliminate Spanning Tree for L2 topologies Utilize full Bi-Sectional bandwidth with ECMP Provides ability to transport various traffic types (e.g. Storage, RDMA)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

93

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb

Benefit
Provides class of service flow control. Ability to support storage traffic

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

94

Data Center Ethernet Features - PFC


Priority-Based Flow Control (PFC)

Enables lossless Fabrics for each class of service PAUSE sent per virtual lane when buffers limit exceeded Network resources are partitioned between VLs (E.g. input buffer and output queue) The switch behavior is negotiable per VL
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

95

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

96

Data Center Ethernet Features - ETS


Enhanced Transmission Selection (ETS)

Enables Intelligent sharing of bandwidth between traffic classes control of bandwidth Being Standardized in IEEE 802.1Qaz Also known as Priority Grouping
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

97

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz Congestion Notification (BCN/QCN) - 802.1Qau

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

98

Data Center Ethernet Features


Congestion Management

Moves congestion out of the core to avoid congestion spreading Allows End-to-End congestion management Standards track in 802.1Qau
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

99

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz Congestion Notification (BCN/QCN) - 802.1Qau Data Center Bridging p y g Capability Exchange Protocol - 802.1AB (LLDP)

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

Auto-negotiation for Enhanced Ethernet capabilities DCBX

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

100

Data Center Ethernet Features - DCBX


Data Center Bridging Capability eXchange Protocol

Data Center Ethernet

Data Center Ethernet

Handshaking Negotiation for: CoS BW Management Class Based Flow Control Congestion Management (BCN/QCN) Application (user_priority usage) Logical Link Down
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

101

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz Congestion Notification (BCN/QCN) - 802.1Qau Data Center Bridging p y g Capability Exchange Protocol - 802.1AB (LLDP) L2 Multi-path for Unicast & Multicast

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

Auto-negotiation for Enhanced Ethernet capabilities DCBX Eliminate Spanning Tree for L2 topologies Utilize full Bi-Sectional bandwidth with ECMP

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

102

Data Center Ethernet Features L2MP


Layer 2 Multi-Pathing
Phase 1
LAN

Phase 2
LAN
Virtual Switch

Phase 3
LAN

MAC A

MAC B Active-Active

L2 ECMP vPC

MAC A

MAC B

L2 ECMP

We are here
Eliminates STP on Uplink Bridge Ports Allows Multiple Active Uplinks Switch to Network Prevents Loops by Pinning a MAC Address to Only One Port Completely Transparent to Next Hop Switch Virtual Switch retains physical switches independent control and data planes Virtual port channel mechanism is transparent to hosts or switches connected to the virtual switch STP as fail-safe mechanism to prevent loops even in the case of control plane failure Uses ISIS based topology Eliminates STP from L2 domain Preferred path selection TRILL is the work in progress standard

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

103

Data Center Ethernet Standards and Features


Overview
Feature
Priority-based Flow Control (PFC) - 802.1Qbb Enhanced Transmission Selection - 802.1Qaz Congestion Notification (BCN/QCN) - 802.1Qau Data Center Bridging p y g Capability Exchange Protocol - 802.1AB (LLDP) L2 Multi-path for Unicast & Multicast Lossless Service

Benefit
Provides class of service flow control. Ability to support storage traffic Grouping classes of traffic into Service Lanes IEEE 802.1Qaz, CoS based Enhanced Transmission End to End Congestion Management for L2 network

Auto-negotiation for Enhanced Ethernet capabilities DCBX Eliminate Spanning Tree for L2 topologies Utilize full Bi-Sectional bandwidth with ECMP Provides ability to transport various traffic types (e.g. Storage, RDMA)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

104

Virtual Links
An Example
Up to 8 VLs per physical link Ability to support QoS queues within the lanes
VL2 - No Drop Service - Storage
DCE
CNA

VL1 LAN Service LAN/IP LAN/IP Gateway

DCE
CNA

VL1 VL2 VL3

DCE
CNA

Campus Core/ Internet


VL3 D l Delayed Drop S i - IPC dD Service

Storage Gateway

Storage Area Network


105

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Fibre Channel over Ethernet How it Works


Direct mapping of Fibre Channel over Ethernet
FC-4 FC-3 FC-2 FC-1 FC-0 FC-4
CRC C
SOF

FC-3 FC-2
FCoE Mapping MAC PHY

FC Frame
Ethernet Payload

EOF

Ethernet Header

Ethernet FCS

(a) Protocol Layers

(b) Frame Encapsulation

Leverages standards-based extensions to Ethernet (DCE) to provide reliable I/O delivery


Priority Flow Control (PFC) Data Center Bridging Capability eXchange Protocol (DCBX)
10GE Lossless Ethernet Link (DCE)
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

FCoE Traffic Other Networking Traffic


106

FCoE Enablers
10Gbps Ethernet Lossless Ethernet
Matches the lossless behavior guaranteed in FC by B2B credits

Ethernet jumbo frames


Max FC frame payload = 2112 bytes
Normal et e et frame, et e type = FCoE o a ethernet a e, ethertype Co Same as a physical FC frame
Ethernet Header FCoE Header FC Header CRC EOF

FC Payload

Control information: version, ordered sets (SOF, EOF)


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

FCS
107

Encapsulation Technologies

Operating System / Applications SCSI Layer FCP iSCSI TCP IP FC


1, 2, 4, 8, 10 Gbps

FCP FCIP TCP IP

FCP iFCP TCP IP

FCP

SRP

FCoE IB
10, 20 Gbps

Ethernet
1, 10 . . . Gbps

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

108

Encapsulation Technologies
FCP layer is untouched
OS / Applications SCSI Layer FCP

Allows same management tools for Fibre Channel Allows same Fibre Channel drivers Allows same Multipathing software

FCoE E. Ethernet
1, 10 . . . Gbps

Simplifies certifications with ith OSMs Evolution rather than Revolution

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

109

Unified I/O (FCoE) Why ?


Fewer CNAs (Converged Network adapters) instead of NICs, HBAs and HCAs Limited number of interfaces for Blade Servers
FC HBA FC HBA NIC NIC NIC NIC HCA
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved.

FC Traffic FC Traffic
CNA

LAN Traffic LAN Traffic Mgmt Traffic Backup Traffic IPC Traffic
Cisco Public

CNA

All traffic goes over 10GE

110

Unified I/O: What Changes on the Network ?


Today: LAN
Management

SAN A
FC HBA FC HBA NIC

SAN B

Core switches

NIC

Access Top of the Rack switches Servers

Ethernet FC
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

111

Unified I/O: Just the Access Layer


Today Unified I/O LAN
Management

Unified I/O
SAN B
Reduction of server adapters Fewer Cables Simplification of access layer & cabling Gateway free implementation - fits in installed base of existing LAN and SAN L2 Multipathing Access Distribution Lower TCO Investment Protection (LANs and SANs) Consistent Operational Model One set of ToR Switches

SAN A

FCoE Switch

FCoE Ethernet FC
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

112

Converged Network Adapters (CNA)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

113

CNA View on Host


10 GE/FCoE

Cisco ASIC

10 GE

FC

PCIe Bus

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

114

CNA View on VMware ESX Fibre Channel

Emulex

Qlogic

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

115

CNA View on VMware ESX 10 GE

Both Emulex and Qlogic are using Intel Oplin 10 Gigabit Ethernet chip
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

116

Disk Management
Storage is zoned to FC initiator f h t i iti t of host.

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

117

Example: CNA Configuration

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

118

Common SAN/LAN Architecture


Administrative Boundaries
LAN SAN A SAN B

Network Admin
Login: Net_admin Password: abc1234

SAN Admin
Login: SAN_admin Password: xyz6789

Ethernet FC
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

119

Common SAN/LAN Architecture


Administrative Boundaries
LAN SAN A SAN B

Network Admin
Login: Net_admin Password: abc1234

SAN Admin
Login: SAN_admin Password: xyz6789

NX5000

CNA CNA

CNA CNA

DataCenter Ethernet with FCoE


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Ethernet

FC
120

Unified IO Deployment - Unified IO


Core
Storage Arrays
Fabric A

SAN Fabric
Fabric B

L3
Aggregation

N7K

N7K MDS9500 MDS9500 MDS9500 MDS9500

L3 L2

C6K N7K N7K

C6K
SAN Edge A SAN Edge B

Access

L2

LAN Access

N5K
A

N5K
VF_Ports
B

N5K
VN_Ports
D

N5K
E

CNA Enet FC FCoE

Unified IO Server Farm Pod Environment


Converged Edge Infrastructure: Unified/IO using ToR at the edge, and CNA at the hosts ToR 10GE Unified/IO Server Environments Leverage Ethernet and Storage Clouds to reach traditional LAN/SAN services
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

121

Unified IO Farm - Phase 1:


vPC @ Aggregation
Core
Storage Arrays
Fabric A

SAN Fabric
Fabric B

L3
Aggregation

N7K
4 4

N7K MDS9500 MDS9500 C6K


SAN Edge A SAN Edge B

MDS9500 MDS9500

L3 L2

C6K

N7K

Access
4
LAN Access

4 4 4 4

L2

N5K
A

N5K
B

N5K
D

N5K
E

CNA Enet FC FCoE

Unified IO Server Farm using vPC at Aggregation LAN cloud


Access Switches remain as single logical instance Storage connectivity is unchanged

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

122

Unified IO Farm - Phase 2:


vPC @ Aggregation and Access
Core
Storage Arrays
Fabric A

SAN Fabric
Fabric B

L3
Aggregation

N7K
4 4

N7K MDS9500 MDS9500 C6K


SAN Edge A SAN Edge B

MDS9500 MDS9500

L3 L2

C6K

N7Ks

Access
4
LAN Access

L2

N5Ks
A B D

N5Ks
E

CNA Enet FC FCoE

Unified IO Server Farm using vPC at Aggregation LAN cloud


Access Switches provide vPC for LAN connectivity Storage connectivity is unchanged (different physical paths for SAN Fabric A and B)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

123

Nexus 5000 on the Aggregation Layer VE Interfaces are NOT Supported so Far

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

124

Cisco Unified Computing System (UCS)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

125

The Unified Computing Journey

Unified Fabric
Wire once infrastructure Low-latency lossless Virtualization aware
TECDCT-3873

Unified Virtual Machines


VN - Link Application Mobility

Unified Computing
Consolidated Fabric & I/O Stateless Vn-tagging Management

Data Center 3.0


Business service focused Resilient Distributed Standardsbased

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

126

Unified Computing Building Blocks


Unified Fabric Introduced with the Cisco Nexus Series
Physical
Wire once infrastructure (Nexus 5000) Fewer switches, adapters, cables

Ethernet

Fibre Channel

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

127

Unified Computing Building Blocks


Unified Fabric Introduced with the Cisco Nexus Series
Physical
Wire once infrastructure (Nexus 5000) Fewer switches, adapters, cables

Virtual

Ethernet

Fibre Channel

Virtual
VN-Link (Nexus 1000v) Manage virtual the same as physical

Physical

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

128

Unified Computing Building Blocks


Unified Fabric Introduced with the Cisco Nexus Series
Physical
Wire once infrastructure (Nexus 5000) Fewer switches, adapters, cables

Virtual

Ethernet

Fibre Channel

Virtual
VN-Link (Nexus 1000v) Manage virtual the same as physical

Scale Physical
Fabric Extender (Nexus 2000) Scale without increasing points of management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

129

Cisco Unified Computing Solution


Mgmt Server

Embed management Unify fabrics Optimize virtualization p Remove unnecessary


switches, adapters, management modules

Mgmt Server

Less than 1/2 the support pp infrastructure for a given workload

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

130 130

Cisco Unified Computing Solution


Mgmt system A single Server that encompasses:

Network: Unified fabric Compute: Industry standard x86 Storage: Access options Virtualization optimized

Unified management model


Dynamic resource provisioning

Efficient Scale
Cisco network scale & services Fewer servers with more memory

Lower cost
Fewer servers, switches, adapters, cables Lower power consumption Fewer points of management
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

131 131

Cisco Unified Computing Solution

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

132 132

Cisco Unified Computing Solution


Single, scalable integrated system Network + compute virtualization Dynamic resource provisioning
Mgmt LAN SAN A SAN B

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

133 133

UCS Building Blocks


UCS Manager Embedded manages entire system UCS Fabric Interconnect 20 Port 10Gb FCoE 40 Port 10Gb FCoE UCS Fabric Extender Remote line card

UCS Blade Server Chassis Flexible bay configurations UCS Blade Server Industry-standard architecture UCS Virtual Adapters Choice of multiple adapters
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

134

Cisco UCS and Nexus Technology


UCS Components
UCS Manager Embedded Manages entire system UCS Fabric Interconnect 20 Port 10Gb FCoE 40 Port 10Gb FCoE UCS Fabric Extender Remote line card UCS Blade Server Chassis Flexible bay configurations UCS Blade Server Industry-standard architecture UCS Virtual Adapters Choice of multiple adapters
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Nexus Products

Nexus 5000 Unified Fabric Nexus 2000 Fabric Extender

VN-Link Nexus 1000V CNAs with FCoE


135

Cisco Unified Computing System (UCS) Physical

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

136

Cisco Unified Computing System (UCS) Physical


SAN LAN MGMT SAN

Top of Rack Interconnect


G A G G G S S G G Fabric Interconnect G Fabric A Interconnect G G G

(40 or 20 10GE ports) + (2 or 1 GEM uplink slots) l t )

Chassis
Blade Enclosure Fabric I R Extender
x8 x8

Up to 8 half width blades or 4 full width blades


I
x8

R
x8

Fabric Extender

Fabric Extender
Host to uplink traffic engineering
M Adapter B P Adapter B P Adapter

Up to 80Gb Flexible bandwidth allocation

X X x86 Computer

X X X X x86 Computer

Adapter 3 options
Cisco Virtualized adapter Compatibility CNAs (Emulex and QLogic) Native FC + Intel Oplin Intel Oplin - (10GE only)

Compute Blade (Half slot)


TECDCT-3873

Compute Blade (Full slot)


Cisco Public

Compute Blade
137

2009 Cisco Systems, Inc. All rights reserved.

Enclosure, Fabric Switch, and Blades (Front)


Redundant, Hot Swap Power Supply Redundant, Hot Swap Fan 1U or 2U Fabric Switch

Half width server blade Up to eight per enclosure Hot Swap SAS drive (Optional)

Full width server blade Up to four per enclosure Mix blade types

6U Enclosure

Ejector Handles
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved.

Redundant, Hot Swap Power Supply


Cisco Public

138

Rear View of Enclosure and Fabric Switch


10GigE Ports Expansion Bay

Redundant Hot Swap Fan Module

Redundant Fabric Extender

Fan Handle

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

139

UCS Adapters Options


Virtualization
Virtual Machine Aware: Virtualization and Consolidation

Compatibility
Existing Driver Stacks

Cost
Proven 10GbE Technology T h l

Converged network adapters (CNA) Ability to mix and match adapter types within a system Automatic discovery of component types
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

140

UCS Adapters: Interface Views


10 GigE Backplane interfaces to IOMs Physical Interfaces vHBAs & vNICs will be bound to these physical interface Intel Oplin will not have HBA component. Could run FCoE software stack

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

141

UCS Adapters: CLI View


Required to scope to correct chassis/blade/adaptor
rtp 6100 B# rtp-6100-B# scope adapter 1/5/1

Note: Only one adaptor on the half slot blade rtp-6100-B# scope adapter 1/5/2 Error: Managed object does not exist
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

142

UCS Adapters: vHBA Detail Identification


Vendor

Provisioned WWN and if bound to Profile

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

143

UCS Adapters: Ethernet vNIC Details

Ethernet stats

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

144

Cisco Unified Computing System (UCS) Logical

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

145

Unified Computing Key Value Propositions:


Drivers for Use Cases
Hardware State Abstraction Service Profiles

Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

146

Server Attributes / Configuration Points 1/3

Server
Identity (UUID) Adapters Number Type: FC, Ethernet Identity Characteristics Firmware Revisions Configuration settings

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

147

Server Attributes / Configuration Points 2/3

Server
Identity (UUID) Adapters Number Type: FC, Ethernet Identity Characteristics Firmware Revisions Configuration settings Uplinks

Network
LAN settings vLAN, QoS, etc SAN settings vSAN Firmware Revisions

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

148

Server Attributes / Configuration Points 3/3

Storage
Optional Disk usage SAN settings LUNs Persistent Binding Firmware Revisions Adapters

Server
Identity (UUID) Uplinks

Network
LAN settings vLAN, QoS, etc SAN settings vSAN Firmware Revisions

Number Type: FC, Ethernet Identity Characteristics Firmware Revisions Configuration settings

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

149

Traditional Server Deployment


Storage Administrator: Configure LUN access C fi
Masking, binding, boot LUN

Server Administrator: Configure management LAN C fi t Upgrade firmware versions


Chassis, BMC, BIOS, adapters

Network Administrator: Configure LAN access


Uplinks, VLANs

Configure switch
Zoning, VSANs, QoS

Configure policies
QoS, ACLs

Configure BIOS settings Configure NIC settings Configure HBA settings Configure boot parameters

Perform tasks for each server Inhibits pay-as-you-grow incremental deployment pay-as-you-grow
Needs admin coordination every time May incur downtime during deployments

Complex server replacement, upgrade, migration process


Most of these tasks need to be performed for replacement server

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

150

UCS Server Profile Opt-in Choices


Fixed Attributes
Processors Memory Capacity Bandwidth Capacity

Definable Attributes
Disks & usage Network Type: FC, Ethernet, etc. Number Identity Characteristics LAN settings vLAN, QoS, etc SAN settings g LUNs vSAN & Persistent Binding Firmware Revisions Configuration settings Identity (BIOS)

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

151

UCS Service Profile

Storage
Optional Disk usage SAN settings LUNs Persistent Binding SAN settings vSAN Firmware Revisions

Server
Identity (UUID) Adapters Number Type: FC, Ethernet y Identity Characteristics Firmware Revisions Configuration settings

Network
Uplinks LAN settings vLAN QoS etc Firmware Revisions

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

152

UCS Service Profiles


Hardware State Abstraction
LAN Connectivity OS & Application SAN Connectivity

MAC Address NIC Firmware NIC Settings

Drive Controller F/W Drive Firmware

UUID BIOS Firmware BIOS Settings Boot Order

BMC Firmware

WWN Address HBA Firmware HBA Settings

State abstracted from hardware

Separate firmware, addresses, and parameter settings from server hardware Separate access port settings from physical ports Physical servers become interchangeable hardware components Easy to move OS & applications across server hardware
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

154

Dont I Get this Already from VMware?


Hypervisors & Hardware State
Virtual Machine Virtual Machine Virtual Machine Virtual Machine Virtual Machine Virtual Machine

Server Virtualization (VMware, Xen, HyperV, etc.) HYPERVISOR

MAC Address NIC Firmware NIC Settings

Drive Controller F/W Drive Firmware

UUID BIOS Firmware BIOS Settings Boot Order

BMC Firmware

WWN Address HBA Firmware HBA Settings

Hardware State Virtualization

Server virtualization & hardware state abstraction are independent of each other Hypervisor (or OS) is unaware of underlying hardware state abstraction

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

155

UCS Service Profiles


End to End Configure of Full UCS HW Stack

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

156

Server Upgrades:
Within a UCS

Server Name: finance-01 UUID: 56 4d cd 3f 59 5b 61 MAC : 08:00:69:02:01:FC WWN: 5080020000075740 Boot Order: SAN, LAN Firmware: xx.yy.zz

Old Server

New Server

Disassociate server profile from old server Associate server profile to new server Old server can be retired or re-purposed

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

157

Server Upgrades:
Across UCS Instances
Old UCS System New UCS System

Server Name: finance-01 UUID: 56 4d cd 3ffinance-01 Server Name: 59 5b 61 MAC :Server4d cd 3ffinance-01 08:00:69:02:01:FC UUID: 56 Name: 59 5b 61 WWN: 5080020000075740 5b 61 MAC : 08:00:69:02:01:FC UUID: 56 4d cd 3f 59 Boot Order: 08:00:69:02:01:FC WWN: 5080020000075740 MAC : SAN, LAN Firmware: xx.yy.zz LAN Boot Order: SAN, WWN: 5080020000075740 Firmware: xx.yy.zz LAN Boot Order: SAN, Firmware: xx.yy.zz

1. Disassociate server profiles from servers in old UCS System


2. 3. Migrate server profiles to new UCS system Associate server profiles to hardware in new UCS system

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

158

Server Upgrades:
Across UCS Instances
Old System New System
Server Name: finance-01 UUID: 56 Name:3f 59 5b 61 Server 4d cd finance-01 Server Name: finance-01 MAC : 08:00:69:02:01:FC 61 UUID: 56 4d cd 3f3f 59 5b 61 UUID: 56 4d cd 59 5b WWN: 5080020000075740 MAC : 08:00:69:02:01:FC MAC : 08:00:69:02:01:FC Boot Order: SAN, LAN WWN: 5080020000075740 WWN: 5080020000075740 Firmware: xx.yy.zz LAN Boot Order: SAN, LAN Boot Order: SAN, Firmware: xx.yy.zz Firmware: xx.yy.zz

1.

Disassociate server profiles from servers in old UCS system

2. Migrate server profiles to new UCS system


3. Associate server profiles to hardware in new UCS system

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

159

Server Upgrades:
Across UCS Instances
Old System New System
Server Name: finance-01 UUID: 56 4d cd 3ffinance-01 Server Name: 59 5b 61 MAC :Server4d cd 3ffinance-01 08:00:69:02:01:FC UUID: 56 Name: 59 5b 61 WWN: 5080020000075740 5b 61 MAC : 08:00:69:02:01:FC UUID: 56 4d cd 3f 59 Boot Order: 08:00:69:02:01:FC WWN: 5080020000075740 MAC : SAN, LAN Firmware: xx.yy.zz LAN Boot Order: SAN, WWN: 5080020000075740 Firmware: xx.yy.zz LAN Boot Order: SAN SAN, Firmware: xx.yy.zz

1. 2. 3.

Disassociate server profiles from servers in old UCS system Migrate server profiles to new UCS system Associate server profiles to hardware in new UCS system
160

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Dynamic Server Provisioning


Profiles for Web Servers
Server Name: web-server-01 UUID: 56 4d cd 3f 59 5b 61 MAC : 08:00:69:02:01:FC WWN: 5080020000075740 Boot Order: SAN, LAN Firmware: xx yy zz xx.yy.zz

Profiles for App Servers


Server Name: app-server-01 UUID: 65 d4 cd f3 59 5b 16 MAC : 08:00:69:02:01:16 WWN: 5080020000076789 Boot Order: SAN, LAN Firmware: xx.yy.zz yy

Apply appropriate profile to provision a specific server type Same hardware can dynamically be deployed as different server types No need to purchase custom configured servers for specific applications

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

161

Server Profiles - Reduce Overall Server CAPEX


Todays Deployment:
Provisioned for peak capacity Spare node per workload
Oracle RAC
Blade Blade Blade Blade Blade

With Server Profiles:


Resources provisioned as needed Same availability with fewer spares
Oracle RAC
Blade Blade Blade

Web Servers
Blade Blade Blade Blade Blade Blade

Web Servers
Blade Blade Blade Blade Blade

VMware
Blade Blade Blade

VMware
Blade Blade Blade Blade Blade

Blade Blade

Burst Capacity
Blade

Total Servers: 18
Normal use Burst Capacity Spare Hot Spare

HA Spare
Blade

Blade

Total Servers: 14

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

162

Unified Computing Key Value Propositions:


Drivers for Use Cases
Hardware State Abstraction Service Profiles

Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

165

Unified Fabric
SAN IPC LAN

Todays Approach
All fabric types have switches in each chassis Repackaged switches Complex to manage Blade-chassis configuration Bl d h i fi ti dependency Costly Small network domain Blade Chassis

Unified Fabric
Fewer switches Fewer adapters 10GE/FCoE Blade Blade Blade All I/O types available in each chassis
10GE & FCoE LAN, SAN, IPC

Blade

Easier to manage Blades can work with any chassis Small network domain

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

166

Backplane and Fabric Extender


High performance backplane
2x 40G total bandwidth per half slot - 8 lanes of 10G (half-slot) (full slot) - 16 lanes of 10G (full-slot) Redundant data and management paths Support auto discover of all component

Compute blade

Backplane

Fabric Extender

Compute blade Compute blade Compute blade Compute blade

Fabric extender
Manage oversubscription 2:1 to 8:1 FCoE from blade to fabric switch Customizable bandwidth
Compute blade Compute blade Compute blade
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

167

UCS: Overall System (Rear)


Uplinks

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

168

Unified Computing Key Value Propositions:


Drivers for Use Cases
Hardware State Abstraction Service Profiles

Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

169

What Is SR-IOV About?


Single Root IO Virtualization (SR-IOV) allows virtualizing the 10 GigE link (via the PCI-Express bus) into multiple virtual links. SR-IOV is a PCI-Sig standard In other words you can create multiple vmnics each with its own bandwidth allocation This could be Nexus 1000v
Server VM1 vnic VM2 vnic Virtual Switch VM3 vnic VM4 vnic Virtual Switch

vmnic

vmnic

This is what SR-IOV enables


pNIC: 10 Gbps

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

170

UCS Adapters Options


Virtualization
VM I/O Virtualization and Consolidation

Compatibility
Existing Driver Stacks

Cost
Free SAN Access for Any Ethernet Equipped Host

10GbE/FCoE

10GbE/FCoE

Eth

QP FC FC C Eth

vNICs
0 1 2 3 127

10GbE

FC

Software FCoE

PCIe x16 PCIe Bus


171

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Cisco UCS Virtualized Adapter


Virtualized adapter designed for both single-OS and VM-based deployments P id mobility, isolation, and management from the network bilit i l ti d tf th t k Provides
Secure Transparent to hosts
10GE/FCoE
MAC 0 MAC 1

Cut-through architecture High Performance


2x 10Gb Low latency High BW IPC support
User Defineable vNICs
Eth
0

FC
1

SCSI
2

FC
3

Eth
127

128 vNICs
Ethernet, FC or SCSI 500K IOPS Initiator and Target mode
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

PCIe x16
172

Enables Passthrough I/O


Guest OS
Device Driver

Guest OS
Device Driver

Guest OS
Device Driver

vNICs appear as independent PCIe devices


Centrally manageable and configurable Hot-pluggable Virtual NICs

Virtualization Layer Host

Device Manager

Different types: Eth, FC, SCSI, IPC

IOMMU

Guest drives device directly


vNIC

vNIC

vNIC

Use Cases:
I/O Appliances High Performance VMs
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

173

Cisco UCS Virtualized Adapter

FC Eth

Eth FC

SCSI Eth

SCSI Eth

Eth

FC

IPC

NIV Adapter

OS Compute Blade

Network Interface Virtualization adapter Vary nature and number of PCIe interfaces
Ethernet, FC, SCSI, IPC

Up to 128 different PCIe devices


Hot-pluggable - only appear when defined PCI-Sig IOV compliant

Part of Server Array fabric


Centrally managed and configured
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

174

User Configuration Example


Global System Class Definitions Class Name
COS Value Drop/No-Drop Strict Priority Bandwidth/Weight

FC
3 No-Drop No 1 (20%)

Gold
1 Drop No 3 (60%)

Ethernet BE
0 Drop No 1 (20%)

FC Traffic

High Priority Ethernet

Best Effort Ethernet

vNIC1
Class Rate Burst FC 4000 300

vNIC2
FC 4000 400

vNIC3
Eth. BE 5000 100 Class Rate Burst

vNIC1
Gold 600 100

vNIC2
Eth. BE 4000 300

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

175

Unified Computing Key Value Propositions:


Drivers for Use Cases
Hardware State Abstraction Service Profiles

Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

178

Blade Overview
Half-width blade Common Attributes 2 x Intel Nehalem-EP processors 2 x SAS hard drives (optional) Blade Service processor Blade and HDD hot plug support Stateless blade design 4x the memory Full-width blade 10Gb CNA and 10GbE adapter options Differences Half-width blade 4x memory 12 x DIMM slots 2x I/O bandwidth
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Full-width blade 48 x DIMM slots


2 x dual port adapters
179

1 x dual port adapter

Full-Height Blade
2 socket Nehalem-EP blade 48 x DDR3 DIMMs 2 x Mezzanine Cards 2 x Hot swap disk drives Up to 384GB per 2 socket blade Transparent to OS and applications

Reduced server costs


Purchase fewer servers for memory bound applications memory-bound

Reduced power and cooling costs Reduced software costs


Most software is licensed on a per-socket basis
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

180

Expanded Memory Blade


Nehalem-EP Processor

Physical View
8GB

Nehalem-EP Processor

Logical View
Slot 23

Slot 23 Slot 22 Slot 21 Slot 20 Slot 19 Slot 18 Slot 17 Slot 16 Slot 15 Slot 14 Slot 13 Slot Sl t 12 Slot 11 Slot 10 Slot 9 Slot 8 Slot 7 Slot 6 Slot 5 Slot 4 Slot 3 Slot 2 Slot 1 Slot 0
Cisco Public

8GB

32GB
Channel 2
(red)

Slot 22 Slot 21 Slot 20 Slot 19

8GB

Channel 2
(red)

8GB

8GB

8GB

8GB

32GB

Slot 18 Slot 17 Slot 16 Slot 15

8GB

8GB

8GB

32GB
Channel 1
(blue)

Slot 14 Slot 13 Slot Sl t 12 Slot 11

8GB

8GB

Channel 1
(blue)
8GB 8GB

8GB

32GB

Slot 10 Slot 9 Slot 8 Slot 7

8GB

8GB

8GB

32GB
Channel 0
(green)

Slot 6 Slot 5 Slot 4 Slot 3

8GB

Channel 0
(green)

8GB

8GB

8GB

8GB

32GB

Slot 2 Slot 1 Slot 0

8GB

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

181

Expanded Memory Architecture


Increases number of DIMMs the system can use
Makes the system think it has high-capacity DIMMs when using larger number of lower-capacity lower capacity DIMMs

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

182

I/O

CPU Memory

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

183

Unified Computing Key Value Propositions:


Drivers for Use Cases
Hardware State Abstraction Service Profiles

Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

184

Unified Management (1/2)


SAN A SAN B LAN

Infrastructure Management Centralize chassis management


Intrinsic system management

Two Failure Domains Separate fabrics Central supervisor, forwarding logic Distributed Fabric Extenders

Single management domain Scalable architecture

10GE/FCoE
Blade Chassis Blade Chassis Blade Chassis Blade Chassis

Traffic isolation Oversubscription

Chassis Management

Chassis Management

Chassis Management

Chassis Management

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

185

Unified Management (2/2)


Single point of device management
View 1 View 2 Adapters, blades, chassis, LAN & SAN connectivity Embedded manager g GUI & CLI Systems Management Software

Custom Portal

GUI

Standard APIs for systems management


XML, SMASH-CLP, WSMAN, IPMI, SNMP SDK for commercial & custom implementations

CLI

XML API

Standard APIs

UCS Manager

Designed for multi-tenancy


RBAC, organizations, pools & policies

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

186

UCS Conceptual Overview

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

187

UCS Resources - Example

Physical
Server Blades Adapters

Logical
UUIDs VLANs IP Address MAC Address VSANs WWNs

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

188

Resource Pools - Example


Blades blade 3 blade 2 blade 1 blade 0 Blade pool

MACs
01:23:45:67:89:0d 01:23:45:67:89:0c 01:23:45:67:89:0b 01:23:45:67:89:0a

MAC pool

WWN pool WWNs


05:00:1B:32:00:00:00:04 05:00:1B:32:00:00:00:03 05:00:1B:32:00:00:00:02

05:00:1B:32:00:00:00:01

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

189

How They Work Together


UCS Server

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

190

Profiles Example
Servers Virtual Machines Ethernet Adapters Fibre Channel Adapters IPMI Profiles

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

191

Out-of-the-Box Protocol Support


SNMP SMASH CLP

IPMI

CIM XML

Remote KVM

UCS CLI and GUI

Serial Over LAN

UCS XML API

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

192

UCS Manager Loaded from 6100 Switch


Point a browser at IP Address of Switch

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

193

UCS Graphical Interface


Top directory map tells you where you are in tree

NAVIGATION PANE
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

CONTENT PANE
194

Navigation Pane Tabs


Equipment | Servers | LAN | SAN | VM | Admin

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

195

Creation Wizards

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

196

Multi-Tenancy Model (Opt-In)


Company Finance

Network Management

HR

Policies

Server Server Server

Server Server Server

Server Server Server

Server Server Server

Policies

Compute Blade

Compute Blade

Compute Blade

Compute Blade

Compute Blade

Blade Chassis

Blade Chassis

Blade Chassis

Blade Chassis

Blade Chassis

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Fabric Extender

Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade

Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade

Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade

Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade

Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade Compute Blade

Facilities
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

Fabric Extender

197

Tenant Portal for Multi-Tenant Deployment


Server Array Manager supports
Multiple hierarchical server organizations Network organization Infrastructure organization
Cisco UCS GUI

Custom Portal

RBAC and object-level security XML API

Cisco UCS GUI


Designed for enterprise deployment Provides a global view

Single tenant custom views


Through custom portals Typically as plugin of an existing data center infrastructure

Server Array

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

198

Unified Compute Integration in the Data Center:


Use Cases
Hardware State Abstraction Service Profiles Unified Fabric - FCOE

Virtualized Adapter

Expanded Memory Server

Unified Management

UCS Integration

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

199

UCS and Nexus in the Data Center


Nexus 7010

Core Layer

Nexus 7010

Distribution Layer

10GE
Access Layer

Nexus 5000
GigE 10GE

GigE 10GE

FEX
10GE Servers

Rack 1 Rack 1 1GE to Servers Rack 1 Row 1 / Domain 1 / POD 1

...

..
Rack 12 200

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

UCS and Nexus in the Data Center


Nexus 7010

Core Layer

Nexus 7010

Distribution Layer

10GE
Access Layer

Nexus 5000
GigE 10GE

GigE 10GE

UCS 6100

FEX
10GE Servers
blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8 blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8 blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

Rack 1 Rack 1 1GE to Servers Rack 1


TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved.

blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

blade1 slot 1 blade2 slot 2

blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8 blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

Row 1 / Domain 1 / POD 1

...

..
Rack 12

blade1 slot 1 blade2 slot 2 blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

blade1 slot 1 blade2 slot 2

blade3 slot 3 blade4 slot 4 blade5 slot 5 blade6 slot 6 blade7 slot 7 blade8 slot 8

Cisco Public

201

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

202

Interested in Data Center?


Discover the Data Center of the Future
Cisco booth: #617 See a simulated data center and discover the benefits including investing to save, energy efficiency and innovation.

Data Center Booth


Come by and see whats happening in the world of Data Center demos; social media activities; bloggers; author signings Demos include: Unified Computing Systems Cisco on Cisco Data Center Interactive Tour Unified Service Delivery for Service Providers Advanced Services
TECDCT-3873 2009 Cisco Systems, Inc. All rights reserved. Cisco Public

203

Interested in Data Center?


Data Center Super Session
Data Center Virtualization Architectures, Road to Cloud Computing (UCS) Wednesday, Ju y 1, 2:30 3:30 p , Hall D ed esday, July , 30 3 30 pm, a Speakers: John McCool and Ed Bugnion

Panel: 10 Gig LOM


Wednesday 08:00 AM Moscone S303

Panel: Next Generation Data Center


Wednesday 04:00 PM Moscone S303

Panel: Mobility in the DC Data


Thursday 08:00 AM Moscone S303
204

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

Please Visit the Cisco Booth in the World of Solutions


See the technology in action
Data Center and Virtualization
DC1 Cisco Unified Computing System p g y DC2 Data Center Switching: Cisco Nexus and Catalyst DC3 Unified Fabric Solutions DC4 Data Center Switching: Cisco Nexus and Catalyst DC5 Data Center 3.0: Accelerate Your Business, Optimize Your Future DC6 Storage Area Networking: MDS DC7 Application Networking Systems: WAAS and ACE

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

205

Recommended Readings

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

206

Complete Your Online Session Evaluation


Give us your feedback and you could win fabulous prizes. Winners announced daily. Receive 20 Passport points for each session evaluation you complete. Complete your session evaluation online now (open a browser through our wireless network to access our portal) or visit one of the Internet stations throughout the Convention Center.

Dont forget to activate your Cisco Live Virtual account for access to all session material, communities, and on-demand and live activities throughout the year. Activate your account at the Cisco booth in the World of Solutions or visit www.ciscolive.com.
207

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

TECDCT-3873

2009 Cisco Systems, Inc. All rights reserved.

Cisco Public

208

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy