FC forwards frames based on 3 byte Fibre Channel ID (FCID) FCID has three fields Domain ID Each switch gets a Domain ID First Byte of the FCID Used to identify the Switch in the Fabric SPT FSFP users Domain ID as SPF Node ID There is a hard limit number of switches per Fabric is 256 Some ID's are reserved so only 239 are useable Qualified limit is ~50 FCID's No vendor will support your large Fabric when it crashes FSPFP is a very simple protocol Area ID Group of ports on a switch have an Area ID Port ID End station connected to switch pets a PORT ID Node Port Virtualization (NPV) NPV fixes the Domain ID problem by removing the need for a switch to participate in Fabric Services I.e. NO FSPF, FCNS, Zoning, etc. Switches running NVP appear to the rest of the fabric as an end host Switch will look like a N_Port Upstream facing link on the NPV switch is call NP_Port Proxy Node Port Node Port ID Virtualization (NPIV) Switch upstream of NPV switch is the NPV Core Switch NPV Core switches run Node Port ID Virtualization (NPIV) Allows multiple FlOGI's and FCID assignments on it's F port facing downstream NPIV is also in virtualization environments E.g. VMware hosts assigning separate pWWN/FCID's to the VMware Guests Brining down the SAN to each VMware vm's NPV/NPIV Configuration feature npv (requires that feature fcoe is loaded already) This will require a write erase and reload On 5500UP, will reallocate all ports as Ethernet and you will have to have to assign type FC after reload which will require another reload (ahhh!) Configure NP Ports on NPV switch switchport mode np Enable NPIV on NPV Core Switch feature npiv Configure previous E ports as F ports on NPV Core Switch switchport mode f Remember now that the NPV switch doesn't participate inFabric Services no floig database no sfpf no fcns NPV/NPIV Topology NPV/NPIV Configuration N5K1 and N5K2 conf t feature fcoe (have to turn on to run NPV) feature npv (!!! THIS WILL REQUIRE A RELOAD AND ERASE THE CURRENT CONFIGURATION!!) MDS1 show interface brief (all ports should be down) conf t int fc1/13, fc1/17 - 18 no shut int fc1/17 - 18 switchport mode fl int fc1/13 switchport rate-mode dedicated switchport mode e int fc1/1 - 2 switchport mode f no shut feature npiv zoneset name VSAN1 vsan1 zone name SERVER3 member pwwn 21:00:00:04:cf:f3:73:65 (JBOD1 DISK0) member pwwn 21:00:00:1b:32:0b:2f:61 (SERVER3) zoneset activate name VSAN1 vsan 1 MDS 2 conf t int fc1/10, fc1/13 - 14 no shut int fc1/13 - 14 switchport mode fl int fc1/10 switchport rate-mode dedicated switchport mode e int fc1/5 switchport mode f no shut feature npiv N5K1 conf t int fc1/31 - 32 no shut switchport mode f (should automatically do this) int fc1/27 - 28 switchport mode np no shut N5K12 conf t int fc1/31 - 32 no shut switchport mode f (should automatically do this) int fc1/29 switchport mode np no shut show npv flogi-table MDS1 show flogi database As you can see from uplinks fc1/1 - 2 it see's both SERVER3 and SERVER2 which where distributed to two separate uplinks (like dynamic pinning) N5K1 show npv external-interface-usaage From this output you see that fc1/31 is pinned to uplink fc1/27 and fc1/32 is pinned to fc1/27. Because each server interface is pinned to a external interface if one of the interfaces fails the server port on that interface will go down then backup and resend a flogi to the network and will then be pinned to the other up interface. Would want to make FC1/27 - 28 part of a SAN port-channel |
Blog >