HPE Community read-only access December 15, 2018
This is a maintenance upgrade. You will be able to read articles and posts, but not post or reply.
Hours:
Dec 15, 4:00 am to 10:00 am UTC
Dec 14, 10:00 pm CST to Dec 15, 4:00 am CST
Dec 14, 8:00 pm PST to Dec 15, 2:00 am PST
BladeSystem Virtual Connect
cancel
Showing results for 
Search instead for 
Did you mean: 

Virtual Connect FlexFabric connectivity to Nexus5K

 
chuckk281
Trusted Contributor

Virtual Connect FlexFabric connectivity to Nexus5K

Avi had a configuration question:

 

***********

 

Hi all,

 

We want to implement the following connectivity between the VC FlexFabric and the Nexus5K.

Port x3 and x4 connected with LACP , each port to different Nexus , SUS for VM’s LAN.

Port x6 , no lacp, SUS for mgmt + vMotion. (there are no free ports on the FlexFbric module)

 

 webpic299.jpg

 

 

Now the network eng. Says that if the vPC peer  between the two nexus fails,   all the vPC connection  from the secondary nexus will be disabled, but the single connection which is not vPC , remain linked (PORT x6) , because of that the smart link will not function and the MGMT+Vmotion will not worked.

 

They suggest to change the connectivity to this one .

 

 webpic300.png

 

 

I know it’s not the best practice from the VC FlexFabric and the VMware side , Also I thinks it’s not recommended to make ODD number aggregation.

 

Any comment how to implement this scenario ?

 

*************

 

Input from Robert:

The first configuration is common throughout VC environments.  You may need to get more detail from the network engineer regarding the specific Nexus failure they are describing.  The failure scenario being described is sometimes referred to as a black hole. 

 

Typically if the vPC fails, then the Nexus port channels will dissolve.  When the Nexus port channels dissolve, the VC X3/X4 port channels will also dissolve.  Each VC would still have a surviving 10G linked-active path to forward traffic, and each would have one port linked-standby.  Both X6 ports should remain in linked-active state and continue to forward traffic, unless the one of the Nexus disables the ports.  This probably does not happen if I understand the failure correctly.  When the vPC fails, there is no longer a peer connection.  With no peer connection, there should be no potential for a loop between the adjacent Nexus.  The Nexus should continue to forward traffic from X6 from both VC modules onto the next layer switching.  I do not believe the Nexus will stop forwarding traffic on non-peer links when the vPC ports all fail.

 

I’m sure someone will correct me if my brief assessment is wrong. 

 

*****************

 

Reply from Avi:

Thanks for reply,

As far as I understand , If something went wrong with the vPC peer connection between the two nexus,

The secondary switch will suspend ONLY  the uplink and the downlink ports  that configure as vPC ports , But single port like x6  in our case, will get into orphan state , which mean that  the port remain linked but no traffic allowed from this port. Because of that the smart link will not function and we will lose traffic.

 

webpic298.jpg

 

**************

 

Input from Michael:

The network engineer’s assumption is correct. The Non-vPC ports will stay active with the peer-link down. However, this is by Cisco design and that traffic should still forward in a split brain scenario upstream (i.e. when the vPC link is down). But, the Cisco recommendation in your deployment scenario is to have a separate non-vPC link between N5Ks so that if the vPC peer link is down, that vMotion/management traffic can still forward east-west.

 

This is documented in their best practice guide: http://www.cisco.com/c/dam/en/us/td/docs/switches/datacenter/sw/design/vpc_design/vpc_best_practices_design_guide.pdf

“We recommend that you create an additional Layer 2 trunk port-channel as an interswitch link to transport

non-vPC VLAN traffic”

 

A quick topology of what I mean (sorry for the basic boxes):      

 

webpic297.png

 

This will solve the polarization issue with non-even links and maintain vMotion/Management traffic while the vPC peer link is down as well.

 

Let me know if you have any questions.

 

**************

 

Comments?