- Integrated Systems
- About Us
- Integrated Systems
- About Us
Dec 15, 4:00 am to 10:00 am UTC
Dec 14, 10:00 pm CST to Dec 15, 4:00 am CST
Dec 14, 8:00 pm PST to Dec 15, 2:00 am PST
03-30-2015 02:07 PM
Virtual Connect Flex-10 problem?
A Virtual Connect question from Zdravko:
On the picture is shown diagram of deployment which we are trying to implement using HP C7000 Chassis with 2 HP Flex 10 and HP Proliang 460c Gen 8 with VMware ESXi on it.
On Vmware we have approximate 126 VLANs, and Nexus 1000v as a DVS. We’ve tried two scenarios to worked this out.
We’ve configured VC connect Domain with vNet in Tunnel mode. In this scenario, the upstream network switch is configured to pass multiple VLANs to two ports on each VC module. The upstream switch ports is configured as “trunk” ports (using LACP) for several VLANs. However in this Scenario we’ve experienced partial or total loss of connectivity for servers in VLANs which are in “bridged-transparent” mode by F5 LB.
We believe we have encounter a scenario described in: http://h20565.www2.hp.com/hpsc/doc/public/display?calledBy=&ac.admitted=1410467226755.876444892.492883150&docId=emr_na-c02684783-2&docLocale=
According to recommendation we’ve disabled tunneling of VLAN tags and implemented map VLAN tags instead with Shared Uplink Set. With this, problem with loss of connectivity was solved.
However again we’ve encounter another limiting factor for scalability, number of VLANs – 162 Unique Server Mapped VLAN; now we are using 126. (http://h20565.www2.hp.com/hpsc/doc/public/display?docId=emr_na-c03278211 – page 4). This scenario is unacceptable since we have growth plans of nearly triple in next 4 years, so we might hit the limit of 162 VLANs very soon. We use this HW for private/custom Cloud solutions, and every customer is a separate VLAN, or a possible tweak/workaround to scenario 1 so that we don’t encounter those problems?
Is there another recommended/supported deployment for our Scenario using HP Flex 10, which does not have above limitation/problems?
Input from Dan:
First off, you cannot use LACP to Aggregate 2 uplinks from different Flex-10 modules as their picture and text describes. LACP requires that the ports on each side terminate in the same switch. Each VC module is considered a different switch. So if they only have 1 uplink from VC Bay 1 and 1 uplink from VC Bay 2, they need to delete the Port Channel interface from the Cisco side.
As far as the Customer Advisory, you need to confirm with them if they are using 2 VLANs or just 1 for their Load Balancer.
There were some VLAN enhancements in 4.30, but if they are truly USING 126 VLANs on each server, then these enhancements won’t help them break the 162 barrier.
The enhancements are such that you can plumb more than 1000 VLANs into VC Mapped Mode as long as the servers are using less than 162 per port and the total in use per VC Domain is less than 1000.
In a pinch, an alternative might be Broadcom NPAR with a 6125XLG instead of Flex-10.
I would have them test that though before they make the decision to switch.