Avatar of snyderkv
 asked on

Standard practice vSwitch design?

We have sites with inconsistent vSwitching configurations so I wanted to ask for a good vSwitching document that lays this out based on most setups in the field.

One layout we have is 4 vmnics going to  two upstream Nexus switches.
vmnic0 and 2 going to Nexus A
vmnic1 and 3 going to Nexus B

vmnic0 and vmnic 1 connected to Mgmt, server port groups and vMotion.
vmnic 2 and vmnic 3 connected to Mgmt, Storage and vMotion.

It's all on vSwitch0. It just doesn't look right and theirs configuration drift among other hosts as well.

Avatar of undefined
Last Comment

8/22/2022 - Mon
Andrew Hancock (VMware vExpert PRO / EE Fellow/British Beekeeper)

Basically what you should have is at least dual network interfaces per service,


Management Network - 2 nics
vMotion - 2 nics
Storage - 2 nics

Virtual Machines - 2 nics

and then the teaming policy per 2 nics, need to be configured to match your physical switches.

Then, Management, vMotion, Storage should all be on sperate individual networks/subnets or VLANs - what ever is in use at your organisation.

You can then also spread vMotion, Storage, Management, Virtual Machines across vSwitches.

It's very flexible, and there is not really a right or wrong way. BUT it should be identical on ALL hosts to make it easier to document and rebuild.

Ok but what if you only have 4 nics? That's my situation. Management is already dedicated to Management traffic via check box and vMotion as well so perhaps they can both use the same two nics? Then put storage on its own two nics? Storage traffic is already on different subnets so that traffic is already separated.

So why not all 4 nics for all traffic since they are already tagged for their separate traffic and on separate subnets/vlans.

Also, why should I add a another vSwitch if the upstream switches is the single point of failure? Some sites have a separate vSwitch for NFS but I don't see how that makes them redundant as the vSwitches never really fail anyways.

vmware has to have some detailed docs on this.
Andrew Hancock (VMware vExpert PRO / EE Fellow/British Beekeeper)

NIC teams or multipath provide redundancy, but if upstream there is no redundancy, then there is little point...

Every network organisation is different, and ESXi provides lots of flexibility to craft your own.

VMware does have documentation, but it's in basic form


You need to establish, if you have enough bandwidth with 2 or 4 nics for everything.

Seperate vSwitches, seperates traffic, and iSCSI needs MPIO so a nic per VMKernel, sometimes it can be difficult to bung it all on a single vSwitch.

(because you need to use port overrides and binding)
All of life is about relationships, and EE has made a viirtual community a real community. It lifts everyone's boat
William Peck


Log in or sign up to see answer
Become an EE member today7-DAY FREE TRIAL
Members can start a 7-Day Free trial then enjoy unlimited access to the platform
Sign up - Free for 7 days
Learn why we charge membership fees
We get it - no one likes a content blocker. Take one extra minute and find out why we block content.
Not exactly the question you had in mind?
Sign up for an EE membership and get your own personalized solution. With an EE membership, you can ask unlimited troubleshooting, research, or opinion questions.
ask a question