Why would you want to use an internal PCIe Switch?
Current generation Intel processors are limited to 48 PCIe lanes per CPU. Given some of these lanes need to be used for non-storage uses (Internal LOM connection, Network Adapters, GPUs) this often leaves at most 40 to 44 PCIe lanes for storage devices. The most common U.2 form factor NVMe drives connect to a 4x PCIe backplane connector to talk to the CPU. This presents a problem on dense NVMe configurations with a large number of drives where you potentially need more PCIe lanes than physically exist. Even with 2 Intel CPU sockets (which adds additional PCIe lanes as well as memory bandwidth), you will need PCIe switches for dense NVMe drive configurations. When looking to deploy vSAN hosts with more than 10 drives per Intel CPU, be aware that you will need to use certified vSAN ReadyNode designs to accommodate for the included PCIe switches that will be required for these configurations to work.
Be sure to ask your VAR or server OEM, if a PCIe switch will be required for dense NVMe configurations. Do note, that less dense configurations (Example 2 NVMe cache devices and 8 SAS capacity devices) would not need a PCIe switch as the 8 lanes that would be used by these 2 NVMe drives would not necessitate oversubscription. Also when choosing chassis for future expandability.
What about SAS Expanders?
So what if you want to “build your own” VSAN server today, and want to use more than 8 devices? The simple and easy way is to purchase a SAS pass-through HBA for each drive group. This has a number of advantages over SAS expanders.
- Dedicating a SAS HBA to each disk group breaks up internal fault domains, and reduces the impact of a failed controller.
- Dedicating a SAS HBA to each disk group increases the available queue depth and throughput for performance and faster rebuilds. We previously covered on the blog that this is a great way to linearly “scale up” performance inside each host. For all-flash configurations, this is especially important.
- Dedicating a SAS HBA to each disk group reduces the amount of firmware/drivers that must be tracked and accounted for in updates. The vSAN health service can help with this task.
Another question that comes up is using external storage expansion. The compute module and an external storage module are certified as a whole VMware vSAN ReadyNode. This is primarily for blade or modular systems that lack a number of internal drive bays. As with the SAS expanders, pay attention to the number of drives supported.
Examples of this include:
- AF-8: DELL FX2 FC630 – A modular server/blade hybrid solution with incredible density can leverage FD332 systems for expansion.
- HPE Synergy
- Dell MX MX5016s Storage Sled provides up to 16 extra SAS drives
VSAN offers extensive flexibility and performance across a huge number of server and storage device vendors. SAS expanders, and pass through HBA’s allow for flexibility incapacity, and performance when designing. The VSAN ready nodes act as a great simple quick place to start when designing a VSAN solution.