In this section:
To install the SBC on a virtual machine (VM) with Direct I/O pass-through, you must first configure the pass-through I/O devices on a ESXi host and create a VM and allocate its resources (for example CPU, memory, and NICs), as well as configure a datastore to contain the SBC operating system and application software. VMDirectPath I/O is a VMware technology that can be used with I/O hardware to reduce the CPU impact of high-bandwidth throughput workload's by "bypassing" the hypervisor. VM DirectPath I/O allows guest operating systems to directly access an I/O device, bypassing the virtualization layer and enhancing the performance.
You must follow the BIOS settings recommendations for the particular server. Refer to BIOS Settings Recommendations section for guidance.
To configure the Passthrough I/O devices on a ESXi host:
On the Configuration tab, click Advanced Settings.
The Passthrough Configuration page appears, listing all the available Passthrough devices.
A green icon indicates that a device is enabled and active. An orange icon indicates that the state of the device has changed and the host must be rebooted before the device can be used.
Select the devices to be used for Passthrough and click OK.
The figures shown in this procedure are intended as examples of the user interface and might not match the presented image exactly.
Reboot the Host.
Pass-through devices are not detected in VMware ESXi version Build–1483097 and VM does not get powered on.
Make sure all your NICs are physically plugged in (link light on) before creating your VM. Otherwise, when you perform the ISO, an incorrect port mapping occurs (logical to physical), and your SBC does not function properly.
Perform the following steps to create newSBC VM.
Login as user root on VMware vSphere client.
The vSphere Client main window appears.
The figures shown in this procedure are intended as examples of the user interface and might not match the presented image exactly.
In VMware vSphere Client main window, click Getting Started tab. Click Create a new virtual machine.
Select Custom in Configuration window, and click Next. Provide a Name for your SBC. The name can be up to 80 characters. Click Next.
Avoid special characters for name. Refer to System Name and Hostname Naming Conventions for details.
From the Storage screen, select a datastore1 and click Next. Ensure datastore has at least 100 GB or more free space. This datastore is required to store all log-related data files.
Select Virtual Machine Version: 11. Refer to For VMware for more information.
From Guest Operating System screen, make the following selections, and then click Next:
Select Linux as the Guest Operating System.
Select Debian GNU/Linux 9 (64-bit) from Version drop-down menu.
From CPUs screen, make following selections, and click Next. (In the following screenshot example, 4 cores are chosen). For CPU, memory configuration details, see For VMware.
Number of Virtual sockets: 1
Number of cores per virtual socket: 4 or above (depending on whether all virtual NICs are used)
From the Memory screen, assign memory for the virtual machine, and then click Next. For fewer than 6000 calls, the minimum reservation must be 10 GB. Anything more than 6000 calls require at least 14GB vRAM. Refer to SBC SWe Performance Metrics for more information.
Define Virtual Machine network interface connections (NICs) using following options from the drop-down menus. Then click Next to continue.
Select first network adapter (NIC 1:) for management interface. For example, the label can be MGMT which is created using Configuring vNetwork Standard Switches (vSwitches).
Select second network adapter (NIC 2:) for HA interface. For example, the label can be HA.
These network adapters and labels are already created on the ESX host server. If you are installing for the first time on new ESXi host server, these network adapters and corresponding labels (MGMT, HA) needs to be created. For details, see Configuring vNetwork Standard Switches (vSwitches). The PKT0, and PKT1 is created using the PCI devices. Refer step 14 through 19 for guidance.
Make sure that Network Adapters are mapped exactly in the following order as shown below:
Select Adapter type as VMXNET 3 for both MGMT and HA interfaces.
Ensure all Connect at Power On check boxes are checked. This must be Power On always.
From the SCSI Controller screen, select LSI Logic Parallel as the SCSI Controller option, then click Next to continue.
From the Select a Disk screen, select Create a new virtual disk option, then click Next to continue.
From the Create a Disk screen, make the following selections:
In Capacity section, assign a minimum of 100 GB or more disk space.
In Disk Provisioning section, choose only Thick Provisioning Eager Zeroed option:
Thick Provision Lazy Zeroed - Allocates the requested hard disk (virtual) during the VM creation. This type of disk pre-allocates and dedicates a pre-defined amount of a space for a virtual machine's disk operations, but it does not write zeroes to a virtual machine file system block until the first write within that region at run time.
Thick Provision Eager Zeroed(recommended) - Pre-allocates and dedicates a user defined amount of space for a VM disk operations.
Thin Provision - Creates virtual hard disk during runtime (on write operations). This provides more optimal hard disk usage, but it has some performance impact until it creates maximum requested virtual hard disk.
In Location section, select Specify a datastore or datastore cluster and click Browse... Select the datastore or NFS where ISO image is available.
Click Next to continue.
From the Advanced Options screen, keep the default value SCSI (0:0) and click Next. The default virtual device node is SCSI (0:0).
From the summary screen, review your VM settings and select the Edit the virtual machine settings before completion option. Click Continue. To go back and change the settings, click Back.
Click Add to continue the configuration on PKT0 and PKT1.
Choose the type of device to add and click Next. For Direct I/O interface, you must select the PCI device.
Select the physical PCI/PCI2 device to connect the packet interface PKT0 and click Next.
Find the PKT port NUMA affinity by executing the following command on the EXSi host:
vsish -e get /net/pNics/<PKT port name - vmnicX>/properties | grep "NUMA"
Edit Settings -> VM options -> Configuration Parameters -> Add Parameters.
Add the following parameters:
numa.nodeAffinity’ = 0 or 1 (based on PKT port NIC affinity) numa.autosize.once = FALSE
Click Finish.
Repeat the step 14 through 17 to add the physical device to connect the PKT1 interface.
After adding the PCI devices for both PKT0 and PKT1 interfaces, click Finish to complete.
All the NIC ports of the same NIC card has to be fully configured as either pass-through devices or for virtual networking (vSwitches). This is a VMware ESXi limitation/restriction.
The virtual machine is created under host IP address with the specified configuration. See the below example screen.
This creates the virtual machine.
Select the host IP address from left-pane and click Configuration Tab. The Virtual Machine Startup and Shutdown section is displayed.
Click Properties link displayed towards the top-right window.
In System Settings pane, you must ensure to select the Allow virtual machine to start and stop automatically with the system check box. By default it is selected.
In Startup Order pane, select the VM, which you want to automatically start and click Move Up.
The selected VM is displayed underneath Automatic Startup.