I test seperate VM on local SSD & ESXi 6.5 host.
Here is briefs
ESX host 01 - iPerf VM 01 - SX6036G - ESX host 02 - iPerf VM 02
All switch & vSwitch, vKernel adapter, nmlx_en adapter set to 9k MTU.
And ESXi 6.5 update 1 inbox driver include nmlx4_core that support Mellanox ConnectX-3, Pro and belows.
Why did you comment that based your personal memorandum?
In another thread on this community said CX-3 supported with ESXi 6.5 inbox driver.
I’ll test Arista 40Gb switch with same configuration then I’ll update result.
Update 01. Direct connection between 2 of CX-3s test
This test shows almost 19Gb/s grade performance
Update 02. SX6036G Global Pause On switched test
This test also shows almost 19Gb/s grade performance
Update 03. SX6036G all port 10GbE switched test
This test also shows almost 10Gb/s grade performance
Update 04. ESXi inbox Ethernet driver packet drop bug - ESXi host iPerf test results…:(
It must have a packet drop & error bug in Mellanox inbox ESXi drivers.
How do you think about below?
Update 05. Windows, Linux Host test
All Windows, Linux Host show me a greate performance on 40GbE configurations that include CX-3 HCAs & SX6036G switches.
I can build a 56Gb Ethernet TCP/IP network with ConnectX-3 (firmware 2.42.5000) & SX6036G (firmware 3.6.4006), now.
packet drop was not appeared that I don’t use PFC configuration with ConnectX-3 ESXi 6.5 update 1.
40, 56GbE Link-up require small delay to establish requested speed.
But performance level is terrible like below.
In 56GbE mode, I’m test iPerf test between physical ESXi host then I met a almost 6.11Gb/s performance only.
* 56GbE iPerf client - physical ESXi 6.5 update 1 host 01 with MTU 9000
* 56GbE iPerf server - physical ESXi 6.5 update 1 host 02 with MTU 9000
Is there any configuration shold be need on SX6036G?
* This is a traditional TCP/IP Network, not RDMA Network.
* When 56Gb ethernet show me a 56GbE throughput, I’ll switch to another test like iSER driver 188.8.131.52 and so on.