Hi,
I have two Nvidia ConnectX-7 200GbE / NDR200 / MCX755106AC-HEAT network cards
installed in two Lenovo ThinkSystem SR630 V3 servers (PCIe Gen 5).
These cards are connected directly to each other (back-to-back) on both ports using two QSFP56 200Gbps cables (link: https://marketplace.nvidia.com/en-us/enterprise/networking/200gbeqsfp56cables/).
I am running Ubuntu 24.04 on both servers.
And transmitting packets (iperf2, 16 tcp streams) from server 1 to server 2 on both ports (2 network interfaces) in parallel .
I can’t seem to exceed total of100Gbps across the two ports together:
- 1 network interface achieve 100 Gbps
- 2 network interfaces together achieve 50Gbps each. (100 Gbps total)
According to the spec, I should be able to achieve 400Gbps via Ethernet.
Things I’ve done:
-
updated the firmware to the latest versions:
- FW: 28.43.2026
- PXE: 3.7.0500
- UEFI: 14.36.0021
-
Increased MTU on both interfaces to 9000
-
sysfs indicate on the correct speed on both network interfaces
$ cat /sys/class/net/ens1f0np0/speed
200000
- ethtool indicate on 200Gbps link speed on each network interface
$ ethtool ens2f0np0
Settings for ens2f0np0:
Supported ports: [ Backplane ]
...
Supports auto-negotiation: Yes
...
Advertised pause frame use: Symmetric
Advertised auto-negotiation: Yes
Advertised FEC modes: RS
Link partner advertised link modes: Not reported
Link partner advertised pause frame use: No
Link partner advertised auto-negotiation: Yes
Link partner advertised FEC modes: Not reported
Speed: 200000Mb/s
Duplex: Full
Auto-negotiation: on
Port: Direct Attach Copper
PHYAD: 0
Transceiver: internal
Supports Wake-on: d
Wake-on: d
Link detected: yes
- mlxconfig output: https://pastebin.com/raw/RLKckxH8
- Checked all tips from ESPCommunity
- Checked all tips from ESPCommunity
Any insights or suggestions would be greatly appreciated.
Thanks!