[gpfsug-discuss] Anyone using RoCE?

Luke Sudbery l.r.sudbery at bham.ac.uk
Fri Jan 23 16:38:43 GMT 2026


Is anyone using RoCE with good results? We are planning on it, but initial tests are not great - we get much better performance using plain Ethernet over the exact same links.

It's up and working, I can see RDMA connections and counters, no errors, but performance is unstable. And worse than Ethernet, which was just meant to be a sanity check!

Things I've looked at based on Lenovo and IBM guides, which I think are all configured correctly:

  *   RoCE interfaces all on the same subnet
  *   They all have IPv6 enabled with  addresses using eui64 addr-gen-mode
  *   DSCP trust mode on NICs
  *   PFC flow control on NICs
  *   Global Pause disabled on NICs
  *   ToS configured for RDMA_CM
  *   Source based routing for multiple interfaces on the same subnet.
  *   Switches (nvidia cumulus) all enabled for RoCE QOS

Iperf and GPFS over plain Ethernet get nearly 3GB/s, which is near the line speed of the NIC in question - 25Gbps. Testing basic RDMA connections with ib_send_bw gets about the same. But GPFS over RoCE gets from 0.7GB/s to 1.9GB/s.

The servers have 4x 200G Mellanox cards. The client has 1x 25G card. What's frustrating and confusing is that we get better performance when we just enable 1 card at the server end, and also get better performance if we have 1 fabric ID per NIC on the server (with all 4 fabric ID on the same NIC at the client end).

I can go into more details if anyone has experience! Does this sound familiar to anyone? I am planning to open a call with Lenovo and/or IBM as I'm not quite sure where to look next.

Cheers,

Luke

--
Luke Sudbery
Principal Engineer (HPC and Storage).
Architecture, Infrastructure and Systems
Advanced Research Computing, IT Services
Room 132, Computer Centre G5, Elms Road

Please note I don't work on Monday.

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20260123/20ef5a3b/attachment.html>


More information about the gpfsug-discuss mailing list