Iโm new to Linstor and trying to make a 2 node proxmox setup with as much redundancy as I can within my cluster size constraints.
Both nodes have 2x mirrored nvme drives with an lvm that is then used by DRBD.
The nodes have a 25Gb link directly between them for DRBD replication. But the servers also have a 1Gb interface (management and proxmox quorum), and a 10Gb interface (NAS, internet, and VM migration).
I would like to use the 10Gb interface as a failover in case the direct link goes down for some reason, but it should not usually be used by DRBD. I couldnโt find a way to do this properly with DRBD networks. So, Iโve created a primary/backup bond in Linux and use the bond interface for DRBD. That way Linux handles all failover logic.
On my NAS (truenas) I have a VM that will be a diskless witness. This VM has a loop back interface with an ip on the DRBD network, but uses static routes to route that traffic over either the 1Gb interface or the 10Gb interface. This way itโs also protected from a single link failure.
My problem is that when trying to move a VM disk over to the DRBD storage for testing, the performance is horrible. Looking at the network interfaces, it starts out at around 3Gb, but soon drops to around 1Gb or lower. Doing a iperf3 test gives 24Gb (with MTU 9000), so itโs not a network problem. I also have the same issue if I remove the witnesses, so thatโs not the cause either.
Is it just my whole implementation thatโs stupid? Which config files or logs would be most useful for debugging this?