Hi,
I am just setup proxmox cluster with 3 node, use linstor as storage backend with 3 replica, and create some VM with linstor storage backend. I install linstor gui as well. However, when I see a resource from linstor gui, there is different size. Please see below
If you want extra assurance your data is not corrupted or out of sync, you can always perform a quick validation on the resource:
The resource cannot be Primary (or InUse), make sure the VM has been stopped.
On each node (pve21, pve22, pve23) perform a sha256sum /dev/drbd1019. Do this in order, one at a time. Each host should show as Primary when calculating the sum.
Verify each host has matching hash values for the /dev/drbd1019 volume.
Assuming the hash values match, there isn’t anything you need to do, but you can delete the volumes containing the larger data allocations and re-add them. This should have the added effect of only consuming ~10 MiBs on each node. This is due to LINSTOR’s ability to only synchronize the allocated data when using thinly provisioned volumes.
With that said, LINSTOR cannot control the “drift” that is possible with thinly provisioned storage across multiple nodes in the cluster.
Additional information. The resource on pve22 is diskless previously. The resource automatic created when I tick “Diskless on remaining” on resource group
unrelated to the rest, but: don’t do that. the proxmox plugin creates (and deletes) them on the fly as required. “diskless on remaining” is only there if there is no other way to create diskless assignments on the fly, but in this case there is.
Given that the replica on pve22 also shows almost zero usage, it does seem like that replica is broken. It would be interesting to know exactly how you got it into that state.
I suggest removing and re-adding the resource on pve22, and check the responses carefully for any errors while you do that.
As you’ve already been told, please don’t tick this. You don’t want it. Proxmox will add diskless resources where required.
Once you’ve re-added the disk, you can repeat the sha256sum exercise given before. If pve22’s replica is still different, then something is bad. You might want to start by describing the exact versions of Linstor, drbd9 kernel module, and underlying OS that you’re using.
If pve22 is still different… I am wondering whether there’s some chance that when the logical volume is being recreated, it’s reusing the same extents without zeroing them - i.e. there is valid metadata, and so drbd thinks the sync is already complete.
In that case, it might be best to force a full resync of the volume.
I think you would do that by logging into pve22 and doing drbdsetup invalidate XXXX (where XXXX is the minor number, matching /dev/drbdXXXX) but I’m not 100% sure.
I was unable to recreate this in my test cluster using the same steps you mentioned using in the LINSTOR GUI. Did you see this happen more than once, or was it something you could reproduce?