[PVE-User] Analysis of free space...
Matthieu Dreistadt
matthieu at 3-stadt.de
Sat Sep 27 18:06:42 CEST 2025
Hi Marco,
you can check "zfs list -o space", which will give you a more detailed
view of what is using the space:
root at xxx:~# zfs list -o space
NAME AVAIL USED USEDSNAP USEDDS
USEDREFRESERV USEDCHILD
rpool 507G 354G 0B 104K 0B
354G
rpool/ROOT 507G 4.40G 0B 96K 0B
4.40G
rpool/ROOT/pve-1 507G 4.40G 1.05G 3.35G 0B
0B
rpool/data 507G 312G 0B 112K 0B
312G
rpool/data/subvol-105-disk-0 8.62G 11.4G 49.2M 11.4G 0B
0B
Used = overall used
Usedsnap = Used by Snapshots
Usedds = Used Disk Space (not counting snapshots, only live data)
Usedchild = Used by datasets/zvols further down in the same path (in my
example, rpool has the same amount of Used and Usedchild space, since
there is nothing directly inside of rpool itself)
Cheers,
Matthieu
Am 24.09.2025 um 18:29 schrieb Marco Gaiarin:
> Mandi! Marco Gaiarin
> In chel di` si favelave...
>
>> Uh, wait... effectively we forgot to enable 'discard' on volumes, and we have
>> enabled afterward (but rebooted the VM).
>> I'll check refreservation property and report back.
> No, volumes seems have all refreservation to 'none', as expected; current
> situation is:
>
> root at lamprologus:~# zfs list | grep ^rpool-data
> rpool-data 54.2T 3.84T 171K /rpool-data
> rpool-data/vm-100-disk-0 1.11T 3.84T 1.11T -
> rpool-data/vm-100-disk-1 2.32T 3.84T 2.32T -
> rpool-data/vm-100-disk-10 1.82T 3.84T 1.82T -
> rpool-data/vm-100-disk-11 2.03T 3.84T 2.03T -
> rpool-data/vm-100-disk-12 1.96T 3.84T 1.96T -
> rpool-data/vm-100-disk-13 2.48T 3.84T 2.48T -
> rpool-data/vm-100-disk-14 2.21T 3.84T 2.21T -
> rpool-data/vm-100-disk-15 2.42T 3.84T 2.42T -
> rpool-data/vm-100-disk-16 2.15T 3.84T 2.15T -
> rpool-data/vm-100-disk-17 2.14T 3.84T 2.14T -
> rpool-data/vm-100-disk-18 3.39T 3.84T 3.39T -
> rpool-data/vm-100-disk-19 3.40T 3.84T 3.40T -
> rpool-data/vm-100-disk-2 1.32T 3.84T 1.32T -
> rpool-data/vm-100-disk-20 3.36T 3.84T 3.36T -
> rpool-data/vm-100-disk-21 2.50T 3.84T 2.50T -
> rpool-data/vm-100-disk-22 3.22T 3.84T 3.22T -
> rpool-data/vm-100-disk-23 2.73T 3.84T 2.73T -
> rpool-data/vm-100-disk-24 2.53T 3.84T 2.53T -
> rpool-data/vm-100-disk-3 213K 3.84T 213K -
> rpool-data/vm-100-disk-4 213K 3.84T 213K -
> rpool-data/vm-100-disk-5 2.33T 3.84T 2.33T -
> rpool-data/vm-100-disk-6 2.28T 3.84T 2.28T -
> rpool-data/vm-100-disk-7 2.13T 3.84T 2.13T -
> rpool-data/vm-100-disk-8 2.29T 3.84T 2.29T -
> rpool-data/vm-100-disk-9 2.11T 3.84T 2.11T -
>
> and a random volume (but all are similar):
>
> root at lamprologus:~# zfs get all rpool-data/vm-100-disk-18 | grep refreservation
> rpool-data/vm-100-disk-18 refreservation none default
> rpool-data/vm-100-disk-18 usedbyrefreservation 0B -
>
> Another strange thing is that all are 2TB volumes:
>
> root at lamprologus:~# cat /etc/pve/qemu-server/100.conf | grep vm-100-disk-19
> scsi20: rpool-data:vm-100-disk-19,backup=0,discard=on,replicate=0,size=2000G
>
> but:
>
> root at lamprologus:~# zfs list rpool-data/vm-100-disk-19
> NAME USED AVAIL REFER MOUNTPOINT
> rpool-data/vm-100-disk-19 3.40T 3.84T 3.40T -
>
> why 'USED' is 3.40T?
>
>
> Thanks.
>
More information about the pve-user
mailing list