[PVE-User] truenas high i/o wait

Tonči Stipičević tonci at suma-informatika.hr
Sun Oct 6 15:01:10 CEST 2024


continued from previous post :

on the other host  TrueNAS VM (TrueNAS-13.0-U6.2)   is causing 1GB/s 
read ... :    ...   Is this host issue or VM internal issue ?  ... where 
to "intervene"?


IOTOP:

Total DISK READ:      1037.23 M/s | Total DISK WRITE: 1060.48 K/s
Current DISK READ:     702.02 M/s | Current DISK WRITE:    1679.83 K/s
     TID  PRIO  USER    DISK READ>  DISK WRITE COMMAND
1803753 be/4 root      283.67 M/s    0.00 B/s kvm -id 106 -name 
TN0103,debug-threads=on -no-shutdown -chardev 
socket,id=~024,tx_queue_size=256,bootindex=102 -machine type=pc+pve0 
[iou-wrk-1726849]
1803546 be/0 root       59.29 M/s  811.01 B/s [zvol_tq-3]
1803545 be/0 root       56.26 M/s    0.00 B/s [zvol_tq-3]
1803544 be/0 root       54.22 M/s    0.00 B/s [zvol_tq-3]


Thank you

and

BR

srdačan pozdrav / best regards

Tonči Stipičević, dipl. ing. elektr.
direktor / manager

SUMA Informatika d.o.o., Badalićeva 27, OIB 93926415263

Podrška / Upravljanje IT sustavima za male i srednje tvrtke
Small & Medium Business IT Support / Management

mob: 091 1234003
www.suma-informatika.hr

On 06. 10. 2024. 14:42, Tonči Stipičević wrote:
> Hello to all,
>
> I've been using for years TrueNAS as full VM in Prox host.  No sata 
> pass-through.
>
> Host data pool consists of 6 x 4T enterprise sata drives in raid10  
> and that is where TrueNAS  disks images reside. ( 1 x boot, 3 x 2T 
> virtual disk stripped)
>
> 'Till last week everything was working smoothly (in real time) but 
> after upgrade to latest Prox host (community/no-subs)  TrueNAS VM 
> (also on latest version)  started generating very high i/o waits 
> (60-70%).
>
> This host hosts 10 VMs and when TrueeNAS VM is not running , i/o wait 
> lowers down to 2-3% and CPU %  runs between 20-30% depending on the VM 
> load.  Most important -> CPU % utilization is much higher than i/o 
> wait ...
>
> But , as soon as TrueNAS boots  i/o waits jumps up to 60-70% ... and 
> everything slows down ...
>
> I moved all other VMs to another host so to eliminate all possible 
> influences , but TrueNAS alone generates the same high i/o wait ...
>
> Is there any way to debug this high i/o wait  problem ? ...
>
> Total DISK READ:       208.71 M/s | Total DISK WRITE:         0.00 B/s
> Current DISK READ:     171.33 M/s | Current DISK WRITE:       0.00 B/s
>     TID  PRIO  USER    DISK READ>  DISK WRITE  SWAPIN      IO COMMAND
> 1189075 be/4 root       50.80 M/s    0.00 B/s  ?unavailable?  kvm -id 
> 100 -name TN02,debug-threads=on -no-shutdown -chardev 
> socket,id=qmp,path=/v~d=net0,rx_queue_size=1024,tx_queue_size=1024 
> -machine type=pc+pve0 [iou-wrk-876705]
>     370 be/0 root       40.69 M/s    0.00 B/s  ?unavailable? [z_rd_int]
> 1189069 be/4 root       17.61 M/s    0.00 B/s  ?unavailable?  kvm -id 
> 100 -name TN02,debug-threads=on -no-shutdown -chardev 
> socket,id=qmp,path=/v~d=net0,rx_queue_size=1024,tx_queue_size=1024 
> -machine type=pc+pve0 [iou-wrk-876705]
> 1188348 be/0 root       10.78 M/s    0.00 B/s  ?unavailable? [zvol]
>
> iotop shows that  z_rd_int and kvm (TrueNAS VM)  reads a lot ... Why ? 
> ... why is so much reading needed   TrueNAS is doing nothing (VM 
> internal 4-5% cpu utilization) , no file transfer or such ....
>
>
> Till this weekend everything worked seemless and I 've never thought 
> about migrating TrueNAS on baremetal or sata-passthrough
>
>
> Thank you very very much for your help and support
>
>
> Best regards
>
> Tonči
>



More information about the pve-user mailing list