[PVE-User] High ceph OSD latency
Fabrizio Cuseo
f.cuseo at panservice.it
Thu Jan 15 13:21:38 CET 2015
I will check, but the latency in osd tree is for each disk, and i have high latency on all osd; this is why i don't think that the problem is related to one host or disk.
Inviato da iPad
> Il giorno 15/gen/2015, alle ore 13:17, Lindsay Mathieson <lindsay.mathieson at gmail.com> ha scritto:
>
>> On Thu, 15 Jan 2015 11:25:44 AM Fabrizio Cuseo wrote:
>> What is strange is that on OSD tree I have high latency: tipically Apply
>> latency is between 5 and 25, but commit lattency is between 150 and 300
>> (and sometimes 5/600), with 5/10 op/s and some B/s rd/wr (i have only 3
>> vms, and only 1 is working now, so the cluster is really unloaded).
>>
>> I am using a pool with 3 copies, and I have increased pg_num to 256 (the
>> default value of 64 is too low); but OSD latency is the same with a
>> different pg_num value.
>>
>> I have other clusters (similar configuration, using dell 2950, dual ethernet
>> for ceph and proxmox, 4 x OSD with 1Tbyte drive, perc 5i controller), with
>> several vlms, and the commit and apply latency is 1/2ms.
>>
>> Another cluster (test cluster) with 3 x dell PE860, with only 1 OSD per
>> node, have better latency (10/20 ms).
>>
>> What can i check ?
>
>
> POOMA U, but if you have one drive or controller that is marginal or failing,
> it can slow down the whole cluster.
>
> Might be worth while benching individual osd's
More information about the pve-user
mailing list