[PVE-User] Help with Ceph in PVE6

Gilberto Nunes gilberto.nunes32 at gmail.com
Sat Mar 28 15:08:24 CET 2020


[UPDATE]
I notice that in [node] -> Ceph -> Pool in Used % column the values is
decrease over time! Perhaps need wait to adjust it and than see if
active+remapped+backfill_wait and  active+remapped+backfilling end it's
operations...
---
Gilberto Nunes Ferreira


Em sáb., 28 de mar. de 2020 às 11:04, Gilberto Nunes <
gilberto.nunes32 at gmail.com> escreveu:

> Help with Ceph in PVE 6
>
> Hi
>
> I have a ceph cluster created with 3 server....
> ServerA has 3 SAS 512GB HDD and 1 SAS 1.3 TB
> ServerB has 3 SAS 512GB HDD and 1 SAS 1.3 TB
> ServerS has 3 SAS 512GB HDD and 1 SAS 1.3 TB
>
> I have one pool named VMS with size/min 3/2 and pg_num initially created
> with 256 but I have increased to 512 and an hour ago to 768 but it's see
> hava not effect on it...
>
> Ceph health apparently is ok but get this with ceph -s command:
>
> ceph -s
>   cluster:
>     id:     93c55c6b-ce64-4e1a-92bc-0bc529d695f2
>     health: HEALTH_OK
>
>   services:
>     mon: 5 daemons, quorum pve3,pve4,pve5,pve7,pve6 (age 15h)
>     mgr: pve3(active, since 15h), standbys: pve4, pve5, pve7, pve6
>     osd: 12 osds: 12 up (since 10m), 12 in (since 10m); 497 remapped pgs
>
>   data:
>     pools:   1 pools, 768 pgs
>     objects: 279.34k objects, 1.1 TiB
>     usage:   3.0 TiB used, 6.2 TiB / 9.1 TiB avail
>     pgs:     375654/838011 objects misplaced (44.827%)
>              494 active+remapped+backfill_wait
>              271 active+clean
>              3   active+remapped+backfilling
>
>   io:
>     client:   140 KiB/s rd, 397 KiB/s wr, 12 op/s rd, 64 op/s wr
>     recovery: 52 MiB/s, 14 objects/s
>
>
> Is there any action I can take to fix this?
>
> Thanks
> ---
> Gilberto Nunes Ferreira
>
> (47) 3025-5907
> (47) 99676-7530 - Whatsapp / Telegram
>
> Skype: gilberto.nunes36
>
>
>
>



More information about the pve-user mailing list