[PVE-User] Ceph df

Сергей Цаболов tsabolov at t8.ru
Tue Feb 1 14:59:59 CET 2022


Hello Alwin,

In this post 
https://forum.proxmox.com/threads/ceph-octopus-upgrade-notes-think-twice-before-enabling-auto-scale.80105/#post-399654

I read about *target ratio to 1 and call it a day *, in my case I set to 
vm.pool  Target ratio 1 :

ceph osd pool autoscale-status
POOL                                         SIZE TARGET SIZE      RATE  
RAW CAPACITY   RATIO  TARGET RATIO EFFECTIVE RATIO  BIAS  PG_NUM  NEW 
PG_NUM  AUTOSCALE
device_health_metrics          22216k       500.0G   2.0 106.4T  0.0092 
                                                     1.0            8 on
vm.pool                                     2734G             3.0        
106.4T  0.0753        1.0000                                       
0.8180 1.0             512 on
cephfs_data                                    0              2.0        
106.4T  0.0000             0.2000 0.1636   1.0             128         
     on
cephfs_metadata                  27843k         500.0G 2.0        
106.4T  0.0092    4.0              32              on

What you think  I need to set target ratio on cephfs_metadata & 
device_health_metrics?

To pool  cephfs_data I set the target ratio 0.2  .

Or the target ration on vm.pool need not the *1* but more?


*
*

31.01.2022 15:05, Alwin Antreich пишет:
> Hello Sergey,
>
> January 31, 2022 9:58 AM, "Сергей Цаболов"<tsabolov at t8.ru>  wrote:
>> My question is how I can  decrease MAX AVAIL in default pool
>> device_health_metrics + cephfs_metadata and set it to vm.pool and
>> cephfs_data
> The max_avail is calculated by the cluster-wide AVAIL and pool USED, with respect to the replication size / EC profile.
>
> Cheers,
> Alwin
>
Sergey TS
The best Regard

_______________________________________________
pve-user mailing list
pve-user at lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-user



More information about the pve-user mailing list