[PVE-User] Proxmox Ceph with differents HDD Size
Gilberto Nunes
gilberto.nunes32 at gmail.com
Fri Aug 31 16:08:33 CEST 2018
Thanks for all buddies that replied my messages.
Indeed I used
ceph osd primary-affinity <osd-id> <weight>
And we felt some performance increment.
What's help here is that we have 6 proxmox ceph server:
ceph01 - HDD with 5 900 rpm
ceph02 - HDD with 7 200 rpm
ceph03 - HDD with 7 200 rpm
ceph04 - HDD with 7 200 rpm
ceph05 - HDD with 5 900 rpm
ceph06 - HDD with 5 900 rpm
So what I do is define weight 0 to HDD's with 5 900 rpm and define weight 1
to HDD's with 7 200 rpm.
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 62.31059 root default
-3 14.55438 host pve-ceph01
0 hdd 3.63860 osd.0 up 1.00000 0
1 hdd 3.63860 osd.1 up 1.00000 0
2 hdd 3.63860 osd.2 up 1.00000 0
3 hdd 3.63860 osd.3 up 1.00000 0
-5 10.91559 host pve-ceph02
4 hdd 2.72890 osd.4 up 1.00000 1.00000
5 hdd 2.72890 osd.5 up 1.00000 1.00000
6 hdd 2.72890 osd.6 up 1.00000 1.00000
7 hdd 2.72890 osd.7 up 1.00000 1.00000
-7 7.27708 host pve-ceph03
8 hdd 2.72890 osd.8 up 1.00000 1.00000
9 hdd 2.72890 osd.9 up 1.00000 1.00000
10 hdd 1.81929 osd.10 up 1.00000 1.00000
-9 7.27716 host pve-ceph04
11 hdd 1.81929 osd.11 up 1.00000 1.00000
12 hdd 1.81929 osd.12 up 1.00000 1.00000
13 hdd 1.81929 osd.13 up 1.00000 1.00000
14 hdd 1.81929 osd.14 up 1.00000 1.00000
-11 14.55460 host pve-ceph05
15 hdd 7.27730 osd.15 up 1.00000 0
16 hdd 7.27730 osd.16 up 1.00000 0
-13 7.73178 host pve-ceph06
17 hdd 0.90959 osd.17 up 1.00000 0
18 hdd 2.72890 osd.18 up 1.00000 0
19 hdd 1.36440 osd.19 up 1.00000 0
20 hdd 2.72890 osd.20 up 1.00000 0
Tha's it! Thanks again.
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
2018-08-30 11:47 GMT-03:00 Phil Schwarz <infolist at schwarz-fr.net>:
> Hope you did change a single disk at a time !
>
> Be warned (if not) that moving an OSD from a server to another triggers
> a rebalancing of almost the complete datas stored upon in order to
> follow crushmap.
>
> For instance exchanging two OSDs between servers result in a complete
> rebalance of the two OSDS,a ccording to my knowledge.
>
> 16% of misplaced datas could be acceptable or not depending on your
> needs of redundancy and throughput, but it's not a low value that could
> be underestimated.
>
> Best regards
>
>
>
> Le 30/08/2018 à 15:27, Gilberto Nunes a écrit :
> > Right now the ceph are very slow
> >
> > 343510/2089155 objects misplaced (16.443%)
> > Status
> >
> > HEALTH_WARN
> > Monitors
> > pve-ceph01:
> > pve-ceph02:
> > pve-ceph03:
> > pve-ceph04:
> > pve-ceph05:
> > pve-ceph06:
> > OSDs
> > In Out
> > Up 21 0
> > Down 0 0
> > Total: 21
> > PGs
> > active+clean:
> > 157
> >
> > active+recovery_wait+remapped:
> > 1
> >
> > active+remapped+backfill_wait:
> > 82
> >
> > active+remapped+backfilling:
> > 2
> >
> > active+undersized+degraded+remapped+backfill_wait:
> > 8
> >
> > Usage
> > 7.68 TiB of 62.31 TiB
> > Reads:
> > Writes:
> > IOPS: Reads:
> > IOPS: Writes:
> > <http://www.proxmox.com/products/proxmox-ve/subscription-service-plans>
> > ()
> > Degraded data redundancy: 21495/2089170 objects degraded (1.029%), 8 pgs
> > degraded, 8 pgs undersized
> >
> > pg 21.0 is stuck undersized for 63693.346103, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,9]
> > pg 21.2 is stuck undersized for 63693.346973, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,10]
> > pg 21.6f is stuck undersized for 62453.277248, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,5]
> > pg 21.8b is stuck undersized for 63693.361835, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,8]
> > pg 21.c3 is stuck undersized for 63693.321337, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,9]
> > pg 21.c5 is stuck undersized for 66587.797684, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,8]
> > pg 21.d4 is stuck undersized for 62453.047415, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,6]
> > pg 21.e1 is stuck undersized for 62453.276631, current state
> > active+undersized+degraded+remapped+backfill_wait, last acting [2,5]
> >
> >
> >
> >
> > ---
> > Gilberto Nunes Ferreira
> >
> > (47) 3025-5907
> > (47) 99676-7530 - Whatsapp / Telegram
> >
> > Skype: gilberto.nunes36
> >
> >
> >
> >
> > 2018-08-30 10:23 GMT-03:00 Gilberto Nunes <gilberto.nunes32 at gmail.com>:
> >
> >> SO, what you guys think about this HDD distribuiton?
> >>
> >> CEPH-01
> >> 1x 3 TB
> >> 1x 2 TB
> >>
> >> CEPH-02
> >> 1x 4 TB
> >> 1x 3 TB
> >>
> >> CEPH-03
> >> 1x 4 TB
> >> 1x 3 TB
> >>
> >> CEPH-04
> >> 1x 4 TB
> >> 1x 3 TB
> >> 1x 2 TB
> >>
> >> CEPH-05
> >> 1x 8 TB
> >> 1x 2 TB
> >>
> >> CEPH-06
> >> 1x 3 TB
> >> 1x 1 TB
> >> 1x 8 TB
> >>
> >>
> >> ---
> >> Gilberto Nunes Ferreira
> >>
> >> (47) 3025-5907
> >> (47) 99676-7530 - Whatsapp / Telegram
> >>
> >> Skype: gilberto.nunes36
> >>
> >>
>
More information about the pve-user
mailing list