[PVE-User] Broken Ceph Cluster when adding new one - Proxmox 5.0 & Ceph Luminous
Phil Schwarz
infolist at schwarz-fr.net
Fri Jul 21 23:22:12 CEST 2017
Hi,
after some investigations (and getting the cluster back) , seems that
i've got an issue with pveceph creating OSD:
ceph-disk zap /dev/sdc
pveceph createosd /dev/sdc -bluestore 0 -fstype xfs
Unknown option: bluestore
According to the doc (1) , it should be OK.
pveceph createosd /dev/sdc -fstype xfs
--> Should be using Filestore, that's on purpose.
create OSD on /dev/sdc (xfs)
Caution: invalid backup GPT header, but valid main header; regenerating
backup header from main header.
****************************************************************************
Caution: Found protective or hybrid MBR and corrupt GPT. Using GPT, but disk
verification and recovery are STRONGLY recommended.
****************************************************************************
GPT data structures destroyed! You may now partition the disk using fdisk or
other utilities.
Creating new GPT entries.
The operation has completed successfully.
Setting name!
partNum is 0
REALLY setting name!
The operation has completed successfully.
Setting name!
partNum is 1
REALLY setting name!
The operation has completed successfully.
The operation has completed successfully.
meta-data=/dev/sdc1 isize=2048 agcount=4, agsize=6400 blks
= sectsz=512 attr=2, projid32bit=1
= crc=1 finobt=1, sparse=0,
rmapbt=0, reflink=0
data = bsize=4096 blocks=25600, imaxpct=25
= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0 ftype=1
log =internal log bsize=4096 blocks=864, version=2
= sectsz=512 sunit=0 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
The operation has completed successfully.
But OSD never gets into GUI, neither in crushmap :
ceph osd tree
ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 7.47186 root default
-2 3.17259 host H1
1 1.81360 osd.1 up 1.00000 1.00000
3 1.35899 osd.3 up 1.00000 1.00000
-3 0.67699 host H2
0 0.67699 osd.0 up 1.00000 1.00000
-4 1.80869 host H3
2 0.44969 osd.2 up 1.00000 1.00000
4 1.35899 osd.4 up 1.00000 1.00000
-5 1.81360 host H4
5 1.81360 osd.5 up 1.00000 1.00000
6 0 osd.6 down 0 1.00000
The osd.6 shloud appear under H5.
Thanks
Best regards
(1) : https://pve.proxmox.com/pve-docs/pveceph.1.html
Le 15/07/2017 à 16:02, Phil Schwarz a écrit :
> Hi,
>
> short version :
> I broke my cluster !
>
> Long version , with context:
> With a 4 nodes Proxmox Cluster
> The nodes are all Pproxmox 5.05+Ceph luminous with filestore
> -3 mon+OSD
> -1 LXC+OSD
>
> Was working fine
> Added a fifth node (proxmox+ceph) today a broke everything..
>
> Though every node can ping each other, the web GUI is full of red
> crossed nodes. No LXC is seen though there up and alive.
> However, every other proxmox is manageable through the web GUI....
>
> In logs, i've tons of same message on 2 over 3 mons :
>
> " failed to decode message of type 80 v6: buffer::malformed_input: void
> pg_history_t::decode(ceph::buffer::list::iterator&) unknown encoding
> version > 7"
>
> Thanks for your answers.
> Best regards
>
> While investigating, i wondered about my config :
> Question relative to /etc/hosts file :
> Should i use private_replication_LAN Ip or public ones ?
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
More information about the pve-user
mailing list