[PVE-User] pveceph createosd after destroyed osd
mark at openvs.co.uk
Tue Jul 3 02:41:39 CEST 2018
Hi, Thanks for your response!
No, I didn't do any of that on the cli - I just did stop in the webgui,
then out, then destroy.
Note that there was no VM's or data at all on this test ceph cluster - I
had deleted it all before doing this. I was basically just removing it all
so the OSD numbers looked "nicer" for the final setup.
It's not a huge deal, I can just reinstall proxmox. But it concerns me that
it seems so fragile using the webgui to do this. I want to know where I
went wrong? Is there somewhere that a signature is being stored so when you
try to add that same drive again (even though I ticked "remove partitions")
it doesn't add back in to the ceph cluster in the next sequential order
from the last current "live" or "valid" drive?
Is it just a rule that you never actually remove drives? you just set them
On 3 July 2018 at 01:34, Woods, Ken A (DNR) <ken.woods at alaska.gov> wrote:
> Are you sure you followed the directions?
> From: pve-user <pve-user-bounces at pve.proxmox.com> on behalf of Mark Adams
> <mark at openvs.co.uk>
> Sent: Monday, July 2, 2018 4:05:51 PM
> To: pve-user at pve.proxmox.com
> Subject: [PVE-User] pveceph createosd after destroyed osd
> Currently running the newest 5.2-1 version, I had a test cluster which was
> working fine. I since added more disks, first stopping, then setting out,
> then destroying each osd so I could recreate it all from scratch.
> However, when adding a new osd (either via GUI or pveceph CLI) it seems to
> show a successful create, however does not show in the gui as an osd under
> the host.
> It's like the osd information is being stored by proxmox/ceph somewhere
> else and not being correctly removed and recreated?
> I can see that the newly created disk (after it being destroyed) is
> Is this by design? is there a way to force the disk back? shouldn't it show
> in the gui once you create it again?
More information about the pve-user