[PVE-User] Ceph Cluster with proxmox failure
Gilberto Nunes
gilberto.nunes32 at gmail.com
Fri Sep 28 22:24:59 CEST 2018
I thing I figure out what happens...
For my lucky one of the servers, pve-ceph05 that remains, has the OSD crush
marked to 0.
My environment consist a mix of different disks sizes....
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
Em sex, 28 de set de 2018 às 17:23, Mark Adams <mark at openvs.co.uk> escreveu:
> the exact same 3 servers have been down and everything has worked? do you
> run ceph mons on every server?
>
> On Fri, 28 Sep 2018, 21:19 Gilberto Nunes, <gilberto.nunes32 at gmail.com>
> wrote:
>
> > Everything is working couple of hours ago!
> > Due a power failure, 3 hosts are down, but I suppose with 3 host the
> > cluster garantee the quorum and allow ceph continues to work.
> > This already happen before and ceph works perfectly with 3 servers
> >
> > ---
> > Gilberto Nunes Ferreira
> >
> > (47) 3025-5907
> > (47) 99676-7530 - Whatsapp / Telegram
> >
> > Skype: gilberto.nunes36
> >
> >
> >
> >
> >
> > Em sex, 28 de set de 2018 às 17:13, Woods, Ken A (DNR) <
> > ken.woods at alaska.gov>
> > escreveu:
> >
> > > So the times are massively different. Fix that.
> > >
> > > And corosync needs multicast , and that’s tested using omping, not
> ping.
> > >
> > > Go back through the initial set up documentation and read what is
> > required
> > > for the basic network configuration.
> > >
> > > If corosync and ceph are both not working, start there.
> > >
> > >
> > > > On Sep 28, 2018, at 12:08, Gilberto Nunes <
> gilberto.nunes32 at gmail.com>
> > > wrote:
> > > >
> > > > pve-ceph01:~# ssh pve-ceph01 date
> > > > Fri Sep 28 17:06:34 -03 2018
> > > > pve-ceph01:~# ssh pve-ceph02 date
> > > > Fri Sep 28 17:06:37 -03 2018
> > > > pve-ceph01:~# ssh pve-ceph05 date
> > > > Fri Sep 28 17:06:39 -03 2018
> > > >
> > > > pve-ceph01:~# ping -c 1 pve-ceph01
> > > > PING pve-ceph01.cepam.com.br (10.10.10.100) 56(84) bytes of data.
> > > > 64 bytes from pve-ceph01.cepam.com.br (10.10.10.100): icmp_seq=1
> > ttl=64
> > > > time=0.020 ms
> > > >
> > > > --- pve-ceph01.cepam.com.br ping statistics ---
> > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
> > > > rtt min/avg/max/mdev = 0.020/0.020/0.020/0.000 ms
> > > > pve-ceph01:~# ping -c 1 pve-ceph02
> > > > once.
> > > > PING pve-ceph02.cepam.com.br (10.10.10.110) 56(84) bytes of data.
> > > > 64 bytes from pve-ceph02.cepam.com.br (10.10.10.110): icmp_seq=1
> > ttl=64
> > > > time=0.120 ms
> > > >
> > > > --- pve-ceph02.cepam.com.br ping statistics ---
> > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
> > > > rtt min/avg/max/mdev = 0.120/0.120/0.120/0.000 ms
> > > > pve-ceph01:~# ping -c 1 pve-ceph05
> > > > PING pve-ceph05.cepam.com.br (10.10.10.140) 56(84) bytes of data.
> > > > 64 bytes from pve-ceph05.cepam.com.br (10.10.10.140): icmp_seq=1
> > ttl=64
> > > > time=0.078 ms
> > > >
> > > > --- pve-ceph05.cepam.com.br ping statistics ---
> > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
> > > > rtt min/avg/max/mdev = 0.078/0.078/0.078/0.000 ms
> > > >
> > > > I can communicate with other...
> > > > ceph command stuck
> > > > all ceph service it's appears to be running...
> > > >
> > > >
> > > > ---
> > > > Gilberto Nunes Ferreira
> > > >
> > > > (47) 3025-5907
> > > > (47) 99676-7530 - Whatsapp / Telegram
> > > >
> > > > Skype: gilberto.nunes36
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > Em sex, 28 de set de 2018 às 17:02, Woods, Ken A (DNR) <
> > > ken.woods at alaska.gov>
> > > > escreveu:
> > > >
> > > >> Network issues?
> > > >> Time issues?
> > > >>
> > > >>
> > > >>> On Sep 28, 2018, at 11:50, Gilberto Nunes <
> > gilberto.nunes32 at gmail.com>
> > > >> wrote:
> > > >>>
> > > >>> Hi there
> > > >>> I have a 6 server Ceph Cluster maded with proxmox 5.2
> > > >>> Suddenly, after power failure, I have only 3 servers UP, but even
> > with
> > > 3
> > > >>> server, Ceph Cluster doesn't work.
> > > >>> pveceph status give me a timeout
> > > >>> pveceph status got timeout
> > > >>>
> > > >>> Any advice?
> > > >>>
> > > >>>
> > > >>> ---
> > > >>> Gilberto Nunes Ferreira
> > > >>>
> > > >>> (47) 3025-5907
> > > >>> (47) 99676-7530 - Whatsapp / Telegram
> > > >>>
> > > >>> Skype: gilberto.nunes36
> > > >>> _______________________________________________
> > > >>> pve-user mailing list
> > > >>> pve-user at pve.proxmox.com
> > > >>>
> > > >>
> > >
> >
> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=9BUtASlCjgOxfXJ55imNPY2xKhHLVcP7E8Klu3Ddc0E&s=XjLyjlncigLbXaMeQ-15anc5Wjnneqwz-n1VWZ_zj_E&e=
> > > >> _______________________________________________
> > > >> pve-user mailing list
> > > >> pve-user at pve.proxmox.com
> > > >>
> > >
> >
> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e=
> > > >>
> > > > _______________________________________________
> > > > pve-user mailing list
> > > > pve-user at pve.proxmox.com
> > > >
> > >
> >
> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e=
> > > _______________________________________________
> > > pve-user mailing list
> > > pve-user at pve.proxmox.com
> > > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> > >
> > _______________________________________________
> > pve-user mailing list
> > pve-user at pve.proxmox.com
> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >
> _______________________________________________
> pve-user mailing list
> pve-user at pve.proxmox.com
> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
More information about the pve-user
mailing list