[PVE-User] Ceph Cluster with proxmox failure
Mark Adams
mark at openvs.co.uk
Fri Sep 28 22:23:41 CEST 2018
also, 3 out of 6 servers is not quorum. you need a majority of the total.
On Fri, 28 Sep 2018, 21:22 Mark Adams, <mark at openvs.co.uk> wrote:
> the exact same 3 servers have been down and everything has worked? do you
> run ceph mons on every server?
>
> On Fri, 28 Sep 2018, 21:19 Gilberto Nunes, <gilberto.nunes32 at gmail.com>
> wrote:
>
>> Everything is working couple of hours ago!
>> Due a power failure, 3 hosts are down, but I suppose with 3 host the
>> cluster garantee the quorum and allow ceph continues to work.
>> This already happen before and ceph works perfectly with 3 servers
>>
>> ---
>> Gilberto Nunes Ferreira
>>
>> (47) 3025-5907
>> (47) 99676-7530 - Whatsapp / Telegram
>>
>> Skype: gilberto.nunes36
>>
>>
>>
>>
>>
>> Em sex, 28 de set de 2018 às 17:13, Woods, Ken A (DNR) <
>> ken.woods at alaska.gov>
>> escreveu:
>>
>> > So the times are massively different. Fix that.
>> >
>> > And corosync needs multicast , and that’s tested using omping, not ping.
>> >
>> > Go back through the initial set up documentation and read what is
>> required
>> > for the basic network configuration.
>> >
>> > If corosync and ceph are both not working, start there.
>> >
>> >
>> > > On Sep 28, 2018, at 12:08, Gilberto Nunes <gilberto.nunes32 at gmail.com
>> >
>> > wrote:
>> > >
>> > > pve-ceph01:~# ssh pve-ceph01 date
>> > > Fri Sep 28 17:06:34 -03 2018
>> > > pve-ceph01:~# ssh pve-ceph02 date
>> > > Fri Sep 28 17:06:37 -03 2018
>> > > pve-ceph01:~# ssh pve-ceph05 date
>> > > Fri Sep 28 17:06:39 -03 2018
>> > >
>> > > pve-ceph01:~# ping -c 1 pve-ceph01
>> > > PING pve-ceph01.cepam.com.br (10.10.10.100) 56(84) bytes of data.
>> > > 64 bytes from pve-ceph01.cepam.com.br (10.10.10.100): icmp_seq=1
>> ttl=64
>> > > time=0.020 ms
>> > >
>> > > --- pve-ceph01.cepam.com.br ping statistics ---
>> > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
>> > > rtt min/avg/max/mdev = 0.020/0.020/0.020/0.000 ms
>> > > pve-ceph01:~# ping -c 1 pve-ceph02
>> > > once.
>> > > PING pve-ceph02.cepam.com.br (10.10.10.110) 56(84) bytes of data.
>> > > 64 bytes from pve-ceph02.cepam.com.br (10.10.10.110): icmp_seq=1
>> ttl=64
>> > > time=0.120 ms
>> > >
>> > > --- pve-ceph02.cepam.com.br ping statistics ---
>> > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
>> > > rtt min/avg/max/mdev = 0.120/0.120/0.120/0.000 ms
>> > > pve-ceph01:~# ping -c 1 pve-ceph05
>> > > PING pve-ceph05.cepam.com.br (10.10.10.140) 56(84) bytes of data.
>> > > 64 bytes from pve-ceph05.cepam.com.br (10.10.10.140): icmp_seq=1
>> ttl=64
>> > > time=0.078 ms
>> > >
>> > > --- pve-ceph05.cepam.com.br ping statistics ---
>> > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms
>> > > rtt min/avg/max/mdev = 0.078/0.078/0.078/0.000 ms
>> > >
>> > > I can communicate with other...
>> > > ceph command stuck
>> > > all ceph service it's appears to be running...
>> > >
>> > >
>> > > ---
>> > > Gilberto Nunes Ferreira
>> > >
>> > > (47) 3025-5907
>> > > (47) 99676-7530 - Whatsapp / Telegram
>> > >
>> > > Skype: gilberto.nunes36
>> > >
>> > >
>> > >
>> > >
>> > >
>> > > Em sex, 28 de set de 2018 às 17:02, Woods, Ken A (DNR) <
>> > ken.woods at alaska.gov>
>> > > escreveu:
>> > >
>> > >> Network issues?
>> > >> Time issues?
>> > >>
>> > >>
>> > >>> On Sep 28, 2018, at 11:50, Gilberto Nunes <
>> gilberto.nunes32 at gmail.com>
>> > >> wrote:
>> > >>>
>> > >>> Hi there
>> > >>> I have a 6 server Ceph Cluster maded with proxmox 5.2
>> > >>> Suddenly, after power failure, I have only 3 servers UP, but even
>> with
>> > 3
>> > >>> server, Ceph Cluster doesn't work.
>> > >>> pveceph status give me a timeout
>> > >>> pveceph status got timeout
>> > >>>
>> > >>> Any advice?
>> > >>>
>> > >>>
>> > >>> ---
>> > >>> Gilberto Nunes Ferreira
>> > >>>
>> > >>> (47) 3025-5907
>> > >>> (47) 99676-7530 - Whatsapp / Telegram
>> > >>>
>> > >>> Skype: gilberto.nunes36
>> > >>> _______________________________________________
>> > >>> pve-user mailing list
>> > >>> pve-user at pve.proxmox.com
>> > >>>
>> > >>
>> >
>> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=9BUtASlCjgOxfXJ55imNPY2xKhHLVcP7E8Klu3Ddc0E&s=XjLyjlncigLbXaMeQ-15anc5Wjnneqwz-n1VWZ_zj_E&e=
>> > >> _______________________________________________
>> > >> pve-user mailing list
>> > >> pve-user at pve.proxmox.com
>> > >>
>> >
>> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e=
>> > >>
>> > > _______________________________________________
>> > > pve-user mailing list
>> > > pve-user at pve.proxmox.com
>> > >
>> >
>> https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e=
>> > _______________________________________________
>> > pve-user mailing list
>> > pve-user at pve.proxmox.com
>> > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>> >
>> _______________________________________________
>> pve-user mailing list
>> pve-user at pve.proxmox.com
>> https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>
More information about the pve-user
mailing list