[pve-devel] corosync bug: cluster break after 1 node clean shutdown
Alexandre DERUMIER
aderumier at odiso.com
Sun Sep 6 10:43:36 CEST 2020
Maybe something interesting, the only survived node was node7, and it was the crm master
I'm also seein crm disabling watchdog, and also some "loop take too long" messages
(some migration logs from node2 to node1 before the maintenance)
Sep 3 10:36:29 m6kvm7 pve-ha-crm[16196]: service 'vm:992': state changed from 'migrate' to 'started' (node = m6kvm1)
Sep 3 10:36:29 m6kvm7 pve-ha-crm[16196]: service 'vm:993': state changed from 'migrate' to 'started' (node = m6kvm1)
Sep 3 10:36:29 m6kvm7 pve-ha-crm[16196]: service 'vm:997': state changed from 'migrate' to 'started' (node = m6kvm1)
....
Sep 3 10:40:41 m6kvm7 pve-ha-crm[16196]: node 'm6kvm2': state changed from 'online' => 'unknown'
Sep 3 10:40:50 m6kvm7 pve-ha-crm[16196]: got unexpected error - error during cfs-locked 'domain-ha' operation: no quorum!
Sep 3 10:40:51 m6kvm7 pve-ha-lrm[16140]: loop take too long (87 seconds)
Sep 3 10:40:51 m6kvm7 pve-ha-crm[16196]: loop take too long (92 seconds)
Sep 3 10:40:51 m6kvm7 pve-ha-crm[16196]: lost lock 'ha_manager_lock - cfs lock update failed - Permission denied
Sep 3 10:40:51 m6kvm7 pve-ha-lrm[16140]: lost lock 'ha_agent_m6kvm7_lock - cfs lock update failed - Permission denied
Sep 3 10:40:56 m6kvm7 pve-ha-lrm[16140]: status change active => lost_agent_lock
Sep 3 10:40:56 m6kvm7 pve-ha-crm[16196]: status change master => lost_manager_lock
Sep 3 10:40:56 m6kvm7 pve-ha-crm[16196]: watchdog closed (disabled)
Sep 3 10:40:56 m6kvm7 pve-ha-crm[16196]: status change lost_manager_lock => wait_for_quorum
others nodes timing
--------------------
10:39:16 -> node2 shutdown, leave coroync
10:40:25 -> other nodes rebooted by watchdog
----- Mail original -----
De: "aderumier" <aderumier at odiso.com>
À: "dietmar" <dietmar at proxmox.com>
Cc: "Proxmox VE development discussion" <pve-devel at lists.proxmox.com>, "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Dimanche 6 Septembre 2020 07:36:10
Objet: Re: [pve-devel] corosync bug: cluster break after 1 node clean shutdown
>>But the pve logs look ok, and there is no indication
>>that we stopped updating the watchdog. So why did the
>>watchdog trigger? Maybe an IPMI bug?
do you mean an ipmi bug on all 13 servers at the same time ?
(I also have 2 supermicro servers in this cluster, but they use same ipmi watchdog driver. (ipmi_watchdog)
I had same kind of with bug once (when stopping a server), on another cluster, 6 months ago.
This was without HA, but different version of corosync, and that time, I was really seeing quorum split in the corosync logs of the servers.
I'll try to reproduce with a virtual cluster with 14 nodes (don't have enough hardware)
Could I be a bug in proxmox HA code, where watchdog is not resetted by LRM anymore?
----- Mail original -----
De: "dietmar" <dietmar at proxmox.com>
À: "aderumier" <aderumier at odiso.com>
Cc: "Proxmox VE development discussion" <pve-devel at lists.proxmox.com>, "pve-devel" <pve-devel at pve.proxmox.com>
Envoyé: Dimanche 6 Septembre 2020 06:21:55
Objet: Re: [pve-devel] corosync bug: cluster break after 1 node clean shutdown
> >>So you are using ipmi hardware watchdog?
>
> yes, I'm using dell idrac ipmi card watchdog
But the pve logs look ok, and there is no indication
that we stopped updating the watchdog. So why did the
watchdog trigger? Maybe an IPMI bug?
_______________________________________________
pve-devel mailing list
pve-devel at lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel
More information about the pve-devel
mailing list