[pve-devel] [PATCH cluster v2 0/8] initial API adaption to corosync 3/kronosnet

Thomas Lamprecht t.lamprecht at proxmox.com
Fri Jun 14 17:50:53 CEST 2019


On 6/14/19 3:03 PM, Fabian Grünbichler wrote:
> ISSUE 1
> 
> creating a cluster with
> 
>  pvesh create /cluster/config -clustername thomastest -link0 192.168.21.71 -link5 10.0.0.71 
> 
> creates an invalid corosync.conf:
> 
> Jun 14 14:23:50 clustertest71 systemd[1]: Starting Corosync Cluster Engine...
> Jun 14 14:23:50 clustertest71 corosync[2160]:   [MAIN  ] Corosync Cluster Engine 3.0.1-dirty starting up
> Jun 14 14:23:50 clustertest71 corosync[2160]:   [MAIN  ] Corosync built-in features: dbus monitoring watchdog systemd xmlconf snmp pie relro bindnow
> Jun 14 14:23:50 clustertest71 corosync[2160]:   [MAIN  ] parse error in config: Not all nodes have the same number of links
> Jun 14 14:23:50 clustertest71 corosync[2160]:   [MAIN  ] Corosync Cluster Engine exiting with status 8 at main.c:1386.
> Jun 14 14:23:50 clustertest71 systemd[1]: corosync.service: Main process exited, code=exited, status=8/n/a
> Jun 14 14:23:50 clustertest71 systemd[1]: corosync.service: Failed with result 'exit-code'.
> Jun 14 14:23:50 clustertest71 systemd[1]: Failed to start Corosync Cluster Engine.
> 
> $ cat /etc/corosync/corosync.conf
> 
> logging {
>   debug: off
>   to_syslog: yes
> }
> 
> nodelist {
>   node {
>     name: clustertest71
>     nodeid: 1
>     quorum_votes: 1
>     ring0_addr: 192.168.21.71
>     ring5_addr: 10.0.0.71
>   }
> }
> 
> quorum {
>   provider: corosync_votequorum
> }
> 
> totem {
>   cluster_name: thomastest
>   config_version: 1
>   interface {
>     linknumber: 0
>   }
>   interface {
>     linknumber: 5
>   }
>   ip_version: ipv4-6
>   link_mode: passive
>   secauth: on
>   version: 2
> }
> 
> doing the same with link0 and link1 instead of link0 and link5 works.
> subsequently changing corosync.conf to have link0 and linkX with X != 1
> also works, although the reload complains with the same error message
> (cmap and corosync-cfgtool show the updated status just fine).
> restarting corosync fails, again with the status shown above.
> 
> haven't checked yet whether that is an issue on our side or corosync,
> but probably worth an investigation ;)

this is a "bug" of corosync..

the following check in exec/totemconfig fails:

for (i=0; i<num_configured; i++) {
	if (totem_config->interfaces[i].member_count != members) err...
}

here, num_configured is the correct number of configured interfaces
(2), the struct entry member_count is 1 (one node, which seems OK here
too) but members is 0...

members is set a bit above with:
members = totem_config->interfaces[0].member_count;


but totem_config->interfaces gets dynamically allocated with:
totem_config->interfaces = malloc (sizeof (struct totem_interface) * INTERFACE_MAX);

So it's not the configured interfaces (0 being the lowest one
configured, 1 the next, ...) but the _actual_ links from 0 to
INTERFACE_MAX - 1 (== 7)

So here it _always_ gets the membercount from link0, if that is
non-existent in the config then it's the default 0...

So either, link0 isn't as optional as you meant/wished or they have
at least one, and probably a few more, bugs where they falsely assume
that interfaces[0] is the first configured not link0...





More information about the pve-devel mailing list