[pve-devel] [PATCH ha-manager] always queue service stop if node shuts down
Thomas Lamprecht
t.lamprecht at proxmox.com
Fri May 26 15:45:38 CEST 2017
Commit 61ae38eb6fc5ab351fb61f2323776819e20538b7 which ensured that
services get freezed on a node reboot had a side effect where running
services did not get gracefully shutdown on node reboot.
This may lead to data loss as the services then get hard killed, or
they may even prevent a node reboot because a storage cannot get
unmounted as a service still access it.
This commits addresses this issue but does not changes behavior of
the freeze logic for now, but we should evaluate if a freeze makes
really sense here or at least make it configurable.
The changed regression test is a result of the fact that we did not
adapt the correct behavior for the is_node_shutdown command in the
problematic commit. The simulation envrionment returned true
everytime a node shutdown (reboot and poweroff) and the real world
environment just returned true if a poweroff happened but not on a
reboot.
Now the simulation acts the same way as the real environment.
Further I moved the simulation implemenentation to the base class so
that both simulator and regression test system behave the same.
Signed-off-by: Thomas Lamprecht <t.lamprecht at proxmox.com>
---
Important changes happen in LRM and Env/PVE2 the rest is "noise" from the
regression/simulation framework.
Please review as a candidate for PVE 4.4 and PVE 5.
src/PVE/HA/Env/PVE2.pm | 6 ++++--
src/PVE/HA/LRM.pm | 14 ++++++++------
src/PVE/HA/NodeStatus.pm | 6 +++++-
src/PVE/HA/Sim/Env.pm | 20 +++++++++++++++++++-
src/PVE/HA/Sim/TestEnv.pm | 11 -----------
src/test/test-reboot1/log.expect | 12 +++++++-----
6 files changed, 43 insertions(+), 26 deletions(-)
diff --git a/src/PVE/HA/Env/PVE2.pm b/src/PVE/HA/Env/PVE2.pm
index 382b61c..2d8f219 100644
--- a/src/PVE/HA/Env/PVE2.pm
+++ b/src/PVE/HA/Env/PVE2.pm
@@ -85,18 +85,21 @@ sub is_node_shutdown {
my ($self) = @_;
my $shutdown = 0;
+ my $reboot = 0;
my $code = sub {
my $line = shift;
# ensure we match the full unit name by matching /^JOB_ID UNIT /
- $shutdown = 1 if ($line =~ m/^\d+\s+(poweroff|halt)\.target\s+/);
+ # see: man systemd.special
+ $shutdown = 1 if ($line =~ m/^\d+\s+shutdown\.target\s+/);
+ $reboot = 1 if ($line =~ m/^\d+\s+reboot\.target\s+/);
};
my $cmd = ['/bin/systemctl', '--full', 'list-jobs'];
eval { PVE::Tools::run_command($cmd, outfunc => $code, noerr => 1); };
- return $shutdown;
+ return ($shutdown, $reboot);
}
sub queue_crm_commands {
diff --git a/src/PVE/HA/LRM.pm b/src/PVE/HA/LRM.pm
index 26c5c89..86ec608 100644
--- a/src/PVE/HA/LRM.pm
+++ b/src/PVE/HA/LRM.pm
@@ -51,14 +51,12 @@ sub shutdown_request {
my $nodename = $haenv->nodename();
- my $shutdown = $haenv->is_node_shutdown();
+ my ($shutdown, $reboot) = $haenv->is_node_shutdown();
if ($shutdown) {
- $haenv->log('info', "shutdown LRM, stop all services");
- $self->{mode} = 'shutdown';
-
- # queue stop jobs for all services
-
+ # *always* queue stop jobs for all services if the node shuts down,
+ # independent if it's a reboot or a poweroff, else we may corrupt
+ # services or hinder node shutdown
my $ss = $self->{service_status};
foreach my $sid (keys %$ss) {
@@ -68,7 +66,11 @@ sub shutdown_request {
# Note: use undef uid to mark shutdown/stop jobs
$self->queue_resource_command($sid, undef, 'request_stop');
}
+ }
+ if ($shutdown && !$reboot) {
+ $haenv->log('info', "shutdown LRM, stop all services");
+ $self->{mode} = 'shutdown';
} else {
$haenv->log('info', "restart LRM, freeze all services");
$self->{mode} = 'restart';
diff --git a/src/PVE/HA/NodeStatus.pm b/src/PVE/HA/NodeStatus.pm
index b99689c..493298f 100644
--- a/src/PVE/HA/NodeStatus.pm
+++ b/src/PVE/HA/NodeStatus.pm
@@ -221,8 +221,12 @@ sub fence_node {
if ($success) {
my $msg = "fencing: acknowledged - got agent lock for node '$node'";
$haenv->log("info", $msg);
+
+ if ($state eq 'fence') {
+ &$send_fence_state_email($self, 'SUCCEED', $msg, $node);
+ }
+
&$set_node_state($self, $node, 'unknown');
- &$send_fence_state_email($self, 'SUCCEED', $msg, $node);
}
return $success;
diff --git a/src/PVE/HA/Sim/Env.pm b/src/PVE/HA/Sim/Env.pm
index cd1574c..fce688a 100644
--- a/src/PVE/HA/Sim/Env.pm
+++ b/src/PVE/HA/Sim/Env.pm
@@ -158,7 +158,25 @@ sub write_lrm_status {
sub is_node_shutdown {
my ($self) = @_;
- return 0; # default to freezing services if not overwritten by subclass
+ my $node = $self->{nodename};
+ my $cstatus = $self->{hardware}->read_hardware_status_nolock();
+
+ die "undefined node status for node '$node'" if !defined($cstatus->{$node});
+
+ my ($shutdown, $reboot) = (0, 0);
+
+ if (my $target = $cstatus->{$node}->{shutdown}) {
+ if ($target eq 'shutdown') {
+ $shutdown = 1;
+ } elsif ($target eq 'reboot') {
+ $shutdown = 1;
+ $reboot = 1;
+ } else {
+ die "unknown shutdown target '$target'";
+ }
+ }
+
+ return ($shutdown, $reboot);
}
sub read_service_config {
diff --git a/src/PVE/HA/Sim/TestEnv.pm b/src/PVE/HA/Sim/TestEnv.pm
index fe08be3..0e6eced 100644
--- a/src/PVE/HA/Sim/TestEnv.pm
+++ b/src/PVE/HA/Sim/TestEnv.pm
@@ -112,17 +112,6 @@ sub loop_end_hook {
$self->{cur_time} += 1; # easier for simulation
}
-sub is_node_shutdown {
- my ($self) = @_;
-
- my $node = $self->{nodename};
- my $cstatus = $self->{hardware}->read_hardware_status_nolock();
-
- die "undefined node status for node '$node'" if !defined($cstatus->{$node});
-
- return defined($cstatus->{$node}->{shutdown}) ? 1 : 0;
-}
-
# must be 0 as we do not want to fork in the regression tests
sub get_max_workers {
my ($self) = @_;
diff --git a/src/test/test-reboot1/log.expect b/src/test/test-reboot1/log.expect
index 840f56d..e2dda5c 100644
--- a/src/test/test-reboot1/log.expect
+++ b/src/test/test-reboot1/log.expect
@@ -21,7 +21,8 @@ info 25 node3/lrm: status change wait_for_agent_lock => active
info 25 node3/lrm: starting service vm:103
info 25 node3/lrm: service status vm:103 started
info 120 cmdlist: execute reboot node3
-info 120 node3/lrm: shutdown LRM, stop all services
+info 120 node3/lrm: restart LRM, freeze all services
+info 120 node1/crm: service 'vm:103': state changed from 'started' to 'freeze'
info 125 node3/lrm: stopping service vm:103
info 125 node3/lrm: service status vm:103 stopped
info 126 node3/lrm: exit (loop end)
@@ -31,9 +32,10 @@ info 145 reboot: execute power node3 off
info 145 reboot: execute power node3 on
info 145 node3/crm: status change startup => wait_for_quorum
info 140 node3/lrm: status change startup => wait_for_agent_lock
-info 145 node3/lrm: got lock 'ha_agent_node3_lock'
-info 145 node3/lrm: status change wait_for_agent_lock => active
-info 145 node3/lrm: starting service vm:103
-info 145 node3/lrm: service status vm:103 started
+info 160 node1/crm: service 'vm:103': state changed from 'freeze' to 'started'
info 164 node3/crm: status change wait_for_quorum => slave
+info 165 node3/lrm: got lock 'ha_agent_node3_lock'
+info 165 node3/lrm: status change wait_for_agent_lock => active
+info 165 node3/lrm: starting service vm:103
+info 165 node3/lrm: service status vm:103 started
info 720 hardware: exit simulation - done
--
2.11.0
More information about the pve-devel
mailing list