use strict;
use warnings;
+
use Digest::MD5 qw(md5_base64);
-use Data::Dumper;
use PVE::Tools;
use PVE::HA::Tools ':exit_codes';
use PVE::HA::NodeStatus;
+use PVE::HA::Usage::Basic;
+use PVE::HA::Usage::Static;
-my $fence_delay = 60;
+## Variable Name & Abbreviations Convention
+#
+# The HA stack has some variables it uses frequently and thus abbreviates it such that it may be
+# confusing for new readers. Here's a short list of the most common used.
+#
+# NOTE: variables should be assumed to be read only if not otherwise stated, only use the specific
+# methods to re-compute/read/alter them.
+#
+# - $haenv -> HA environment, the main interface to the simulator/test/real world
+# - $sid -> Service ID, unique identifier for a service, `type:vmid` is common
+#
+# - $ms -> Master/Manager Status, contains runtime info from the current active manager
+# - $ns -> Node Status, hash holding online/offline status about all nodes
+#
+# - $ss -> Service Status, hash holding the current state (last LRM cmd result, failed starts
+# or migrates, maintenance fallback node, for *all* services ...
+# - $sd -> Service Data, the service status of a *single* service, iow. $ss->{$sid}
+#
+# - $sc -> Service Configuration, hash for all services including target state, group, ...
+# - $cd -> Configuration Data, the service config of a *single* service, iow. $sc->{$sid}
+#
+# Try to avoid adding new two letter (or similar over abbreviated) names, but also don't send
+# patches for changing above, as that set is mostly sensible and should be easy to remember once
+# spending a bit time in the HA code base.
sub new {
my ($this, $haenv) = @_;
my $class = ref($this) || $this;
- my $ms = $haenv->read_manager_status();
+ my $self = bless { haenv => $haenv, crs => {} }, $class;
+
+ my $old_ms = $haenv->read_manager_status();
- $ms->{master_node} = $haenv->nodename();
+ # we only copy the state part of the manager which cannot be auto generated
- my $ns = PVE::HA::NodeStatus->new($haenv, $ms->{node_status} || {});
+ $self->{ns} = PVE::HA::NodeStatus->new($haenv, $old_ms->{node_status} || {});
# fixme: use separate class PVE::HA::ServiceStatus
- my $ss = $ms->{service_status} || {};
+ $self->{ss} = $old_ms->{service_status} || {};
+
+ $self->{ms} = { master_node => $haenv->nodename() };
+
+ # take over node request state to ensure a node in (manual) maintenance mode stays that way
+ # on change of active master.
+ $self->{ms}->{node_request} = $old_ms->{node_request} if defined($old_ms->{node_request});
- my $self = bless {
- haenv => $haenv,
- ms => $ms, # master status
- ns => $ns, # PVE::HA::NodeStatus
- ss => $ss, # service status
- }, $class;
+ $self->update_crs_scheduler_mode(); # initial set, we update it once every loop
return $self;
}
+sub update_crs_scheduler_mode {
+ my ($self) = @_;
+
+ my $haenv = $self->{haenv};
+ my $dc_cfg = $haenv->get_datacenter_settings();
+
+ $self->{crs}->{rebalance_on_request_start} = !!$dc_cfg->{crs}->{'ha-rebalance-on-start'};
+
+ my $old_mode = $self->{crs}->{scheduler};
+ my $new_mode = $dc_cfg->{crs}->{ha} || 'basic';
+
+ if (!defined($old_mode)) {
+ $haenv->log('info', "using scheduler mode '$new_mode'") if $new_mode ne 'basic';
+ } elsif ($new_mode eq $old_mode) {
+ return; # nothing to do
+ } else {
+ $haenv->log('info', "switching scheduler mode from '$old_mode' to '$new_mode'");
+ }
+
+ $self->{crs}->{scheduler} = $new_mode;
+
+ return;
+}
+
sub cleanup {
my ($self) = @_;
$ms->{node_status} = $ns->{status};
$ms->{service_status} = $ss;
$ms->{timestamp} = $haenv->get_time();
-
+
$haenv->write_manager_status($ms);
-}
+}
-sub select_service_node {
- my ($groups, $online_node_usage, $service_conf, $current_node, $try_next) = @_;
+sub get_service_group {
+ my ($groups, $online_node_usage, $service_conf) = @_;
- my $group = { 'nodes' => { $service_conf->{node} => 1 } }; # default group
+ my $group = {};
+ # add all online nodes to default group to allow try_next when no group set
+ $group->{nodes}->{$_} = 1 for $online_node_usage->list_nodes();
- $group = $groups->{ids}->{$service_conf->{group}} if $service_conf->{group} &&
- $groups->{ids}->{$service_conf->{group}};
+ # overwrite default if service is bound to a specific group
+ if (my $group_id = $service_conf->{group}) {
+ $group = $groups->{ids}->{$group_id} if $groups->{ids}->{$group_id};
+ }
+
+ return $group;
+}
+
+# groups available nodes with their priority as group index
+sub get_node_priority_groups {
+ my ($group, $online_node_usage) = @_;
my $pri_groups = {};
my $group_members = {};
if ($entry =~ m/^(\S+):(\d+)$/) {
($node, $pri) = ($1, $2);
}
- next if !defined($online_node_usage->{$node}); # offline
+ next if !$online_node_usage->contains_node($node); # offline
$pri_groups->{$pri}->{$node} = 1;
$group_members->{$node} = $pri;
}
-
# add non-group members to unrestricted groups (priority -1)
if (!$group->{restricted}) {
my $pri = -1;
- foreach my $node (keys %$online_node_usage) {
+ for my $node ($online_node_usage->list_nodes()) {
next if defined($group_members->{$node});
$pri_groups->{$pri}->{$node} = 1;
$group_members->{$node} = -1;
}
}
+ return ($pri_groups, $group_members);
+}
+
+sub select_service_node {
+ my ($groups, $online_node_usage, $sid, $service_conf, $current_node, $try_next, $tried_nodes, $maintenance_fallback, $best_scored) = @_;
+
+ my $group = get_service_group($groups, $online_node_usage, $service_conf);
+
+ my ($pri_groups, $group_members) = get_node_priority_groups($group, $online_node_usage);
my @pri_list = sort {$b <=> $a} keys %$pri_groups;
return undef if !scalar(@pri_list);
-
- if (!$try_next && $group->{nofailback} && defined($group_members->{$current_node})) {
+
+ # stay on current node if possible (avoids random migrations)
+ if ((!$try_next && !$best_scored) && $group->{nofailback} && defined($group_members->{$current_node})) {
return $current_node;
}
my $top_pri = $pri_list[0];
- my @nodes = sort {
- $online_node_usage->{$a} <=> $online_node_usage->{$b} || $a cmp $b
+ # try to avoid nodes where the service failed already if we want to relocate
+ if ($try_next) {
+ foreach my $node (@$tried_nodes) {
+ delete $pri_groups->{$top_pri}->{$node};
+ }
+ }
+
+ return $maintenance_fallback
+ if defined($maintenance_fallback) && $pri_groups->{$top_pri}->{$maintenance_fallback};
+
+ return $current_node if (!$try_next && !$best_scored) && $pri_groups->{$top_pri}->{$current_node};
+
+ my $scores = $online_node_usage->score_nodes_to_start_service($sid, $current_node);
+ my @nodes = sort {
+ $scores->{$a} <=> $scores->{$b} || $a cmp $b
} keys %{$pri_groups->{$top_pri}};
my $found;
my $node = $nodes[$i];
if ($node eq $current_node) {
$found = $i;
- last;
}
}
if ($try_next) {
-
- if (defined($found) && ($found < (scalar(@nodes) - 1))) {
+ if (!$best_scored && defined($found) && ($found < (scalar(@nodes) - 1))) {
return $nodes[$found + 1];
} else {
return $nodes[0];
}
-
} else {
-
- return $nodes[$found] if defined($found);
-
return $nodes[0];
-
}
}
sub compute_new_uuid {
my ($state) = @_;
-
+
$uid_counter++;
return md5_base64($state . $$ . time() . $uid_counter);
}
my $valid_service_states = {
stopped => 1,
request_stop => 1,
+ request_start => 1,
+ request_start_balance => 1,
started => 1,
fence => 1,
+ recovery => 1,
migrate => 1,
relocate => 1,
freeze => 1,
error => 1,
};
+# FIXME with 'static' mode and thousands of services, the overhead can be noticable and the fact
+# that this function is called for each state change and upon recovery doesn't help.
sub recompute_online_node_usage {
my ($self) = @_;
- my $online_node_usage = {};
+ my $haenv = $self->{haenv};
my $online_nodes = $self->{ns}->list_online_nodes();
- foreach my $node (@$online_nodes) {
- $online_node_usage->{$node} = 0;
+ my $online_node_usage;
+
+ if (my $mode = $self->{crs}->{scheduler}) {
+ if ($mode eq 'static') {
+ $online_node_usage = eval {
+ my $scheduler = PVE::HA::Usage::Static->new($haenv);
+ $scheduler->add_node($_) for $online_nodes->@*;
+ return $scheduler;
+ };
+ $haenv->log('warning', "fallback to 'basic' scheduler mode, init for 'static' failed - $@")
+ if $@;
+ } elsif ($mode eq 'basic') {
+ # handled below in the general fall-back case
+ } else {
+ $haenv->log('warning', "got unknown scheduler mode '$mode', using 'basic'");
+ }
+ }
+
+ # fallback to the basic algorithm in any case
+ if (!$online_node_usage) {
+ $online_node_usage = PVE::HA::Usage::Basic->new($haenv);
+ $online_node_usage->add_node($_) for $online_nodes->@*;
}
- foreach my $sid (keys %{$self->{ss}}) {
+ foreach my $sid (sort keys %{$self->{ss}}) {
my $sd = $self->{ss}->{$sid};
my $state = $sd->{state};
- if (defined($online_node_usage->{$sd->{node}})) {
- if (($state eq 'started') || ($state eq 'request_stop') ||
- ($state eq 'fence') || ($state eq 'freeze') || ($state eq 'error')) {
- $online_node_usage->{$sd->{node}}++;
- } elsif (($state eq 'migrate') || ($state eq 'relocate')) {
- $online_node_usage->{$sd->{target}}++;
- } elsif ($state eq 'stopped') {
+ my $target = $sd->{target}; # optional
+ if ($online_node_usage->contains_node($sd->{node})) {
+ if (
+ $state eq 'started' || $state eq 'request_stop' || $state eq 'fence'
+ || $state eq 'freeze' || $state eq 'error' || $state eq 'recovery'
+ ) {
+ $online_node_usage->add_service_usage_to_node($sd->{node}, $sid, $sd->{node});
+ } elsif ($state eq 'migrate' || $state eq 'relocate' || $state eq 'request_start_balance') {
+ my $source = $sd->{node};
+ # count it for both, source and target as load is put on both
+ $online_node_usage->add_service_usage_to_node($source, $sid, $source, $target)
+ if $state ne 'request_start_balance';
+ $online_node_usage->add_service_usage_to_node($target, $sid, $source, $target);
+ } elsif ($state eq 'stopped' || $state eq 'request_start') {
# do nothing
} else {
- die "should not be reached";
+ die "should not be reached (sid = '$sid', state = '$state')";
+ }
+ } elsif (defined($target) && $online_node_usage->contains_node($target)) {
+ if ($state eq 'migrate' || $state eq 'relocate') {
+ # to correctly track maintenance modi and also consider the target as used for the
+ # case a node dies, as we cannot really know if the to-be-aborted incoming migration
+ # has already cleaned up all used resources
+ $online_node_usage->add_service_usage_to_node($target, $sid, $sd->{node}, $target);
}
}
}
my $old_state = $sd->{state};
my $old_node = $sd->{node};
+ my $old_failed_nodes = $sd->{failed_nodes};
+ my $old_maintenance_node = $sd->{maintenance_node};
die "no state change" if $old_state eq $new_state; # just to be sure
$sd->{state} = $new_state;
$sd->{node} = $old_node;
+ $sd->{failed_nodes} = $old_failed_nodes if defined($old_failed_nodes);
+ $sd->{maintenance_node} = $old_maintenance_node if defined($old_maintenance_node);
my $text_state = '';
foreach my $k (sort keys %params) {
$self->recompute_online_node_usage();
$sd->{uid} = compute_new_uuid($new_state);
-
- $text_state = " ($text_state)" if $text_state;
- $haenv->log('info', "service '$sid': state changed from '${old_state}' to '${new_state}' $text_state");
+ $text_state = " ($text_state)" if $text_state;
+ $haenv->log('info', "service '$sid': state changed from '${old_state}' to '${new_state}'$text_state");
+};
+
+# clean up a possible bad state from a recovered service to allow its start
+my $fence_recovery_cleanup = sub {
+ my ($self, $sid, $fenced_node) = @_;
+
+ my $haenv = $self->{haenv};
+
+ my (undef, $type, $id) = $haenv->parse_sid($sid);
+ my $plugin = PVE::HA::Resources->lookup($type);
+
+ # should not happen
+ die "unknown resource type '$type'" if !$plugin;
+
+ # locks may block recovery, cleanup those which are safe to remove after fencing,
+ # i.e., after the original node was reset and thus all it's state
+ my $removable_locks = [
+ 'backup',
+ 'mounted',
+ 'migrate',
+ 'clone',
+ 'rollback',
+ 'snapshot',
+ 'snapshot-delete',
+ 'suspending',
+ 'suspended',
+ ];
+ if (my $removed_lock = $plugin->remove_locks($haenv, $id, $removable_locks, $fenced_node)) {
+ $haenv->log('warning', "removed leftover lock '$removed_lock' from recovered " .
+ "service '$sid' to allow its start.");
+ }
};
-# read LRM status for all nodes
+# read LRM status for all nodes
sub read_lrm_status {
my ($self) = @_;
}
}
-
return ($results, $modes);
}
my ($haenv, $ms, $ns, $ss) = ($self->{haenv}, $self->{ms}, $self->{ns}, $self->{ss});
my $cmdlist = $haenv->read_crm_commands();
-
+
foreach my $cmd (split(/\n/, $cmdlist)) {
chomp $cmd;
if ($cmd =~ m/^(migrate|relocate)\s+(\S+)\s+(\S+)$/) {
- my ($task, $sid, $node) = ($1, $2, $3);
+ my ($task, $sid, $node) = ($1, $2, $3);
if (my $sd = $ss->{$sid}) {
if (!$ns->node_is_online($node)) {
$haenv->log('err', "crm command error - node not online: $cmd");
} else {
if ($node eq $sd->{node}) {
$haenv->log('info', "ignore crm command - service already on target node: $cmd");
- } else {
+ } else {
$haenv->log('info', "got crm command: $cmd");
- $ss->{$sid}->{cmd} = [ $task, $node];
+ $ss->{$sid}->{cmd} = [ $task, $node ];
}
}
} else {
$haenv->log('err', "crm command error - no such service: $cmd");
}
+ } elsif ($cmd =~ m/^stop\s+(\S+)\s+(\S+)$/) {
+ my ($sid, $timeout) = ($1, $2);
+ if (my $sd = $ss->{$sid}) {
+ $haenv->log('info', "got crm command: $cmd");
+ $ss->{$sid}->{cmd} = [ 'stop', $timeout ];
+ } else {
+ $haenv->log('err', "crm command error - no such service: $cmd");
+ }
+ } elsif ($cmd =~ m/^enable-node-maintenance\s+(\S+)$/) {
+ my $node = $1;
+
+ my $state = $ns->get_node_state($node);
+ if ($state eq 'online') {
+ $ms->{node_request}->{$node}->{maintenance} = 1;
+ } elsif ($state eq 'maintenance') {
+ $haenv->log('info', "ignoring crm command - node $node is already in maintenance state");
+ } else {
+ $haenv->log('err', "crm command error - node not online: $cmd");
+ }
+ } elsif ($cmd =~ m/^disable-node-maintenance\s+(\S+)$/) {
+ my $node = $1;
+
+ my $state = $ns->get_node_state($node);
+ if ($state ne 'maintenance') {
+ $haenv->log(
+ 'warn', "clearing maintenance of node $node requested, but it's in state $state");
+ }
+ delete $ms->{node_request}->{$node}->{maintenance}; # gets flushed out at the end of the CRM loop
} else {
$haenv->log('err', "unable to parse crm command: $cmd");
}
my ($haenv, $ms, $ns, $ss) = ($self->{haenv}, $self->{ms}, $self->{ns}, $self->{ss});
- $ns->update($haenv->get_node_info());
+ my ($node_info) = $haenv->get_node_info();
+ my ($lrm_results, $lrm_modes) = $self->read_lrm_status();
+
+ $ns->update($node_info, $lrm_modes);
- if (!$ns->node_is_online($haenv->nodename())) {
+ if (!$ns->node_is_operational($haenv->nodename())) {
$haenv->log('info', "master seems offline");
return;
}
- my ($lrm_results, $lrm_modes) = $self->read_lrm_status();
+ $self->update_crs_scheduler_mode();
my $sc = $haenv->read_service_config();
# add new service
foreach my $sid (sort keys %$sc) {
next if $ss->{$sid}; # already there
- $haenv->log('info', "adding new service '$sid' on node '$sc->{$sid}->{node}'");
+ my $cd = $sc->{$sid};
+ next if $cd->{state} eq 'ignored';
+
+ $haenv->log('info', "adding new service '$sid' on node '$cd->{node}'");
# assume we are running to avoid relocate running service at add
- $ss->{$sid} = { state => 'started', node => $sc->{$sid}->{node},
- uid => compute_new_uuid('started') };
+ my $state = ($cd->{state} eq 'started') ? 'request_start' : 'request_stop';
+ $ss->{$sid} = {
+ state => $state, node => $cd->{node}, uid => compute_new_uuid('started'),
+ };
}
- # remove stale service from manager state
+ # remove stale or ignored services from manager state
foreach my $sid (keys %$ss) {
- next if $sc->{$sid};
- $haenv->log('info', "removing stale service '$sid' (no config)");
+ next if $sc->{$sid} && $sc->{$sid}->{state} ne 'ignored';
+
+ my $reason = defined($sc->{$sid}) ? 'ignored state requested' : 'no config';
+ $haenv->log('info', "removing stale service '$sid' ($reason)");
+
+ # remove all service related state information
delete $ss->{$sid};
}
-
+
$self->update_crm_commands();
for (;;) {
my $repeat = 0;
-
+
$self->recompute_online_node_usage();
- foreach my $sid (keys %$ss) {
+ foreach my $sid (sort keys %$ss) {
my $sd = $ss->{$sid};
my $cd = $sc->{$sid} || { state => 'disabled' };
$self->next_state_started($sid, $cd, $sd, $lrm_res);
- } elsif ($last_state eq 'migrate' || $last_state eq 'relocate') {
+ } elsif ($last_state eq 'request_start') {
+
+ $self->next_state_request_start($sid, $cd, $sd, $lrm_res);
+
+ } elsif ($last_state eq 'migrate' || $last_state eq 'relocate' || $last_state eq 'request_start_balance') {
$self->next_state_migrate_relocate($sid, $cd, $sd, $lrm_res);
# do nothing here - wait until fenced
+ } elsif ($last_state eq 'recovery') {
+
+ $self->next_state_recovery($sid, $cd, $sd, $lrm_res);
+
} elsif ($last_state eq 'request_stop') {
$self->next_state_request_stop($sid, $cd, $sd, $lrm_res);
} elsif ($last_state eq 'freeze') {
my $lrm_mode = $sd->{node} ? $lrm_modes->{$sd->{node}} : undef;
- # unfreeze
- &$change_service_state($self, $sid, 'started')
- if $lrm_mode && $lrm_mode eq 'active';
+ if ($lrm_mode && $lrm_mode eq 'active') { # unfreeze if active again
+ my $state = ($cd->{state} eq 'started') ? 'started' : 'request_stop';
+ $change_service_state->($self, $sid, $state);
+ }
} elsif ($last_state eq 'error') {
my $lrm_mode = $sd->{node} ? $lrm_modes->{$sd->{node}} : undef;
if ($lrm_mode && $lrm_mode eq 'restart') {
- if (($sd->{state} eq 'started' || $sd->{state} eq 'stopped' ||
- $sd->{state} eq 'request_stop')) {
- &$change_service_state($self, $sid, 'freeze');
+ my $state = $sd->{state};
+ if ($state eq 'started' || $state eq 'stopped'|| $state eq 'request_stop') {
+ $change_service_state->($self, $sid, 'freeze');
}
}
# handle fencing
my $fenced_nodes = {};
- foreach my $sid (keys %$ss) {
- my $sd = $ss->{$sid};
- next if $sd->{state} ne 'fence';
+ foreach my $sid (sort keys %$ss) {
+ my ($service_state, $service_node) = $ss->{$sid}->@{'state', 'node'};
+ next if $service_state ne 'fence';
- if (!defined($fenced_nodes->{$sd->{node}})) {
- $fenced_nodes->{$sd->{node}} = $ns->fence_node($sd->{node}) || 0;
+ if (!defined($fenced_nodes->{$service_node})) {
+ $fenced_nodes->{$service_node} = $ns->fence_node($service_node) || 0;
}
- next if !$fenced_nodes->{$sd->{node}};
+ next if !$fenced_nodes->{$service_node};
- # node fence was successful - mark service as stopped
- &$change_service_state($self, $sid, 'stopped');
+ # node fence was successful - recover service
+ $change_service_state->($self, $sid, 'recovery');
+ $repeat = 1; # for faster recovery execution
+ }
+
+ # Avoid that a node without services in 'fence' state (e.g., removed
+ # manually by admin) is stuck with the 'fence' node state.
+ for my $node (sort grep { !defined($fenced_nodes->{$_}) } keys $ns->{status}->%*) {
+ next if $ns->get_node_state($node) ne 'fence';
+
+ $haenv->log('notice', "node '$node' in fence state but no services to-fence! admin interference?!");
+ $repeat = 1 if $ns->fence_node($node);
}
last if !$repeat;
}
}
- if ($ns->node_is_offline_delayed($sd->{node}, $fence_delay)) {
+ if ($ns->node_is_offline_delayed($sd->{node})) {
&$change_service_state($self, $sid, 'fence');
return;
}
# check result from LRM daemon
if ($lrm_res) {
my $exit_code = $lrm_res->{exit_code};
+ my $req_state = $cd->{state} eq 'started' ? 'started' : 'request_stop';
if ($exit_code == SUCCESS) {
- &$change_service_state($self, $sid, 'started', node => $sd->{target});
+ &$change_service_state($self, $sid, $req_state, node => $sd->{target});
return;
+ } elsif ($exit_code == EWRONG_NODE) {
+ $haenv->log('err', "service '$sid' - migration failed: service" .
+ " registered on wrong node!");
+ &$change_service_state($self, $sid, 'error');
+ } elsif ($exit_code == IGNORED) {
+ $haenv->log(
+ "info",
+ "service '$sid' - rebalance-on-start request ignored - service already running",
+ );
+ $change_service_state->($self, $sid, $req_state, node => $sd->{node});
} else {
$haenv->log('err', "service '$sid' - migration failed (exit code $exit_code)");
- &$change_service_state($self, $sid, 'started', node => $sd->{node});
+ &$change_service_state($self, $sid, $req_state, node => $sd->{node});
return;
}
}
- if ($ns->node_is_offline_delayed($sd->{node}, $fence_delay)) {
+ if ($ns->node_is_offline_delayed($sd->{node})) {
&$change_service_state($self, $sid, 'fence');
return;
}
}
-
sub next_state_stopped {
my ($self, $sid, $cd, $sd, $lrm_res) = @_;
if ($sd->{node} ne $cd->{node}) {
# this can happen if we fence a node with active migrations
# hack: modify $sd (normally this should be considered read-only)
- $haenv->log('info', "fixup service '$sid' location ($sd->{node} => $cd->{node}");
- $sd->{node} = $cd->{node};
+ $haenv->log('info', "fixup service '$sid' location ($sd->{node} => $cd->{node})");
+ $sd->{node} = $cd->{node};
}
if ($sd->{cmd}) {
- my ($cmd, $target) = @{$sd->{cmd}};
- delete $sd->{cmd};
+ my $cmd = shift @{$sd->{cmd}};
if ($cmd eq 'migrate' || $cmd eq 'relocate') {
+ my $target = shift @{$sd->{cmd}};
if (!$ns->node_is_online($target)) {
$haenv->log('err', "ignore service '$sid' $cmd request - node '$target' not online");
} elsif ($sd->{node} eq $target) {
$haenv->log('info', "ignore service '$sid' $cmd request - service already on node '$target'");
} else {
- eval {
- $haenv->change_service_location($sid, $sd->{node}, $target);
- $cd->{node} = $sd->{node} = $target; # fixme: $sd is read-only??!!
- $haenv->log('info', "$cmd service '$sid' to node '$target' (stopped)");
- };
- if (my $err = $@) {
- $haenv->log('err', "$cmd service '$sid' to node '$target' failed - $err");
- }
+ &$change_service_state($self, $sid, $cmd, node => $sd->{node}, target => $target);
+ return;
}
+ } elsif ($cmd eq 'stop') {
+ $haenv->log('info', "ignore service '$sid' $cmd request - service already stopped");
} else {
- $haenv->log('err', "unknown command '$cmd' for service '$sid'");
+ $haenv->log('err', "unknown command '$cmd' for service '$sid'");
}
- }
+ delete $sd->{cmd};
+ }
if ($cd->{state} eq 'disabled') {
- # do nothing
+ # NOTE: do nothing here, the stop state is an exception as we do not
+ # process the LRM result here, thus the LRM always tries to stop the
+ # service (protection for the case no CRM is active)
return;
- }
+ }
- if ($cd->{state} eq 'enabled') {
- if (my $node = select_service_node($self->{groups}, $self->{online_node_usage}, $cd, $sd->{node})) {
- if ($node && ($sd->{node} ne $node)) {
- eval {
- $haenv->change_service_location($sid, $sd->{node}, $node);
- $cd->{node} = $sd->{node} = $node; # fixme: $sd is read-only??!!
- };
- if (my $err = $@) {
- $haenv->log('err', "move service '$sid' to node '$node' failed - $err");
- } else {
- &$change_service_state($self, $sid, 'started', node => $node);
- }
- } else {
- &$change_service_state($self, $sid, 'started', node => $node);
- }
- } else {
- # fixme: warn
- }
+ if ($ns->node_is_offline_delayed($sd->{node}) && $ns->get_node_state($sd->{node}) ne 'maintenance') {
+ &$change_service_state($self, $sid, 'fence');
+ return;
+ }
+ if ($cd->{state} eq 'stopped') {
+ # almost the same as 'disabled' state but the service will also get recovered
+ return;
+ }
+
+ if ($cd->{state} eq 'started') {
+ # simply mark it started, if it's on the wrong node next_state_started will fix that for us
+ $change_service_state->($self, $sid, 'request_start', node => $sd->{node});
return;
}
$haenv->log('err', "service '$sid' - unknown state '$cd->{state}' in service configuration");
}
+sub next_state_request_start {
+ my ($self, $sid, $cd, $sd, $lrm_res) = @_;
+
+ my $haenv = $self->{haenv};
+ my $current_node = $sd->{node};
+
+ if ($self->{crs}->{rebalance_on_request_start}) {
+ my $selected_node = select_service_node(
+ $self->{groups},
+ $self->{online_node_usage},
+ $sid,
+ $cd,
+ $sd->{node},
+ 0, # try_next
+ $sd->{failed_nodes},
+ $sd->{maintenance_node},
+ 1, # best_score
+ );
+ my $select_text = $selected_node ne $current_node ? 'new' : 'current';
+ $haenv->log('info', "service $sid: re-balance selected $select_text node $selected_node for startup");
+
+ if ($selected_node ne $current_node) {
+ $change_service_state->($self, $sid, 'request_start_balance', node => $current_node, target => $selected_node);
+ return;
+ }
+ }
+
+ $change_service_state->($self, $sid, 'started', node => $current_node);
+}
+
+sub record_service_failed_on_node {
+ my ($self, $sid, $node) = @_;
+
+ if (!defined($self->{ss}->{$sid}->{failed_nodes})) {
+ $self->{ss}->{$sid}->{failed_nodes} = [];
+ }
+
+ push @{$self->{ss}->{$sid}->{failed_nodes}}, $node;
+}
+
sub next_state_started {
my ($self, $sid, $cd, $sd, $lrm_res) = @_;
my $ns = $self->{ns};
if (!$ns->node_is_online($sd->{node})) {
- if ($ns->node_is_offline_delayed($sd->{node}, $fence_delay)) {
+ if ($ns->node_is_offline_delayed($sd->{node})) {
&$change_service_state($self, $sid, 'fence');
}
- return;
+ if ($ns->get_node_state($sd->{node}) ne 'maintenance') {
+ return;
+ } else {
+ # save current node as fallback for when it comes out of maintenance
+ $sd->{maintenance_node} = $sd->{node};
+ }
}
-
- if ($cd->{state} eq 'disabled') {
+
+ if ($cd->{state} eq 'disabled' || $cd->{state} eq 'stopped') {
&$change_service_state($self, $sid, 'request_stop');
return;
}
- if ($cd->{state} eq 'enabled') {
+ if ($cd->{state} eq 'started') {
if ($sd->{cmd}) {
- my ($cmd, $target) = @{$sd->{cmd}};
- delete $sd->{cmd};
+ my $cmd = shift @{$sd->{cmd}};
if ($cmd eq 'migrate' || $cmd eq 'relocate') {
+ my $target = shift @{$sd->{cmd}};
if (!$ns->node_is_online($target)) {
$haenv->log('err', "ignore service '$sid' $cmd request - node '$target' not online");
} elsif ($sd->{node} eq $target) {
$haenv->log('info', "ignore service '$sid' $cmd request - service already on node '$target'");
} else {
- $haenv->log('info', "$cmd service '$sid' to node '$target' (running)");
+ $haenv->log('info', "$cmd service '$sid' to node '$target'");
&$change_service_state($self, $sid, $cmd, node => $sd->{node}, target => $target);
}
+ } elsif ($cmd eq 'stop') {
+ my $timeout = shift @{$sd->{cmd}};
+ if ($timeout == 0) {
+ $haenv->log('info', "request immediate service hard-stop for service '$sid'");
+ } else {
+ $haenv->log('info', "request graceful stop with timeout '$timeout' for service '$sid'");
+ }
+ &$change_service_state($self, $sid, 'request_stop', timeout => $timeout);
+ $haenv->update_service_config($sid, {'state' => 'stopped'});
} else {
- $haenv->log('err', "unknown command '$cmd' for service '$sid'");
+ $haenv->log('err', "unknown command '$cmd' for service '$sid'");
}
+
+ delete $sd->{cmd};
+
} else {
my $try_next = 0;
+
if ($lrm_res) {
- if ($lrm_res->{exit_code} == ERROR) {
- my $try = $master_status->{relocate_trial}->{$sid} || 0;
+ my $ec = $lrm_res->{exit_code};
+ if ($ec == SUCCESS) {
+
+ if (defined($sd->{failed_nodes})) {
+ $haenv->log('info', "relocation policy successful for '$sid' on node '$sd->{node}'," .
+ " failed nodes: " . join(', ', @{$sd->{failed_nodes}}) );
+ }
+
+ delete $sd->{failed_nodes};
+
+ # store flag to indicate successful start - only valid while state == 'started'
+ $sd->{running} = 1;
+
+ } elsif ($ec == ERROR || $ec == EWRONG_NODE) {
- if ($try < $cd->{max_relocate}) {
+ delete $sd->{running};
- $try++;
- $try_next = 1; # tell select_service_node to relocate
+ # apply our relocate policy if we got ERROR from the LRM
+ $self->record_service_failed_on_node($sid, $sd->{node});
+
+ if (scalar(@{$sd->{failed_nodes}}) <= $cd->{max_relocate}) {
+
+ # tell select_service_node to relocate if possible
+ $try_next = 1;
$haenv->log('warning', "starting service $sid on node".
" '$sd->{node}' failed, relocating service.");
- $master_status->{relocate_trial}->{$sid} = $try;
} else {
- $haenv->log('err', "recovery policy for service".
- " $sid failed, entering error state!");
+ $haenv->log('err', "recovery policy for service $sid " .
+ "failed, entering error state. Failed nodes: ".
+ join(', ', @{$sd->{failed_nodes}}));
&$change_service_state($self, $sid, 'error');
return;
}
- } elsif ($lrm_res->{exit_code} == SUCCESS) {
- $master_status->{relocate_trial}->{$sid} = 0;
+ } else {
+ $self->record_service_failed_on_node($sid, $sd->{node});
+
+ $haenv->log('err', "service '$sid' got unrecoverable error (exit code $ec))");
+ # we have no save way out (yet) for other errors
+ &$change_service_state($self, $sid, 'error');
+ return;
}
}
- my $node = select_service_node($self->{groups}, $self->{online_node_usage},
- $cd, $sd->{node}, $try_next);
+ my $node = select_service_node(
+ $self->{groups},
+ $self->{online_node_usage},
+ $sid,
+ $cd,
+ $sd->{node},
+ $try_next,
+ $sd->{failed_nodes},
+ $sd->{maintenance_node},
+ );
if ($node && ($sd->{node} ne $node)) {
- $haenv->log('info', "migrate service '$sid' to node '$node' (running)");
- &$change_service_state($self, $sid, 'migrate', node => $sd->{node}, target => $node);
+ $self->{online_node_usage}->add_service_usage_to_node($node, $sid, $sd->{node});
+
+ if (defined(my $fallback = $sd->{maintenance_node})) {
+ if ($node eq $fallback) {
+ $haenv->log(
+ 'info',
+ "moving service '$sid' back to '$fallback', node came back from maintenance.",
+ );
+ delete $sd->{maintenance_node};
+ } elsif ($sd->{node} ne $fallback) {
+ $haenv->log('info', "dropping maintenance fallback node '$fallback' for '$sid'");
+ delete $sd->{maintenance_node};
+ }
+ }
+
+ if ($cd->{type} eq 'vm') {
+ $haenv->log('info', "migrate service '$sid' to node '$node' (running)");
+ &$change_service_state($self, $sid, 'migrate', node => $sd->{node}, target => $node);
+ } else {
+ $haenv->log('info', "relocate service '$sid' to node '$node'");
+ &$change_service_state($self, $sid, 'relocate', node => $sd->{node}, target => $node);
+ }
} else {
- # do nothing
+ if ($try_next && !defined($node)) {
+ $haenv->log(
+ 'warning',
+ "Start Error Recovery: Tried all available nodes for service '$sid', retry"
+ ." start on current node. Tried nodes: " . join(', ', @{$sd->{failed_nodes}},
+ )
+ );
+ }
+
+ if ($sd->{maintenance_node} && $sd->{node} eq $sd->{maintenance_node}) {
+ my $node_state = $ns->get_node_state($sd->{node});
+ if ($node_state eq 'online') {
+ # Having the maintenance node set here means that the service was never
+ # started on a different node since it was set. This can happen in the edge
+ # case that the whole cluster is shut down at the same time while the
+ # 'migrate' policy was configured. Node is not in maintenance mode anymore
+ # and service is started on this node, so it's fine to clear the setting.
+ $haenv->log(
+ 'info',
+ "service '$sid': clearing stale maintenance node "
+ ."'$sd->{maintenance_node}' setting (is current node)",
+ );
+ delete $sd->{maintenance_node};
+ }
+ }
+
+ # ensure service get started again if it went unexpected down
+ # but ensure also no LRM result gets lost
+ $sd->{uid} = compute_new_uuid($sd->{state}) if defined($lrm_res);
}
}
return;
- }
+ }
$haenv->log('err', "service '$sid' - unknown state '$cd->{state}' in service configuration");
}
my ($self, $sid, $cd, $sd, $lrm_res) = @_;
my $ns = $self->{ns};
+ my $ms = $self->{ms};
if ($cd->{state} eq 'disabled') {
+ # clean up on error recovery
+ delete $sd->{failed_nodes};
+
&$change_service_state($self, $sid, 'stopped');
return;
}
- if ($ns->node_is_offline_delayed($sd->{node}, $fence_delay)) {
- &$change_service_state($self, $sid, 'fence');
+}
+
+# after a node was fenced this recovers the service to a new node
+sub next_state_recovery {
+ my ($self, $sid, $cd, $sd, $lrm_res) = @_;
+
+ my ($haenv, $ss) = ($self->{haenv}, $self->{ss});
+ my $ns = $self->{ns};
+ my $ms = $self->{ms};
+
+ if ($sd->{state} ne 'recovery') { # should not happen
+ $haenv->log('err', "cannot recover service '$sid' from fencing, wrong state '$sd->{state}'");
return;
}
+ my $fenced_node = $sd->{node}; # for logging purpose
+
+ $self->recompute_online_node_usage(); # we want the most current node state
+
+ my $recovery_node = select_service_node(
+ $self->{groups},
+ $self->{online_node_usage},
+ $sid,
+ $cd,
+ $sd->{node},
+ );
+
+ if ($recovery_node) {
+ my $msg = "recover service '$sid' from fenced node '$fenced_node' to node '$recovery_node'";
+ if ($recovery_node eq $fenced_node) {
+ # can happen if restriced groups and the node came up again OK
+ $msg = "recover service '$sid' to previous failed and fenced node '$fenced_node' again";
+ }
+ $haenv->log('info', "$msg");
+
+ $fence_recovery_cleanup->($self, $sid, $fenced_node);
+
+ $haenv->steal_service($sid, $sd->{node}, $recovery_node);
+ $self->{online_node_usage}->add_service_usage_to_node($recovery_node, $sid, $recovery_node);
+
+ # NOTE: $sd *is normally read-only*, fencing is the exception
+ $cd->{node} = $sd->{node} = $recovery_node;
+ my $new_state = ($cd->{state} eq 'started') ? 'started' : 'request_stop';
+ $change_service_state->($self, $sid, $new_state, node => $recovery_node);
+ } else {
+ # no possible node found, cannot recover - but retry later, as we always try to make it available
+ $haenv->log('err', "recovering service '$sid' from fenced node '$fenced_node' failed, no recovery node found");
+
+ if ($cd->{state} eq 'disabled') {
+ # allow getting a service out of recovery manually if an admin disables it.
+ delete $sd->{failed_nodes}; # clean up on recovery to stopped
+ $change_service_state->($self, $sid, 'stopped'); # must NOT go through request_stop
+ return;
+ }
+ }
}
1;