Virtual Machines High Availability

The goal of this section is to provide information to prepare for failures of the Virtual Machines or Hosts and to recover from them. These failures are categorized depending on whether they come from the physical infrastructure (Host failures) or from the virtualized infrastructure (VM crashes). In both scenarios, OpenNebula provides a cost-effective failover solution to minimize downtime from server and OS failures.

Host Failures

When OpenNebula detects that a Host is down, a hook can be triggered to deal with the situation. OpenNebula comes out-of-the-box with a script that can act as a hook to be triggered when a host enters the ERROR state. This can very useful to limit the downtime of a service due to a hardware failure, since it can redeploy the VMs on another host.

To set up this Host hook to be triggered in the ERROR state, you need to create it using the following template and command:

$ cat /usr/share/one/examples/host_hooks/error_hook

ARGUMENTS       = "$TEMPLATE -m -p 5"
ARGUMENTS_STDIN = "yes"
COMMAND         = "ft/host_error.rb"
NAME            = "host_error"
STATE           = "ERROR"
REMOTE          = "no"
RESOURCE        = HOST
TYPE            = state

$ onehook create /usr/share/one/examples/host_hooks/error_hook

We are defining a host hook, named host_error, that will execute the script ft/host_error.rb locally with the following arguments:

Argument

Description

Default

$TEMPLATE

Template of the HOST which triggered the hook. In XML format, base64 encoded.

NA

Action

This defines the action to be performed upon the VMs that were running in the host that went down.

This can be:

  • -m migrate VMs to another host. Only for images on shared storage (NFS, Ceph, …)

    This option skips VMs deployed on local datastores, so they can be started after host recovery.

  • -r delete and recreate VMs running in the host. Their state will be lost!

  • -d delete VMs running in the host.

NA

Force Inactive

-f force resubmission of VMs in SUSPEND and POWEROFF states

False

Avoid Transient

-p <n> avoid resubmission if host comes back after <n> monitoring cycles

2

Avoid Fencing

--no-fencing avoid host fencing

False

More information on hooks here.

Warning

Note that spurious network errors may lead to a VM being started twice on different hosts and possibly clashing on shared resources. The previous script needs to fence the error host to prevent split brain VMs. You may use any fencing mechanism for the host and invoke it within the error hook.

Enabling Fencing

In order to enable fencing you need to implement file /var/lib/one/remotes/hooks/ft/fence_host.sh:

  • Update your hosts using onehost update <HOST_ID> and add there the attribute FENCE_IP with the fencing device IP.

  • Update the above script and add USERNAME and PASSWORD of your fencing device.

  • Remove the line echo ""Fence host not configured, please edit ft/fence_host.sh"" && exit 1 from above script.

  • Depending on your hardware provider, you will need to use a differnt tool to control the ILO, so please check your hardware manual, for example:

while [ "$RETRIES" -gt 0 ]
do
    fence_ilo5 -P --ip=$FENCE_IP --password="${PASSWORD}" --username="${USERNAME}" --action="${ACTION}" && exit 0
    RETRIES=$((RETRIES-1))
    sleep $SLEEP_TIME
done

Continue with Troubleshooting guide to understand how to recover failed VMs.