mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 08:42:29 +01:00
Host-HA offers investigation, fencing and recovery mechanisms for host that for any reason are malfunctioning. It uses Activity and Health checks to determine current host state based on which it may degrade a host or try to recover it. On failing to recover it, it may try to fence the host. The core feature is implemented in a hypervisor agnostic way, with two separate implementations of the driver/provider for Simulator and KVM hypervisors. The framework also allows for implementation of other hypervisor specific provider implementation in future. The Host-HA provider implementation for KVM hypervisor uses the out-of-band management sub-system to issue IPMI calls to reset (recover) or poweroff (fence) a host. The Host-HA provider implementation for Simulator provides a means of testing and validating the core framework implementation. Signed-off-by: Abhinandan Prateek <abhinandan.prateek@shapeblue.com> Signed-off-by: Rohit Yadav <rohit.yadav@shapeblue.com>
602 lines
22 KiB
Python
602 lines
22 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
from __builtin__ import False
|
|
""" BVT tests for Hosts Maintenance
|
|
"""
|
|
|
|
# Import Local Modules
|
|
from marvin.codes import FAILED
|
|
from marvin.cloudstackTestCase import *
|
|
from marvin.cloudstackAPI import *
|
|
from marvin.lib.utils import *
|
|
from marvin.lib.base import *
|
|
from marvin.lib.common import *
|
|
from nose.plugins.attrib import attr
|
|
|
|
from time import sleep
|
|
|
|
_multiprocess_shared_ = False
|
|
|
|
|
|
class TestHostHA(cloudstackTestCase):
|
|
|
|
def setUp(self):
|
|
self.logger = logging.getLogger('TestHM')
|
|
self.stream_handler = logging.StreamHandler()
|
|
self.logger.setLevel(logging.DEBUG)
|
|
self.logger.addHandler(self.stream_handler)
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.hypervisor = self.testClient.getHypervisorInfo()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.services = self.testClient.getParsedTestDataConfig()
|
|
self.zone = get_zone(self.apiclient, self.testClient.getZoneForTests())
|
|
self.pod = get_pod(self.apiclient, self.zone.id)
|
|
self.cleanup = []
|
|
self.services = {
|
|
"service_offering": {
|
|
"name": "Ultra Tiny Instance",
|
|
"displaytext": "Ultra Tiny Instance",
|
|
"cpunumber": 1,
|
|
"cpuspeed": 100,
|
|
"memory": 128,
|
|
},
|
|
"service_offering_local": {
|
|
"name": "Ultra Tiny Local Instance",
|
|
"displaytext": "Ultra Tiny Local Instance",
|
|
"cpunumber": 1,
|
|
"cpuspeed": 100,
|
|
"memory": 128,
|
|
"storagetype": "local"
|
|
},
|
|
"vm": {
|
|
"username": "root",
|
|
"password": "password",
|
|
"ssh_port": 22,
|
|
# Hypervisor type should be same as
|
|
# hypervisor type of cluster
|
|
"privateport": 22,
|
|
"publicport": 22,
|
|
"protocol": 'TCP',
|
|
},
|
|
"natrule": {
|
|
"privateport": 22,
|
|
"publicport": 22,
|
|
"startport": 22,
|
|
"endport": 22,
|
|
"protocol": "TCP",
|
|
"cidrlist": '0.0.0.0/0',
|
|
},
|
|
"ostype": 'CentOS 5.3 (64-bit)',
|
|
"sleep": 60,
|
|
"timeout": 10,
|
|
}
|
|
|
|
|
|
def tearDown(self):
|
|
try:
|
|
# Clean up, terminate the created templates
|
|
cleanup_resources(self.apiclient, self.cleanup)
|
|
|
|
except Exception as e:
|
|
raise Exception("Warning: Exception during cleanup : %s" % e)
|
|
|
|
return
|
|
|
|
|
|
def createVMs(self, hostId, number, local):
|
|
self.template = get_template(
|
|
self.apiclient,
|
|
self.zone.id,
|
|
self.services["ostype"]
|
|
)
|
|
|
|
if self.template == FAILED:
|
|
assert False, "get_template() failed to return template with description %s" % self.services["ostype"]
|
|
|
|
self.logger.debug("Using template %s " % self.template.id)
|
|
|
|
if local:
|
|
self.service_offering = ServiceOffering.create(
|
|
self.apiclient,
|
|
self.services["service_offering_local"]
|
|
)
|
|
else:
|
|
self.service_offering = ServiceOffering.create(
|
|
self.apiclient,
|
|
self.services["service_offering"]
|
|
)
|
|
|
|
self.logger.debug("Using service offering %s " % self.service_offering.id)
|
|
|
|
vms = []
|
|
for i in range(0, number):
|
|
self.services["vm"]["zoneid"] = self.zone.id
|
|
self.services["vm"]["template"] = self.template.id
|
|
self.services["vm"]["displayname"] = 'vm' + str(i)
|
|
self.services["vm"]["hypervisor"] = self.hypervisor
|
|
vm = VirtualMachine.create(
|
|
self.apiclient,
|
|
self.services["vm"],
|
|
serviceofferingid=self.service_offering.id,
|
|
hostid=hostId
|
|
)
|
|
vms.append(vm)
|
|
self.cleanup.append(vm)
|
|
self.logger.debug("VM create = {}".format(vm.id))
|
|
return vm
|
|
|
|
def noOfVMsOnHost(self, hostId):
|
|
listVms = VirtualMachine.list(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
vmnos = 0
|
|
if (listVms is not None):
|
|
for vm in listVms:
|
|
self.logger.debug('VirtualMachine on Hyp 1 = {}'.format(vm.id))
|
|
vmnos = vmnos + 1
|
|
|
|
return vmnos
|
|
|
|
def checkHostDown(self, fromHostIp, testHostIp):
|
|
try:
|
|
ssh = SshClient(fromHostIp, 22, "root", "password")
|
|
res = ssh.execute("ping -c 1 %s" % testHostIp)
|
|
result = str(res)
|
|
if result.count("100% packet loss") == 1:
|
|
return True, 1
|
|
else:
|
|
return False, 1
|
|
except Exception as e:
|
|
self.logger.debug("Got exception %s" % e)
|
|
return False, 1
|
|
|
|
def checkHostUp(self, fromHostIp, testHostIp):
|
|
try:
|
|
ssh = SshClient(fromHostIp, 22, "root", "password")
|
|
res = ssh.execute("ping -c 1 %s" % testHostIp)
|
|
result = str(res)
|
|
if result.count(" 0% packet loss") == 1:
|
|
return True, 1
|
|
else:
|
|
return False, 1
|
|
except Exception as e:
|
|
self.logger.debug("Got exception %s" % e)
|
|
return False, 1
|
|
|
|
|
|
def isOnlyNFSStorageAvailable(self):
|
|
if self.zone.localstorageenabled:
|
|
return False
|
|
storage_pools = StoragePool.list(
|
|
self.apiclient,
|
|
zoneid=self.zone.id,
|
|
listall=True
|
|
)
|
|
self.assertEqual(
|
|
isinstance(storage_pools, list),
|
|
True,
|
|
"Check if listStoragePools returns a valid response"
|
|
)
|
|
for storage_pool in storage_pools:
|
|
if storage_pool.type == u'NetworkFilesystem':
|
|
return True
|
|
|
|
return False
|
|
|
|
def isOnlyLocalStorageAvailable(self):
|
|
if not(self.zone.localstorageenabled):
|
|
return False
|
|
|
|
storage_pools = StoragePool.list(
|
|
self.apiclient,
|
|
zoneid=self.zone.id,
|
|
listall=True
|
|
)
|
|
self.assertEqual(
|
|
isinstance(storage_pools, list),
|
|
True,
|
|
"Check if listStoragePools returns a valid response"
|
|
)
|
|
for storage_pool in storage_pools:
|
|
if storage_pool.type == u'NetworkFilesystem':
|
|
return False
|
|
|
|
return True
|
|
|
|
def isLocalAndNFSStorageAvailable(self):
|
|
if not(self.zone.localstorageenabled):
|
|
return False
|
|
|
|
storage_pools = StoragePool.list(
|
|
self.apiclient,
|
|
zoneid=self.zone.id,
|
|
listall=True
|
|
)
|
|
self.assertEqual(
|
|
isinstance(storage_pools, list),
|
|
True,
|
|
"Check if listStoragePools returns a valid response"
|
|
)
|
|
for storage_pool in storage_pools:
|
|
if storage_pool.type == u'NetworkFilesystem':
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
def checkHostStateInCloudstack(self, state, hostId):
|
|
try:
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
id=hostId
|
|
)
|
|
self.assertEqual(
|
|
isinstance(listHost, list),
|
|
True,
|
|
"Check if listHost returns a valid response"
|
|
)
|
|
|
|
self.assertEqual(
|
|
len(listHost),
|
|
1,
|
|
"Check if listHost returns a host"
|
|
)
|
|
self.logger.debug(" Host state is %s " % listHost[0].state)
|
|
if listHost[0].state == state:
|
|
return True, 1
|
|
else:
|
|
return False, 1
|
|
except Exception as e:
|
|
self.logger.debug("Got exception %s" % e)
|
|
return False, 1
|
|
|
|
|
|
def disconnectHostfromNetwork(self, hostIp, timeout):
|
|
srcFile = os.path.dirname(os.path.realpath(__file__)) + "/test_host_ha.sh"
|
|
if not(os.path.isfile(srcFile)):
|
|
self.logger.debug("File %s not found" % srcFile)
|
|
raise unittest.SkipTest("Script file %s required for HA not found" % srcFile);
|
|
|
|
ssh = SshClient(hostIp, 22, "root", "password")
|
|
ssh.scp(srcFile, "/root/test_host_ha.sh")
|
|
ssh.execute("nohup sh /root/test_host_ha.sh -t %s -d all > /dev/null 2>&1 &\n" % timeout)
|
|
return
|
|
|
|
def stopAgentOnHost(self, hostIp, timeout):
|
|
srcFile = os.path.dirname(os.path.realpath(__file__)) + "/test_host_ha.sh"
|
|
if not(os.path.isfile(srcFile)):
|
|
self.logger.debug("File %s not found" % srcFile)
|
|
raise unittest.SkipTest("Script file %s required for HA not found" % srcFile);
|
|
|
|
ssh = SshClient(hostIp, 22, "root", "password")
|
|
ssh.scp(srcFile, "/root/test_host_ha.sh")
|
|
ssh.execute("nohup sh /root/test_host_ha.sh -t %s -d agent > /dev/null 2>&1 &\n" % timeout)
|
|
return
|
|
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_01_host_ha_with_nfs_storagepool_with_vm(self):
|
|
raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
|
|
return
|
|
|
|
if not(self.isOnlyNFSStorageAvailable()):
|
|
raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
|
|
return
|
|
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Hypervisor = {}'.format(host.id))
|
|
|
|
|
|
if len(listHost) != 2:
|
|
self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
return
|
|
|
|
no_of_vms = self.noOfVMsOnHost(listHost[0].id)
|
|
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
|
|
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
|
|
|
|
if no_of_vms < 5:
|
|
self.logger.debug("test_01: Create VMs as there are not enough vms to check host ha")
|
|
no_vm_req = 5 - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req, False)
|
|
|
|
ha_host = listHost[1]
|
|
other_host = listHost[0]
|
|
if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
|
|
ha_host = listHost[0]
|
|
other_host = listHost[1]
|
|
|
|
self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
|
|
|
|
hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostDown):
|
|
raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
|
|
|
|
hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Down", ha_host.id)
|
|
#the test could have failed here but we will try our best to get host back in consistent state
|
|
|
|
no_of_vms = self.noOfVMsOnHost(ha_host.id)
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
#
|
|
hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostUp):
|
|
self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
|
|
|
|
|
|
hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
|
|
|
|
if not(hostDownInCloudstack):
|
|
raise self.fail("Host is not down %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
if not(hostUpInCloudstack):
|
|
raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
|
|
return
|
|
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_02_host_ha_with_local_storage_and_nfs(self):
|
|
raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
|
|
return
|
|
if not(self.isLocalAndNFSStorageAvailable()):
|
|
raise unittest.SkipTest("Skipping this test as this is for Local storage and NFS storage only.");
|
|
return
|
|
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Hypervisor = {}'.format(host.id))
|
|
|
|
|
|
if len(listHost) != 2:
|
|
self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
return
|
|
|
|
no_of_vms = self.noOfVMsOnHost(listHost[0].id)
|
|
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
|
|
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
|
|
|
|
if no_of_vms < 5:
|
|
self.logger.debug("test_02: Create VMs as there are not enough vms to check host ha")
|
|
no_vm_req = 5 - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req, True)
|
|
|
|
ha_host = listHost[1]
|
|
other_host = listHost[0]
|
|
if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
|
|
ha_host = listHost[0]
|
|
other_host = listHost[1]
|
|
|
|
self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
|
|
|
|
hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostDown):
|
|
raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
|
|
|
|
hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Down", ha_host.id)
|
|
#the test could have failed here but we will try our best to get host back in consistent state
|
|
|
|
no_of_vms = self.noOfVMsOnHost(ha_host.id)
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
#
|
|
hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostUp):
|
|
self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
|
|
|
|
|
|
hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
|
|
|
|
if not(hostDownInCloudstack):
|
|
raise self.fail("Host is not down %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
if not(hostUpInCloudstack):
|
|
raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
|
|
return
|
|
|
|
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_03_host_ha_with_only_local_storage(self):
|
|
raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
|
|
return
|
|
|
|
if not(self.isOnlyLocalStorageAvailable()):
|
|
raise unittest.SkipTest("Skipping this test as this is for Local storage only.");
|
|
return
|
|
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Hypervisor = {}'.format(host.id))
|
|
|
|
|
|
if len(listHost) != 2:
|
|
self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
return
|
|
|
|
no_of_vms = self.noOfVMsOnHost(listHost[0].id)
|
|
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
|
|
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
|
|
if no_of_vms < 5:
|
|
self.logger.debug("test_03: Create VMs as there are not enough vms to check host ha")
|
|
no_vm_req = 5 - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req, True)
|
|
|
|
ha_host = listHost[1]
|
|
other_host = listHost[0]
|
|
if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
|
|
ha_host = listHost[0]
|
|
other_host = listHost[1]
|
|
|
|
self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
|
|
|
|
hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostDown):
|
|
raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
|
|
|
|
hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Alert", ha_host.id)
|
|
#the test could have failed here but we will try our best to get host back in consistent state
|
|
|
|
no_of_vms = self.noOfVMsOnHost(ha_host.id)
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
#
|
|
hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
|
|
if not(hostUp):
|
|
self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
|
|
|
|
|
|
hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
|
|
|
|
if not(hostDownInCloudstack):
|
|
raise self.fail("Host is not in alert %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
if not(hostUpInCloudstack):
|
|
raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
|
|
return
|
|
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_04_host_ha_vmactivity_check(self):
|
|
|
|
if not(self.isOnlyNFSStorageAvailable()):
|
|
raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
|
|
return
|
|
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Hypervisor = {}'.format(host.id))
|
|
|
|
|
|
if len(listHost) != 2:
|
|
self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
|
|
return
|
|
|
|
no_of_vms = self.noOfVMsOnHost(listHost[0].id)
|
|
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
|
|
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
|
|
|
|
if no_of_vms < 5:
|
|
self.logger.debug("test_01: Create VMs as there are not enough vms to check host ha")
|
|
no_vm_req = 5 - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req, False)
|
|
|
|
ha_host = listHost[1]
|
|
other_host = listHost[0]
|
|
if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
|
|
ha_host = listHost[0]
|
|
other_host = listHost[1]
|
|
|
|
self.stopAgentOnHost(ha_host.ipaddress, 150)
|
|
|
|
hostDisconnectedInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Disconnected", ha_host.id)
|
|
#the test could have failed here but we will try our best to get host back in consistent state
|
|
|
|
no_of_vms = self.noOfVMsOnHost(ha_host.id)
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
|
|
self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
|
|
#
|
|
|
|
hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
|
|
|
|
if not(hostDisconnectedInCloudstack):
|
|
raise self.fail("Host is not disconnected %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
if not(hostUpInCloudstack):
|
|
raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
|
|
|
|
return
|