mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 08:42:29 +01:00
* DB : Add support for MySQL 8
- Splits commands to create user and grant access on database, the old
statement is no longer supported by MySQL 8.x
- `NO_AUTO_CREATE_USER` is no longer supported by MySQL 8.x so remove
that from db.properties conn parameters
For mysql-server 8.x setup the following changes were added/tested to
make it work with CloudStack in /etc/mysql/mysql.conf.d/mysqld.cnf and
then restart the mysql-server process:
server_id = 1
sql-mode="STRICT_TRANS_TABLES,NO_ENGINE_SUBSTITUTION,ERROR_FOR_DIVISION_BY_ZERO,NO_ZERO_DATE,NO_ZERO_IN_DATE,NO_ENGINE_SUBSTITUTION"
innodb_rollback_on_timeout=1
innodb_lock_wait_timeout=600
max_connections=1000
log-bin=mysql-bin
binlog-format = 'ROW'
default-authentication-plugin=mysql_native_password
Notice the last line above, this is to reset the old password based
authentication used by MySQL 5.x.
Developers can set empty password as follows:
> sudo mysql -u root
ALTER USER 'root'@'localhost' IDENTIFIED BY '';
In libvirt repository, there are two related commits
2019-08-23 13:13 Daniel P. Berrangé ● rpm: don't enable socket activation in upgrade if --listen present
2019-08-22 14:52 Daniel P. Berrangé ● remote: forbid the --listen arg when systemd socket activation
In libvirt.spec.in
/bin/systemctl mask libvirtd.socket >/dev/null 2>&1 || :
/bin/systemctl mask libvirtd-ro.socket >/dev/null 2>&1 || :
/bin/systemctl mask libvirtd-admin.socket >/dev/null 2>&1 || :
/bin/systemctl mask libvirtd-tls.socket >/dev/null 2>&1 || :
/bin/systemctl mask libvirtd-tcp.socket >/dev/null 2>&1 || :
Co-authored-by: Wei Zhou <w.zhou@global.leaseweb.com>
Co-authored-by: Abhishek Kumar <abhishek.mrt22@gmail.com>
Co-authored-by: Rohit Yadav <rohit.yadav@shapeblue.com>
721 lines
28 KiB
Python
721 lines
28 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
""" BVT tests for Hosts Maintenance
|
|
"""
|
|
|
|
# Import Local Modules
|
|
from marvin.cloudstackTestCase import *
|
|
from marvin.lib.utils import *
|
|
from marvin.lib.base import *
|
|
from marvin.lib.common import (get_zone, get_pod, get_suitable_test_template, list_ssvms)
|
|
from nose.plugins.attrib import attr
|
|
from marvin.lib.decoratorGenerators import skipTestIf
|
|
from distutils.util import strtobool
|
|
from marvin.sshClient import SshClient
|
|
|
|
_multiprocess_shared_ = False
|
|
MIN_VMS_FOR_TEST = 3
|
|
|
|
class TestHostMaintenanceBase(cloudstackTestCase):
|
|
def get_ssh_client(self, ip, username, password, retries=10):
|
|
""" Setup ssh client connection and return connection """
|
|
try:
|
|
ssh_client = SshClient(ip, 22, username, password, retries)
|
|
except Exception as e:
|
|
raise unittest.SkipTest("Unable to create ssh connection: " % e)
|
|
|
|
self.assertIsNotNone(
|
|
ssh_client, "Failed to setup ssh connection to ip=%s" % ip)
|
|
|
|
return ssh_client
|
|
|
|
def wait_until_host_is_in_state(self, hostid, resourcestate, interval=3, retries=20):
|
|
def check_resource_state():
|
|
response = Host.list(
|
|
self.apiclient,
|
|
id=hostid
|
|
)
|
|
if isinstance(response, list):
|
|
if response[0].resourcestate == resourcestate:
|
|
self.logger.debug('Host with id %s is in resource state = %s' % (hostid, resourcestate))
|
|
return True, None
|
|
else:
|
|
self.logger.debug("Waiting for host " + hostid +
|
|
" to reach state " + resourcestate +
|
|
", with current state " + response[0].resourcestate)
|
|
return False, None
|
|
|
|
done, _ = wait_until(interval, retries, check_resource_state)
|
|
if not done:
|
|
raise Exception("Failed to wait for host %s to be on resource state %s" % (hostid, resourcestate))
|
|
return True
|
|
|
|
def prepare_host_for_maintenance(self, hostid):
|
|
self.logger.debug("Sending Host with id %s to prepareHostForMaintenance" % hostid)
|
|
cmd = prepareHostForMaintenance.prepareHostForMaintenanceCmd()
|
|
cmd.id = hostid
|
|
response = self.apiclient.prepareHostForMaintenance(cmd)
|
|
self.logger.debug("Host with id %s is in prepareHostForMaintenance" % hostid)
|
|
self.logger.debug(response)
|
|
return response
|
|
|
|
def cancel_host_maintenance(self, hostid):
|
|
self.logger.debug("Canceling Host with id %s from maintain" % hostid)
|
|
cmd = cancelHostMaintenance.cancelHostMaintenanceCmd()
|
|
cmd.id = hostid
|
|
res = self.apiclient.cancelHostMaintenance(cmd)
|
|
self.logger.debug("Host with id %s is cancelling maintenance" % hostid)
|
|
return res
|
|
|
|
def revert_host_state_on_failure(self, hostId):
|
|
cmd = updateHost.updateHostCmd()
|
|
cmd.id = hostId
|
|
cmd.allocationstate = "Enable"
|
|
response = self.apiclient.updateHost(cmd)
|
|
self.assertEqual(response.resourcestate, "Enabled")
|
|
|
|
|
|
class TestHostMaintenance(TestHostMaintenanceBase):
|
|
|
|
def setUp(self):
|
|
self.logger = logging.getLogger('TestHM')
|
|
self.stream_handler = logging.StreamHandler()
|
|
self.logger.setLevel(logging.DEBUG)
|
|
self.logger.addHandler(self.stream_handler)
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.hypervisor = self.testClient.getHypervisorInfo()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.services = self.testClient.getParsedTestDataConfig()
|
|
self.zone = get_zone(self.apiclient, self.testClient.getZoneForTests())
|
|
self.pod = get_pod(self.apiclient, self.zone.id)
|
|
self.cleanup = []
|
|
self.hostConfig = self.config.__dict__["zones"][0].__dict__["pods"][0].__dict__["clusters"][0].__dict__["hosts"][0].__dict__
|
|
|
|
|
|
def tearDown(self):
|
|
try:
|
|
# Clean up, terminate the created templates
|
|
cleanup_resources(self.apiclient, self.cleanup)
|
|
|
|
except Exception as e:
|
|
raise Exception("Warning: Exception during cleanup : %s" % e)
|
|
|
|
return
|
|
|
|
def createVMs(self, hostId, number, offering_key="tiny"):
|
|
self.template = get_suitable_test_template(
|
|
self.apiclient,
|
|
self.zone.id,
|
|
None,
|
|
self.hypervisor
|
|
)
|
|
if self.template == FAILED:
|
|
assert False, "get_suitable_test_template() failed to return template"
|
|
|
|
self.logger.debug("Using template %s " % self.template.id)
|
|
|
|
self.service_offering = ServiceOffering.create(
|
|
self.apiclient,
|
|
self.services["service_offerings"][offering_key]
|
|
)
|
|
self.logger.debug("Using service offering %s " % self.service_offering.id)
|
|
self.network_offering = NetworkOffering.create(
|
|
self.apiclient,
|
|
self.services["l2-network_offering"],
|
|
)
|
|
self.network_offering.update(self.apiclient, state='Enabled')
|
|
self.services["network"]["networkoffering"] = self.network_offering.id
|
|
self.l2_network = Network.create(
|
|
self.apiclient,
|
|
self.services["l2-network"],
|
|
zoneid=self.zone.id,
|
|
networkofferingid=self.network_offering.id
|
|
)
|
|
|
|
vms=[]
|
|
for i in range(0, number):
|
|
self.services["virtual_machine"]["zoneid"] = self.zone.id
|
|
self.services["virtual_machine"]["template"] = self.template.id
|
|
self.services["virtual_machine"]["displayname"] = 'vm' + str(i)
|
|
self.services["virtual_machine"]["hypervisor"] = self.hypervisor
|
|
vm = VirtualMachine.create(
|
|
self.apiclient,
|
|
self.services["virtual_machine"],
|
|
serviceofferingid=self.service_offering.id,
|
|
networkids=self.l2_network.id,
|
|
hostid=hostId
|
|
)
|
|
vms.append(vm)
|
|
self.cleanup.append(vm)
|
|
self.logger.debug("VM create = {}".format(vm.id))
|
|
self.cleanup.append(self.l2_network)
|
|
self.cleanup.append(self.network_offering)
|
|
self.cleanup.append(self.service_offering)
|
|
return vms
|
|
|
|
def checkAllVmsRunningOnHost(self, hostId):
|
|
listVms1 = VirtualMachine.list(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
|
|
if (listVms1 is not None):
|
|
self.logger.debug('Vms found to test all running = {} '.format(len(listVms1)))
|
|
for vm in listVms1:
|
|
if (vm.state != "Running"):
|
|
self.logger.debug('VirtualMachine on Host with id = {} is in {}'.format(vm.id, vm.state))
|
|
return (False, None)
|
|
|
|
response = list_ssvms(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
if isinstance(response, list):
|
|
for systemvm in response:
|
|
if systemvm.state != 'Running':
|
|
self.logger.debug("Found not running VM {}".format(systemvm.name))
|
|
return (False, None)
|
|
|
|
return (True, None)
|
|
|
|
def checkVmMigratingOnHost(self, hostId):
|
|
vm_migrating=False
|
|
listVms1 = VirtualMachine.list(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
|
|
if (listVms1 is not None):
|
|
self.logger.debug('Vms found = {} '.format(len(listVms1)))
|
|
for vm in listVms1:
|
|
if (vm.state == "Migrating"):
|
|
self.logger.debug('VirtualMachine on Host with id = {} is in {}'.format(vm.id, vm.state))
|
|
vm_migrating=True
|
|
break
|
|
|
|
return (vm_migrating, None)
|
|
|
|
def migrationsFinished(self, hostId):
|
|
migrations_finished=True
|
|
listVms1 = VirtualMachine.list(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
|
|
if (listVms1 is not None):
|
|
numVms = len(listVms1)
|
|
migrations_finished = (numVms == 0)
|
|
|
|
return (migrations_finished, None)
|
|
|
|
def noOfVMsOnHost(self, hostId):
|
|
listVms = VirtualMachine.list(
|
|
self.apiclient,
|
|
hostid=hostId
|
|
)
|
|
no_of_vms=0
|
|
self.logger.debug("Counting VMs on host " + hostId)
|
|
if (listVms is not None):
|
|
for vm in listVms:
|
|
self.logger.debug("VirtualMachine on Host " + hostId + " = " + vm.id)
|
|
no_of_vms=no_of_vms+1
|
|
self.logger.debug("Found VMs on host " + str(no_of_vms))
|
|
return no_of_vms
|
|
|
|
def hostPrepareAndCancelMaintenance(self, target_host_id, other_host_id):
|
|
# Wait for all VMs to complete any pending migrations.
|
|
if not wait_until(3, 100, self.checkAllVmsRunningOnHost, target_host_id) or \
|
|
not wait_until(3, 100, self.checkAllVmsRunningOnHost, other_host_id):
|
|
raise Exception("Failed to wait for all VMs to reach running state to execute test")
|
|
|
|
self.prepare_host_for_maintenance(target_host_id)
|
|
migrations_finished = wait_until(5, 200, self.migrationsFinished, target_host_id)
|
|
|
|
self.wait_until_host_is_in_state(target_host_id, "Maintenance", 5, 200)
|
|
|
|
vm_count_after_maintenance = self.noOfVMsOnHost(target_host_id)
|
|
|
|
self.cancel_host_maintenance(target_host_id)
|
|
self.wait_until_host_is_in_state(target_host_id, "Enabled", 5, 200)
|
|
|
|
if vm_count_after_maintenance != 0:
|
|
self.fail("Host to put to maintenance still has VMs running")
|
|
|
|
return migrations_finished
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_01_cancel_host_maintenace_with_no_migration_jobs(self):
|
|
"""
|
|
Tests if putting a host with no migrations (0 VMs) work back and forth
|
|
|
|
1) Verify if there are at least 2 hosts in enabled state.
|
|
2) Put the host into maintenance verify success
|
|
3) Put the other host into maintenance, verify success
|
|
"""
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
hypervisor=self.hypervisor,
|
|
resourcestate='Enabled',
|
|
state='Up'
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Found Host = {}'.format(host.id))
|
|
|
|
|
|
if (len(listHost) < 2):
|
|
raise unittest.SkipTest("Canceling tests for host maintenance as we need 2 or more hosts up and enabled")
|
|
|
|
try:
|
|
|
|
migrations_finished = self.hostPrepareAndCancelMaintenance(listHost[0].id, listHost[1].id)
|
|
|
|
if migrations_finished:
|
|
self.hostPrepareAndCancelMaintenance(listHost[1].id, listHost[0].id)
|
|
else:
|
|
raise unittest.SkipTest("VMs are still migrating so reverse migration /maintenace skipped")
|
|
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(listHost[0].id)
|
|
self.revert_host_state_on_failure(listHost[1].id)
|
|
self.logger.debug("Exception {}".format(e))
|
|
self.fail("Host maintenance test failed {}".format(e[0]))
|
|
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_02_cancel_host_maintenace_with_migration_jobs(self):
|
|
"""
|
|
Tests if putting a host with migrations (3 VMs) work back and forth
|
|
|
|
1) Verify if there are at least 2 hosts in enabled state.
|
|
2) Deploy VMs if needed
|
|
3) Put the host into maintenance verify success -ensure existing host has zero running VMs
|
|
4) Put the other host into maintenance, verify success just as step 3
|
|
"""
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
hypervisor=self.hypervisor,
|
|
resourcestate='Enabled',
|
|
state='Up'
|
|
)
|
|
for host in listHost:
|
|
self.logger.debug('Found Host = {}'.format(host.id))
|
|
|
|
if (len(listHost) < 2):
|
|
raise unittest.SkipTest("Canceling tests for host maintenance as we need 2 or more hosts up and enabled")
|
|
|
|
no_of_vms = self.noOfVMsOnHost(listHost[0].id)
|
|
|
|
no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
|
|
|
|
if no_of_vms < MIN_VMS_FOR_TEST:
|
|
self.logger.debug("Create VMs as there are not enough vms to check host maintenance")
|
|
no_vm_req = MIN_VMS_FOR_TEST - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req)
|
|
|
|
try:
|
|
migrations_finished = self.hostPrepareAndCancelMaintenance(listHost[0].id, listHost[1].id)
|
|
|
|
if migrations_finished:
|
|
self.hostPrepareAndCancelMaintenance(listHost[1].id, listHost[0].id)
|
|
else:
|
|
raise unittest.SkipTest("VMs are still migrating so reverse migration /maintenace skipped")
|
|
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(listHost[0].id)
|
|
self.revert_host_state_on_failure(listHost[1].id)
|
|
self.logger.debug("Exception {}".format(e))
|
|
self.fail("Host maintenance test failed {}".format(e[0]))
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"smoke",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="true")
|
|
def test_03_cancel_host_maintenace_with_migration_jobs_failure(self):
|
|
"""
|
|
Tests if putting a host with impossible migrations (2 VMs) work pushes to ErrorInMaintenance state
|
|
|
|
1) Verify if there are at least 2 hosts in enabled state.
|
|
2) Tag the host and deploy tagged VMs which cannot be migrated to other host without tags
|
|
3) Put the host into maintenance verify it fails with it reaching ErrorInMaintenance
|
|
"""
|
|
listHost = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
hypervisor=self.hypervisor,
|
|
resourcestate='Enabled',
|
|
state='Up'
|
|
)
|
|
|
|
for host in listHost:
|
|
self.logger.debug('Found Host = {}'.format(host.id))
|
|
|
|
if (len(listHost) < 2):
|
|
raise unittest.SkipTest("Canceling tests for host maintenance as we need 2 or more hosts up and enabled")
|
|
|
|
target_host_id = listHost[0].id
|
|
|
|
try:
|
|
Host.update(self.apiclient,
|
|
id=target_host_id,
|
|
hosttags=self.services["service_offerings"]["taggedsmall"]["hosttags"])
|
|
|
|
no_of_vms = self.noOfVMsOnHost(target_host_id)
|
|
|
|
# Need only 2 VMs for this case.
|
|
if no_of_vms < 2:
|
|
self.logger.debug("Create VMs as there are not enough vms to check host maintenance")
|
|
no_vm_req = 2 - no_of_vms
|
|
if (no_vm_req > 0):
|
|
self.logger.debug("Creating vms = {}".format(no_vm_req))
|
|
self.vmlist = self.createVMs(listHost[0].id, no_vm_req, "taggedsmall")
|
|
|
|
# Attempt putting host in maintenance and check if ErrorInMaintenance state is reached
|
|
self.prepare_host_for_maintenance(target_host_id)
|
|
error_in_maintenance_reached = self.wait_until_host_is_in_state(target_host_id, "ErrorInMaintenance", 5, 300)
|
|
|
|
self.cancel_host_maintenance(target_host_id)
|
|
self.wait_until_host_is_in_state(target_host_id, "Enabled", 5, 200)
|
|
|
|
Host.update(self.apiclient, id=target_host_id, hosttags="")
|
|
|
|
if not error_in_maintenance_reached:
|
|
self.fail("Error in maintenance state should have reached after ports block")
|
|
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(listHost[0].id)
|
|
self.revert_host_state_on_failure(listHost[1].id)
|
|
Host.update(self.apiclient, id=target_host_id, hosttags="")
|
|
self.logger.debug("Exception {}".format(e))
|
|
self.fail("Host maintenance test failed {}".format(e[0]))
|
|
|
|
|
|
class TestHostMaintenanceAgents(TestHostMaintenanceBase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestHostMaintenanceAgents, cls).getClsTestClient()
|
|
cls.apiclient = cls.testClient.getApiClient()
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
cls.dbclient = cls.testClient.getDbConnection()
|
|
cls.zone = get_zone(cls.apiclient, cls.testClient.getZoneForTests())
|
|
cls.pod = get_pod(cls.apiclient, cls.zone.id)
|
|
cls.services = cls.testClient.getParsedTestDataConfig()
|
|
|
|
cls.logger = logging.getLogger('TestHMAgents')
|
|
cls.stream_handler = logging.StreamHandler()
|
|
cls.logger.setLevel(logging.DEBUG)
|
|
cls.logger.addHandler(cls.stream_handler)
|
|
|
|
cls._cleanup = []
|
|
cls.hypervisorNotSupported = False
|
|
if cls.hypervisor.lower() not in ['kvm', 'lxc']:
|
|
cls.hypervisorNotSupported = True
|
|
|
|
if not cls.hypervisorNotSupported:
|
|
cls.initialsshvalue = cls.is_ssh_enabled()
|
|
|
|
cls.template = get_suitable_test_template(
|
|
cls.apiclient,
|
|
cls.zone.id,
|
|
None,
|
|
cls.hypervisor
|
|
)
|
|
if cls.template == FAILED:
|
|
assert False, "get_suitable_test_template() failed to return template"
|
|
|
|
cls.services["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.services["virtual_machine"]["template"] = cls.template.id
|
|
cls.services["virtual_machine"]["hypervisor"] = cls.hypervisor
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.apiclient,
|
|
cls.services["service_offerings"]["tiny"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
cls.network_offering = NetworkOffering.create(
|
|
cls.apiclient,
|
|
cls.services["l2-network_offering"],
|
|
)
|
|
cls.network_offering.update(cls.apiclient, state='Enabled')
|
|
cls.services["network"]["networkoffering"] = cls.network_offering.id
|
|
cls.l2_network = Network.create(
|
|
cls.apiclient,
|
|
cls.services["l2-network"],
|
|
zoneid=cls.zone.id,
|
|
networkofferingid=cls.network_offering.id
|
|
)
|
|
cls._cleanup.append(cls.l2_network)
|
|
cls._cleanup.append(cls.network_offering)
|
|
|
|
cls.hostConfig = cls.config.__dict__["zones"][0].__dict__["pods"][0].__dict__["clusters"][0].__dict__["hosts"][0].__dict__
|
|
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
try:
|
|
if not cls.hypervisorNotSupported:
|
|
# Revert setting value to the original
|
|
cls.set_ssh_enabled(cls.initialsshvalue)
|
|
cleanup_resources(cls.apiclient, cls._cleanup)
|
|
except Exception as e:
|
|
raise Exception("Warning: Exception during cleanup : %s" % e)
|
|
|
|
def setUp(self):
|
|
if not self.hypervisorNotSupported:
|
|
self.host = self.get_enabled_host_connected_agent()
|
|
self.cleanup = []
|
|
|
|
def tearDown(self):
|
|
try:
|
|
cleanup_resources(self.apiclient, self.cleanup)
|
|
except Exception as e:
|
|
raise Exception("Warning: Exception during cleanup : %s" % e)
|
|
|
|
|
|
@classmethod
|
|
def is_ssh_enabled(cls):
|
|
conf = Configurations.list(cls.apiclient, name="kvm.ssh.to.agent")
|
|
if not conf:
|
|
return False
|
|
else:
|
|
return bool(strtobool(conf[0].value)) if conf[0].value else False
|
|
|
|
@classmethod
|
|
def updateConfiguration(self, name, value):
|
|
cmd = updateConfiguration.updateConfigurationCmd()
|
|
cmd.name = name
|
|
cmd.value = value
|
|
self.apiclient.updateConfiguration(cmd)
|
|
|
|
@classmethod
|
|
def set_ssh_enabled(cls, on):
|
|
value = "true" if on else "false"
|
|
cls.updateConfiguration('kvm.ssh.to.agent', value)
|
|
|
|
def wait_until_agent_is_in_state(self, hostid, state, interval=3, retries=20):
|
|
def check_agent_state():
|
|
response = Host.list(
|
|
self.apiclient,
|
|
id=hostid
|
|
)
|
|
if isinstance(response, list):
|
|
if response[0].state == state:
|
|
self.logger.debug('Host agent with id %s is in state = %s' % (hostid, state))
|
|
return True, None
|
|
return False, None
|
|
|
|
done, _ = wait_until(interval, retries, check_agent_state)
|
|
if not done:
|
|
raise Exception("Failed to wait for host agent %s to be on state %s" % (hostid, state))
|
|
return True
|
|
|
|
def get_enabled_host_connected_agent(self):
|
|
hosts = Host.list(
|
|
self.apiclient,
|
|
type='Routing',
|
|
zoneid=self.zone.id,
|
|
podid=self.pod.id,
|
|
hypervisor=self.hypervisor,
|
|
resourcestate='Enabled',
|
|
state='Up'
|
|
)
|
|
if len(hosts) < 2:
|
|
raise unittest.SkipTest("Host maintenance tests must be tested for 2 or more hosts")
|
|
return hosts[0]
|
|
|
|
def deploy_vm_on_host(self, hostid):
|
|
return VirtualMachine.create(
|
|
self.apiclient,
|
|
self.services["virtual_machine"],
|
|
serviceofferingid=self.service_offering.id,
|
|
networkids=self.l2_network.id,
|
|
hostid=hostid
|
|
)
|
|
|
|
def assert_host_is_functional_after_cancelling_maintenance(self, hostid):
|
|
self.wait_until_agent_is_in_state(hostid, "Up")
|
|
self.logger.debug('Deploying VM on host %s' % hostid)
|
|
vm = self.deploy_vm_on_host(hostid)
|
|
self.assertEqual(
|
|
vm.state,
|
|
"Running",
|
|
"Check VM is running on the host"
|
|
)
|
|
self.cleanup.append(vm)
|
|
|
|
@skipTestIf("hypervisorNotSupported")
|
|
@attr(tags=["advanced", "advancedns", "smoke", "basic", "eip", "sg"], required_hardware="true")
|
|
def test_01_cancel_host_maintenance_ssh_enabled_agent_connected(self):
|
|
"""
|
|
Test cancel maintenance when: 'kvm.ssh.to.agent' = true, agent state = 'Up'
|
|
|
|
1) Put host on Maintenance
|
|
2) Cancel maintenance on host
|
|
4) Assert agent is still connected after cancelling maintenance
|
|
3) Deploy VM on the host after cancelling maintenance
|
|
"""
|
|
|
|
if not self.is_ssh_enabled():
|
|
self.set_ssh_enabled(True)
|
|
|
|
try:
|
|
self.prepare_host_for_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Maintenance")
|
|
self.cancel_host_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Enabled")
|
|
self.assert_host_is_functional_after_cancelling_maintenance(self.host.id)
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(self.host.id)
|
|
self.fail(e)
|
|
|
|
@skipTestIf("hypervisorNotSupported")
|
|
@attr(tags=["boris", "advancedns", "smoke", "basic", "eip", "sg"], required_hardware="true")
|
|
def test_02_cancel_host_maintenance_ssh_enabled_agent_disconnected(self):
|
|
"""
|
|
Test cancel maintenance when: 'kvm.ssh.to.agent' = true, agent state != 'Up'
|
|
|
|
1) Put host on maintenance
|
|
2) SSH into host and stop cloudstack-agent service - host gets Disconnected
|
|
3) Cancel maintenance on host
|
|
4) Assert agent is connected after cancelling maintenance
|
|
5) Deploy VM on the host
|
|
"""
|
|
|
|
if not self.is_ssh_enabled():
|
|
self.set_ssh_enabled(True)
|
|
# username, password = self.get_host_credentials(self.host.id)
|
|
username = self.hostConfig["username"]
|
|
password = self.hostConfig["password"]
|
|
|
|
try:
|
|
self.prepare_host_for_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Maintenance")
|
|
|
|
ssh_client = self.get_ssh_client(self.host.ipaddress, self.hostConfig["username"],
|
|
self.hostConfig["password"])
|
|
ssh_client.execute("service cloudstack-agent stop")
|
|
self.wait_until_agent_is_in_state(self.host.id, "Disconnected")
|
|
|
|
self.cancel_host_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Enabled")
|
|
|
|
self.assert_host_is_functional_after_cancelling_maintenance(self.host.id)
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(self.host.id)
|
|
self.fail(e)
|
|
|
|
@skipTestIf("hypervisorNotSupported")
|
|
@attr(tags=["advanced", "advancedns", "smoke", "basic", "eip", "sg"], required_hardware="true")
|
|
def test_03_cancel_host_maintenance_ssh_disabled_agent_connected(self):
|
|
"""
|
|
Test cancel maintenance when: 'kvm.ssh.to.agent' = false, agent state = 'Up'
|
|
|
|
1) Put host on Maintenance
|
|
2) Cancel maintenance on host
|
|
4) Assert agent is still connected after cancelling maintenance
|
|
3) Deploy VM on the host after cancelling maintenance
|
|
"""
|
|
|
|
if self.is_ssh_enabled():
|
|
self.set_ssh_enabled(False)
|
|
|
|
try:
|
|
self.prepare_host_for_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Maintenance")
|
|
self.cancel_host_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Enabled")
|
|
self.assert_host_is_functional_after_cancelling_maintenance(self.host.id)
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(self.host.id)
|
|
self.fail(e)
|
|
|
|
@skipTestIf("hypervisorNotSupported")
|
|
@attr(tags=["advanced", "advancedns", "smoke", "basic", "eip", "sg"], required_hardware="true")
|
|
def test_04_cancel_host_maintenance_ssh_disabled_agent_disconnected(self):
|
|
"""
|
|
Test cancel maintenance when: 'kvm.ssh.to.agent' = false, agent state != 'Up'
|
|
|
|
1) Put host on maintenance
|
|
2) SSH into host (if possible) and stop cloudstack-agent service - host gets Disconnected.
|
|
Skip test if not possible to SSH into host
|
|
3) Cancel maintenance on host - assert cannot cancel maintenance on disconnected host (exception thwown)
|
|
4( SSH into host and start cloudstack-agent service - host gets connected
|
|
5) Cancel maintenance on host
|
|
4) Assert agent is connected after cancelling maintenance
|
|
5) Deploy VM on the host
|
|
"""
|
|
|
|
if self.is_ssh_enabled():
|
|
self.set_ssh_enabled(False)
|
|
|
|
try:
|
|
self.prepare_host_for_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Maintenance")
|
|
|
|
ssh_client = self.get_ssh_client(self.host.ipaddress, self.hostConfig["username"],
|
|
self.hostConfig["password"])
|
|
ssh_client.execute("service cloudstack-agent stop")
|
|
self.wait_until_agent_is_in_state(self.host.id, "Disconnected")
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(self.host.id)
|
|
self.fail(e)
|
|
|
|
self.assertRaises(Exception, self.cancel_host_maintenance, self.host.id)
|
|
|
|
try:
|
|
ssh_client = self.get_ssh_client(self.host.ipaddress, self.hostConfig["username"],
|
|
self.hostConfig["password"])
|
|
ssh_client.execute("service cloudstack-agent start")
|
|
self.wait_until_agent_is_in_state(self.host.id, "Up")
|
|
|
|
self.cancel_host_maintenance(self.host.id)
|
|
self.wait_until_host_is_in_state(self.host.id, "Enabled")
|
|
self.assert_host_is_functional_after_cancelling_maintenance(self.host.id)
|
|
except Exception as e:
|
|
self.revert_host_state_on_failure(self.host.id)
|
|
self.fail(e)
|