mirror of
				https://github.com/apache/cloudstack.git
				synced 2025-10-26 08:42:29 +01:00 
			
		
		
		
	
		
			
				
	
	
		
			516 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			516 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # Licensed to the Apache Software Foundation (ASF) under one
 | |
| # or more contributor license agreements.  See the NOTICE file
 | |
| # distributed with this work for additional information
 | |
| # regarding copyright ownership.  The ASF licenses this file
 | |
| # to you under the Apache License, Version 2.0 (the
 | |
| # "License"); you may not use this file except in compliance
 | |
| # with the License.  You may obtain a copy of the License at
 | |
| #
 | |
| #   http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing,
 | |
| # software distributed under the License is distributed on an
 | |
| # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 | |
| # KIND, either express or implied.  See the License for the
 | |
| # specific language governing permissions and limitations
 | |
| # under the License.
 | |
| from __builtin__ import False
 | |
| """ BVT tests for Hosts Maintenance
 | |
| """
 | |
| 
 | |
| # Import Local Modules
 | |
| from marvin.codes import FAILED
 | |
| from marvin.cloudstackTestCase import *
 | |
| from marvin.cloudstackAPI import *
 | |
| from marvin.lib.utils import *
 | |
| from marvin.lib.base import *
 | |
| from marvin.lib.common import *
 | |
| from nose.plugins.attrib import attr
 | |
| 
 | |
| from time import sleep
 | |
| 
 | |
| _multiprocess_shared_ = False
 | |
| 
 | |
| 
 | |
| class TestHostHA(cloudstackTestCase):
 | |
| 
 | |
|     def setUp(self):
 | |
|         self.logger = logging.getLogger('TestHM')
 | |
|         self.stream_handler = logging.StreamHandler()
 | |
|         self.logger.setLevel(logging.DEBUG)
 | |
|         self.logger.addHandler(self.stream_handler)
 | |
|         self.apiclient = self.testClient.getApiClient()
 | |
|         self.hypervisor = self.testClient.getHypervisorInfo()
 | |
|         self.dbclient = self.testClient.getDbConnection()
 | |
|         self.services = self.testClient.getParsedTestDataConfig()
 | |
|         self.zone = get_zone(self.apiclient, self.testClient.getZoneForTests())
 | |
|         self.pod = get_pod(self.apiclient, self.zone.id)
 | |
|         self.cleanup = []
 | |
|         self.services = {
 | |
|                             "service_offering": {
 | |
|                                 "name": "Ultra Tiny Instance",
 | |
|                                 "displaytext": "Ultra Tiny Instance",
 | |
|                                 "cpunumber": 1,
 | |
|                                 "cpuspeed": 100,
 | |
|                                 "memory": 128,
 | |
|                             },
 | |
|                             "service_offering_local": {
 | |
|                                 "name": "Ultra Tiny Local Instance",
 | |
|                                 "displaytext": "Ultra Tiny Local Instance",
 | |
|                                 "cpunumber": 1,
 | |
|                                 "cpuspeed": 100,
 | |
|                                 "memory": 128,
 | |
|                                 "storagetype": "local"
 | |
|                             },
 | |
|                             "vm": {
 | |
|                                 "username": "root",
 | |
|                                 "password": "password",
 | |
|                                 "ssh_port": 22,
 | |
|                                 # Hypervisor type should be same as
 | |
|                                 # hypervisor type of cluster
 | |
|                                 "privateport": 22,
 | |
|                                 "publicport": 22,
 | |
|                                 "protocol": 'TCP',
 | |
|                             },
 | |
|                             "natrule": {
 | |
|                                 "privateport": 22,
 | |
|                                 "publicport": 22,
 | |
|                                 "startport": 22,
 | |
|                                 "endport": 22,
 | |
|                                 "protocol": "TCP",
 | |
|                                 "cidrlist": '0.0.0.0/0',
 | |
|                             },
 | |
|                          "ostype": 'CentOS 5.3 (64-bit)',
 | |
|                          "sleep": 60,
 | |
|                          "timeout": 10,
 | |
|                          }
 | |
|         
 | |
| 
 | |
|     def tearDown(self):
 | |
|         try:
 | |
|             # Clean up, terminate the created templates
 | |
|             cleanup_resources(self.apiclient, self.cleanup)
 | |
| 
 | |
|         except Exception as e:
 | |
|             raise Exception("Warning: Exception during cleanup : %s" % e)
 | |
| 
 | |
|         return
 | |
|     
 | |
|     def createVMs(self, hostId, number, local):
 | |
|         
 | |
|         self.template = get_template(
 | |
|             self.apiclient,
 | |
|             self.zone.id,
 | |
|             self.services["ostype"]
 | |
|         )
 | |
|             
 | |
|         if self.template == FAILED:
 | |
|             assert False, "get_template() failed to return template with description %s" % self.services["ostype"]
 | |
|             
 | |
|         self.logger.debug("Using template %s " % self.template.id)
 | |
|         
 | |
|         if local:
 | |
|             self.service_offering = ServiceOffering.create(
 | |
|                 self.apiclient,
 | |
|                 self.services["service_offering_local"]
 | |
|             )
 | |
|         else:
 | |
|             self.service_offering = ServiceOffering.create(
 | |
|                 self.apiclient,
 | |
|                 self.services["service_offering"]
 | |
|             )
 | |
|             
 | |
|             
 | |
|         self.logger.debug("Using service offering %s " % self.service_offering.id)
 | |
|         
 | |
|         vms = []
 | |
|         for i in range(0, number):
 | |
|             self.services["vm"]["zoneid"] = self.zone.id
 | |
|             self.services["vm"]["template"] = self.template.id
 | |
|             self.services["vm"]["displayname"] = 'vm' + str(i)
 | |
|             self.services["vm"]["hypervisor"] = self.hypervisor
 | |
|             vm = VirtualMachine.create(
 | |
|                 self.apiclient,
 | |
|                 self.services["vm"],
 | |
|                 serviceofferingid=self.service_offering.id,
 | |
|                 hostid=hostId
 | |
|             )
 | |
|             vms.append(vm)
 | |
|             self.cleanup.append(vm)
 | |
|             self.logger.debug("VM create = {}".format(vm.id))
 | |
|         return vm
 | |
|     
 | |
|     def noOfVMsOnHost(self, hostId):
 | |
|         listVms = VirtualMachine.list(
 | |
|                        self.apiclient,
 | |
|                        hostid=hostId
 | |
|                        )
 | |
|         vmnos = 0
 | |
|         if (listVms is not None):
 | |
|             for vm in listVms:
 | |
|                 self.logger.debug('VirtualMachine on Hyp 1 = {}'.format(vm.id))
 | |
|                 vmnos = vmnos + 1
 | |
|              
 | |
|         return vmnos
 | |
|     
 | |
|     def checkHostDown(self, fromHostIp, testHostIp):
 | |
|         try:
 | |
|             ssh = SshClient(fromHostIp, 22, "root", "password") 
 | |
|             res = ssh.execute("ping -c 1 %s" % testHostIp)
 | |
|             result = str(res)
 | |
|             if result.count("100% packet loss") == 1:
 | |
|                 return True, 1
 | |
|             else:
 | |
|                 return False, 1
 | |
|         except Exception as e:
 | |
|             self.logger.debug("Got exception %s" % e)
 | |
|             return False, 1
 | |
|         
 | |
|     def checkHostUp(self, fromHostIp, testHostIp):
 | |
|         try:
 | |
|             ssh = SshClient(fromHostIp, 22, "root", "password") 
 | |
|             res = ssh.execute("ping -c 1 %s" % testHostIp)
 | |
|             result = str(res)
 | |
|             if result.count(" 0% packet loss") == 1:
 | |
|                 return True, 1
 | |
|             else:
 | |
|                 return False, 1
 | |
|         except Exception as e:
 | |
|             self.logger.debug("Got exception %s" % e)
 | |
|             return False, 1
 | |
|       
 | |
|         
 | |
|     def isOnlyNFSStorageAvailable(self):
 | |
|         if self.zone.localstorageenabled:
 | |
|             return False
 | |
|         storage_pools = StoragePool.list(
 | |
|                    self.apiclient,
 | |
|                    zoneid=self.zone.id,
 | |
|                    listall=True
 | |
|                     )
 | |
|         self.assertEqual(
 | |
|                            isinstance(storage_pools, list),
 | |
|                            True,
 | |
|                            "Check if listStoragePools returns a valid response"
 | |
|                            )
 | |
|         for storage_pool in storage_pools:
 | |
|             if storage_pool.type == u'NetworkFilesystem':
 | |
|                 return True
 | |
|             
 | |
|         return False
 | |
|     
 | |
|     def isOnlyLocalStorageAvailable(self):
 | |
|         if not(self.zone.localstorageenabled):
 | |
|             return False
 | |
|         
 | |
|         storage_pools = StoragePool.list(
 | |
|                    self.apiclient,
 | |
|                    zoneid=self.zone.id,
 | |
|                    listall=True
 | |
|                     )
 | |
|         self.assertEqual(
 | |
|                            isinstance(storage_pools, list),
 | |
|                            True,
 | |
|                            "Check if listStoragePools returns a valid response"
 | |
|                            )
 | |
|         for storage_pool in storage_pools:
 | |
|             if storage_pool.type == u'NetworkFilesystem':
 | |
|                 return False
 | |
|             
 | |
|         return True
 | |
|     
 | |
|     def isLocalAndNFSStorageAvailable(self):
 | |
|         if not(self.zone.localstorageenabled):
 | |
|             return False
 | |
|         
 | |
|         storage_pools = StoragePool.list(
 | |
|                    self.apiclient,
 | |
|                    zoneid=self.zone.id,
 | |
|                    listall=True
 | |
|                     )
 | |
|         self.assertEqual(
 | |
|                            isinstance(storage_pools, list),
 | |
|                            True,
 | |
|                            "Check if listStoragePools returns a valid response"
 | |
|                            )
 | |
|         for storage_pool in storage_pools:
 | |
|             if storage_pool.type == u'NetworkFilesystem':
 | |
|                 return True
 | |
|             
 | |
|         return False
 | |
|         
 | |
|         
 | |
|     def checkHostStateInCloudstack(self, state, hostId):
 | |
|         try:
 | |
|             listHost = Host.list(
 | |
|                    self.apiclient,
 | |
|                    type='Routing',
 | |
|                    zoneid=self.zone.id,
 | |
|                    podid=self.pod.id,
 | |
|                    id=hostId
 | |
|                )
 | |
|             self.assertEqual(
 | |
|                            isinstance(listHost, list),
 | |
|                            True,
 | |
|                            "Check if listHost returns a valid response"
 | |
|                            )
 | |
|             
 | |
|             self.assertEqual(
 | |
|                            len(listHost),
 | |
|                            1,
 | |
|                            "Check if listHost returns a host"
 | |
|                            )
 | |
|             self.logger.debug(" Host state is %s " % listHost[0].state)
 | |
|             if listHost[0].state == state:
 | |
|                 return True, 1
 | |
|             else:
 | |
|                 return False, 1
 | |
|         except Exception as e:
 | |
|             self.logger.debug("Got exception %s" % e)
 | |
|             return False, 1
 | |
|     
 | |
|     
 | |
|     def disconnectHostfromNetwork(self, hostIp, timeout):
 | |
|         srcFile = os.path.dirname(os.path.realpath(__file__)) + "/test_host_ha.sh"
 | |
|         if not(os.path.isfile(srcFile)):
 | |
|             self.logger.debug("File %s not found" % srcFile)
 | |
|             raise unittest.SkipTest("Script file %s required for HA not found" % srcFile);
 | |
|             
 | |
|         ssh = SshClient(hostIp, 22, "root", "password")
 | |
|         ssh.scp(srcFile, "/root/test_host_ha.sh")
 | |
|         ssh.execute("nohup sh /root/test_host_ha.sh %s > /dev/null 2>&1 &\n" % timeout)
 | |
|         return
 | |
|         
 | |
| 
 | |
|     @attr(
 | |
|         tags=[
 | |
|             "advanced",
 | |
|             "advancedns",
 | |
|             "smoke",
 | |
|             "basic",
 | |
|             "eip",
 | |
|             "sg"],
 | |
|         required_hardware="true")
 | |
|     def test_01_host_ha_with_nfs_storagepool_with_vm(self):
 | |
|         
 | |
|         if not(self.isOnlyNFSStorageAvailable()):
 | |
|             raise unittest.SkipTest("Skipping this test as this is for NFS store only.");
 | |
|             return
 | |
|         
 | |
|         listHost = Host.list(
 | |
|             self.apiclient,
 | |
|             type='Routing',
 | |
|             zoneid=self.zone.id,
 | |
|             podid=self.pod.id,
 | |
|         )
 | |
|         for host in listHost:
 | |
|             self.logger.debug('Hypervisor = {}'.format(host.id))
 | |
|             
 | |
|                   
 | |
|         if len(listHost) != 2:
 | |
|             self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             return
 | |
|         
 | |
|         no_of_vms = self.noOfVMsOnHost(listHost[0].id)
 | |
|         
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
 | |
|         
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|         
 | |
|           
 | |
|         if no_of_vms < 5:
 | |
|             self.logger.debug("test_01: Create VMs as there are not enough vms to check host ha")
 | |
|             no_vm_req = 5 - no_of_vms
 | |
|             if (no_vm_req > 0):
 | |
|                 self.logger.debug("Creating vms = {}".format(no_vm_req))
 | |
|                 self.vmlist = self.createVMs(listHost[0].id, no_vm_req, False)
 | |
|            
 | |
|         ha_host = listHost[1]
 | |
|         other_host = listHost[0]
 | |
|         if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
 | |
|             ha_host = listHost[0]
 | |
|             other_host = listHost[1]
 | |
|                 
 | |
|         self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
 | |
|             
 | |
|         hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostDown): 
 | |
|             raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
 | |
|             
 | |
|         hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Down", ha_host.id)
 | |
|         #the test could have failed here but we will try our best to get host back in consistent state
 | |
|             
 | |
|         no_of_vms = self.noOfVMsOnHost(ha_host.id)
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|             #
 | |
|         hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostUp): 
 | |
|             self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
 | |
|             
 | |
|         
 | |
|         hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
 | |
|         
 | |
|         if not(hostDownInCloudstack): 
 | |
|             raise self.fail("Host is not down %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|         if not(hostUpInCloudstack): 
 | |
|             raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|             
 | |
|         return
 | |
| 
 | |
|     
 | |
|     @attr(
 | |
|         tags=[
 | |
|             "advanced",
 | |
|             "advancedns",
 | |
|             "smoke",
 | |
|             "basic",
 | |
|             "eip",
 | |
|             "sg"],
 | |
|         required_hardware="true")
 | |
|     def test_02_host_ha_with_local_storage_and_nfs(self):
 | |
|     
 | |
|         if not(self.isLocalAndNFSStorageAvailable()):
 | |
|             raise unittest.SkipTest("Skipping this test as this is for Local storage and NFS storage only.");
 | |
|             return
 | |
|         
 | |
|         listHost = Host.list(
 | |
|             self.apiclient,
 | |
|             type='Routing',
 | |
|             zoneid=self.zone.id,
 | |
|             podid=self.pod.id,
 | |
|         )
 | |
|         for host in listHost:
 | |
|             self.logger.debug('Hypervisor = {}'.format(host.id))
 | |
|             
 | |
|                   
 | |
|         if len(listHost) != 2:
 | |
|             self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             return
 | |
|         
 | |
|         no_of_vms = self.noOfVMsOnHost(listHost[0].id)
 | |
|         
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
 | |
|         
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|         
 | |
|           
 | |
|         if no_of_vms < 5:
 | |
|             self.logger.debug("test_02: Create VMs as there are not enough vms to check host ha")
 | |
|             no_vm_req = 5 - no_of_vms
 | |
|             if (no_vm_req > 0):
 | |
|                 self.logger.debug("Creating vms = {}".format(no_vm_req))
 | |
|                 self.vmlist = self.createVMs(listHost[0].id, no_vm_req, True)
 | |
|            
 | |
|         ha_host = listHost[1]
 | |
|         other_host = listHost[0]
 | |
|         if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
 | |
|             ha_host = listHost[0]
 | |
|             other_host = listHost[1]
 | |
|                 
 | |
|         self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
 | |
|             
 | |
|         hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostDown): 
 | |
|             raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
 | |
|             
 | |
|         hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Down", ha_host.id)
 | |
|         #the test could have failed here but we will try our best to get host back in consistent state
 | |
|             
 | |
|         no_of_vms = self.noOfVMsOnHost(ha_host.id)
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|             #
 | |
|         hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostUp): 
 | |
|             self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
 | |
|             
 | |
|         
 | |
|         hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
 | |
|         
 | |
|         if not(hostDownInCloudstack): 
 | |
|             raise self.fail("Host is not down %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|         if not(hostUpInCloudstack): 
 | |
|             raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|             
 | |
|         return
 | |
|     
 | |
|     
 | |
|     
 | |
|     @attr(
 | |
|         tags=[
 | |
|             "advanced",
 | |
|             "advancedns",
 | |
|             "smoke",
 | |
|             "basic",
 | |
|             "eip",
 | |
|             "sg"],
 | |
|         required_hardware="true")
 | |
|     def test_03_host_ha_with_only_local_storage(self):
 | |
|     
 | |
|         if not(self.isOnlyLocalStorageAvailable()):
 | |
|             raise unittest.SkipTest("Skipping this test as this is for Local storage only.");
 | |
|             return
 | |
|         
 | |
|         listHost = Host.list(
 | |
|             self.apiclient,
 | |
|             type='Routing',
 | |
|             zoneid=self.zone.id,
 | |
|             podid=self.pod.id,
 | |
|         )
 | |
|         for host in listHost:
 | |
|             self.logger.debug('Hypervisor = {}'.format(host.id))
 | |
|             
 | |
|                   
 | |
|         if len(listHost) != 2:
 | |
|             self.logger.debug("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             raise unittest.SkipTest("Host HA can be tested with two host only %s, found" % len(listHost));
 | |
|             return
 | |
|         
 | |
|         no_of_vms = self.noOfVMsOnHost(listHost[0].id)
 | |
|         
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(listHost[1].id)
 | |
|         
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|         
 | |
|         if no_of_vms < 5:
 | |
|             self.logger.debug("test_03: Create VMs as there are not enough vms to check host ha")
 | |
|             no_vm_req = 5 - no_of_vms
 | |
|             if (no_vm_req > 0):
 | |
|                 self.logger.debug("Creating vms = {}".format(no_vm_req))
 | |
|                 self.vmlist = self.createVMs(listHost[0].id, no_vm_req, True)
 | |
|            
 | |
|         ha_host = listHost[1]
 | |
|         other_host = listHost[0]
 | |
|         if self.noOfVMsOnHost(listHost[0].id) > self.noOfVMsOnHost(listHost[1].id):
 | |
|             ha_host = listHost[0]
 | |
|             other_host = listHost[1]
 | |
|                 
 | |
|         self.disconnectHostfromNetwork(ha_host.ipaddress, 400)
 | |
|             
 | |
|         hostDown = wait_until(10, 10, self.checkHostDown, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostDown): 
 | |
|             raise unittest.SkipTest("Host %s is not down, cannot proceed with test" % (ha_host.ipaddress))
 | |
|             
 | |
|         hostDownInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Alert", ha_host.id)
 | |
|         #the test could have failed here but we will try our best to get host back in consistent state
 | |
|             
 | |
|         no_of_vms = self.noOfVMsOnHost(ha_host.id)
 | |
|         no_of_vms = no_of_vms + self.noOfVMsOnHost(other_host.id)
 | |
|         self.logger.debug("Number of VMS on hosts = %s" % no_of_vms)
 | |
|             #
 | |
|         hostUp = wait_until(10, 10, self.checkHostUp, other_host.ipaddress, ha_host.ipaddress)
 | |
|         if not(hostUp): 
 | |
|             self.logger.debug("Host is down %s, though HA went fine, the environment is not consistent " % (ha_host.ipaddress))
 | |
|             
 | |
|         
 | |
|         hostUpInCloudstack = wait_until(40, 10, self.checkHostStateInCloudstack, "Up", ha_host.id)
 | |
|         
 | |
|         if not(hostDownInCloudstack): 
 | |
|             raise self.fail("Host is not in alert %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|         if not(hostUpInCloudstack): 
 | |
|             raise self.fail("Host is not up %s, in cloudstack so failing test " % (ha_host.ipaddress))
 | |
|             
 | |
|         return |