# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. import logging import random import SignedAPICall import time import XenAPI from solidfire.factory import ElementFactory from util import sf_util # All tests inherit from cloudstackTestCase from marvin.cloudstackTestCase import cloudstackTestCase # Import Integration Libraries # base - contains all resources as entities and defines create, delete, list operations on them from marvin.lib.base import Account, DiskOffering, ServiceOffering, Snapshot, StoragePool, Template, User, VirtualMachine, Volume # common - commonly used methods for all tests are listed here from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_virtual_machines, \ list_volumes # utils - utility classes for common cleanup, external library wrappers, etc. from marvin.lib.utils import cleanup_resources # Prerequisites: # Only one zone # Only one pod # Only one cluster # # Running the tests: # Change the "hypervisor_type" variable to control which hypervisor type to test. # If using XenServer, verify the "xen_server_hostname" variable is correct. # If using XenServer, change the "supports_cloning" variable to True or False as desired. # # Note: # If you do have more than one cluster, you might need to change this line: cls.cluster = list_clusters(cls.apiClient)[0] class TestData(): # constants account = "account" capacityBytes = "capacitybytes" capacityIops = "capacityiops" clusterId = "clusterId" computeOffering = "computeoffering" diskName = "diskname" diskOffering = "diskoffering" domainId = "domainId" hypervisor = "hypervisor" kvm = "kvm" login = "login" mvip = "mvip" one_GB_in_bytes = 1073741824 password = "password" port = "port" primaryStorage = "primarystorage" provider = "provider" scope = "scope" solidFire = "solidfire" storageTag = "SolidFire_SAN_1" tags = "tags" templateCacheNameKvm = "centos55-x86-64" templateCacheNameXenServer = "centos56-x86-64-xen" testAccount = "testaccount" url = "url" user = "user" username = "username" virtualMachine = "virtualmachine" virtualMachine2 = "virtualmachine2" volume_1 = "volume_1" volume_2 = "volume_2" xenServer = "xenserver" zoneId = "zoneId" # modify to control which hypervisor type to test hypervisor_type = kvm xen_server_hostname = "XenServer-6.5-1" def __init__(self): self.testdata = { TestData.solidFire: { TestData.mvip: "10.117.40.120", TestData.username: "admin", TestData.password: "admin", TestData.port: 443, TestData.url: "https://10.117.40.120:443" }, TestData.kvm: { TestData.username: "root", TestData.password: "solidfire" }, TestData.xenServer: { TestData.username: "root", TestData.password: "solidfire" }, TestData.account: { "email": "test@test.com", "firstname": "John", "lastname": "Doe", "username": "test", "password": "test" }, TestData.testAccount: { "email": "test2@test2.com", "firstname": "Jane", "lastname": "Doe", "username": "test2", "password": "test" }, TestData.user: { "email": "user@test.com", "firstname": "Jane", "lastname": "Doe", "username": "testuser", "password": "password" }, TestData.primaryStorage: { "name": "SolidFire-%d" % random.randint(0, 100), TestData.scope: "ZONE", "url": "MVIP=10.117.40.120;SVIP=10.117.41.120;" + "clusterAdminUsername=admin;clusterAdminPassword=admin;" + "clusterDefaultMinIops=10000;clusterDefaultMaxIops=15000;" + "clusterDefaultBurstIopsPercentOfMaxIops=1.5;", TestData.provider: "SolidFire", TestData.tags: TestData.storageTag, TestData.capacityIops: 4500000, TestData.capacityBytes: 2251799813685248, TestData.hypervisor: "Any" }, TestData.virtualMachine: { "name": "TestVM", "displayname": "Test VM" }, TestData.virtualMachine2: { "name": "TestVM2", "displayname": "Test VM 2" }, TestData.computeOffering: { "name": "SF_CO_1", "displaytext": "SF_CO_1 (Min IOPS = 10,000; Max IOPS = 15,000)", "cpunumber": 1, "cpuspeed": 100, "memory": 128, "storagetype": "shared", "customizediops": False, "miniops": "10000", "maxiops": "15000", "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag }, TestData.diskOffering: { "name": "SF_DO_1", "displaytext": "SF_DO_1 (Min IOPS = 300; Max IOPS = 500)", "disksize": 128, "customizediops": False, "miniops": 300, "maxiops": 500, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" }, TestData.volume_1: { TestData.diskName: "test-volume", }, TestData.volume_2: { TestData.diskName: "test-volume-2", }, TestData.zoneId: 1, TestData.clusterId: 1, TestData.domainId: 1, TestData.url: "10.117.40.114" } class TestVolumes(cloudstackTestCase): _should_only_be_one_vm_in_list_err_msg = "There should only be one VM in this list." _should_only_be_one_volume_in_list_err_msg = "There should only be one volume in this list." _should_only_be_one_host_in_list_err_msg = "There should only be one host in this list." _should_only_be_two_volumes_in_list_err_msg = "There should only be two volumes in this list." _sf_account_id_should_be_non_zero_int_err_msg = "The SolidFire account ID should be a non-zero integer." _volume_size_should_be_non_zero_int_err_msg = "The SolidFire volume size should be a non-zero integer." _volume_vm_id_and_vm_id_do_not_match_err_msg = "The volume's VM ID and the VM's ID do not match." _vm_not_in_running_state_err_msg = "The VM is not in the 'Running' state." _vm_not_in_stopped_state_err_msg = "The VM is not in the 'Stopped' state." _volume_response_should_not_be_zero_err_msg = "The length of the response for the SolidFire-volume query should not be zero." _volume_should_not_be_in_a_vag = "The volume should not be in a volume access group." _volume_size_not_an_int = "'volume_size_in_GB' is not of type 'int'" _only_data_volumes_err_msg = "Only data volumes can be resized via a new disk offering." _to_change_volume_size_err_msg = "To change a volume's size without providing a new disk offering, its current " \ "disk offering must be customizable or it must be a root volume (if providing a disk offering, make sure it is " \ "different from the current disk offering)." _min_iops_err_msg = "The current disk offering does not support customization of the 'Min IOPS' parameter." _this_kind_of_disk_err_msg = "This kind of KVM disk cannot be resized while it is connected to a VM that's not in the Stopped state." _template_creation_did_not_fail_err_msg = "The template creation did not fail (as expected)." _volume_resize_did_not_fail_err_msg = "The volume resize did not fail (as expected)." _volume_attached_to_non_stopped_vm_err_msg = "volume is attached to a non-stopped VM" @classmethod def setUpClass(cls): # Set up API client testclient = super(TestVolumes, cls).getClsTestClient() cls.apiClient = testclient.getApiClient() cls.configData = testclient.getParsedTestDataConfig() cls.dbConnection = testclient.getDbConnection() cls.testdata = TestData().testdata cls._handle_supports_cloning() cls._connect_to_hypervisor() # Set up SolidFire connection solidfire = cls.testdata[TestData.solidFire] cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password]) # Get Resources from Cloud Infrastructure cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId]) cls.cluster = list_clusters(cls.apiClient)[0] cls.template = get_template(cls.apiClient, cls.zone.id, hypervisor=TestData.hypervisor_type) cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId]) # Create test account cls.account = Account.create( cls.apiClient, cls.testdata["account"], admin=1 ) # Set up connection to make customized API calls cls.user = User.create( cls.apiClient, cls.testdata["user"], account=cls.account.name, domainid=cls.domain.id ) url = cls.testdata[TestData.url] api_url = "http://" + url + ":8080/client/api" userkeys = User.registerUserKeys(cls.apiClient, cls.user.id) cls.cs_api = SignedAPICall.CloudStack(api_url, userkeys.apikey, userkeys.secretkey) primarystorage = cls.testdata[TestData.primaryStorage] cls.primary_storage = StoragePool.create( cls.apiClient, primarystorage, scope=primarystorage[TestData.scope], zoneid=cls.zone.id, provider=primarystorage[TestData.provider], tags=primarystorage[TestData.tags], capacityiops=primarystorage[TestData.capacityIops], capacitybytes=primarystorage[TestData.capacityBytes], hypervisor=primarystorage[TestData.hypervisor] ) cls.compute_offering = ServiceOffering.create( cls.apiClient, cls.testdata[TestData.computeOffering] ) cls.disk_offering = DiskOffering.create( cls.apiClient, cls.testdata[TestData.diskOffering] ) # Create VM and volume for tests cls.virtual_machine = VirtualMachine.create( cls.apiClient, cls.testdata[TestData.virtualMachine], accountid=cls.account.name, zoneid=cls.zone.id, serviceofferingid=cls.compute_offering.id, templateid=cls.template.id, domainid=cls.domain.id, startvm=False ) TestVolumes._start_vm(cls.virtual_machine) cls.volume = Volume.create( cls.apiClient, cls.testdata[TestData.volume_1], account=cls.account.name, domainid=cls.domain.id, zoneid=cls.zone.id, diskofferingid=cls.disk_offering.id ) # Resources that are to be destroyed cls._cleanup = [ cls.volume, cls.virtual_machine, cls.compute_offering, cls.disk_offering, cls.user, cls.account ] @classmethod def tearDownClass(cls): try: cleanup_resources(cls.apiClient, cls._cleanup) cls.primary_storage.delete(cls.apiClient) sf_util.purge_solidfire_volumes(cls.sfe) except Exception as e: logging.debug("Exception in tearDownClass(cls): %s" % e) def setUp(self): self.attached = False self.cleanup = [] def tearDown(self): if self.attached: self.virtual_machine.detach_volume(self.apiClient, self.volume) cleanup_resources(self.apiClient, self.cleanup) def test_00_check_template_cache(self): if self._supports_cloning == False: return sf_volumes = self._get_active_sf_volumes() sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, self._get_template_cache_name(), self) self.assertEqual( len(sf_volume.volume_access_groups), 0, "The volume should not be in a VAG." ) sf_account_id = sf_volume.account_id sf_account = self.sfe.get_account_by_id(sf_account_id).account sf_account_name = sf_account.username self.assertEqual( sf_account_name.endswith("_1"), True, "The template cache volume's account does not end with '_1'." ) def test_01_attach_new_volume_to_stopped_VM(self): '''Attach a volume to a stopped virtual machine, then start VM''' self.virtual_machine.stop(self.apiClient) new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_2], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) self.cleanup.append(new_volume) sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) new_volume = self.virtual_machine.attach_volume( self.apiClient, new_volume ) newvolume = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) TestVolumes._start_vm(self.virtual_machine) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( newvolume.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), "running", TestVolumes._vm_not_in_running_state_err_msg ) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, new_volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_iscsi_name = sf_util.get_iqn(self.cs_api, new_volume, self) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, newvolume.name, self) sf_util.check_size_and_iops(sf_volume, newvolume, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) # Detach volume new_volume = self.virtual_machine.detach_volume( self.apiClient, new_volume ) def test_02_attach_detach_attach_volume(self): '''Attach, detach, and attach volume to a running VM''' TestVolumes._start_vm(self.virtual_machine) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) ####################################### ####################################### # STEP 1: Attach volume to running VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) ######################################### ######################################### # STEP 2: Detach volume from running VM # ######################################### ######################################### self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, "The volume should not be in a VAG." ) self._check_host_side(sf_iscsi_name, vm.hostid, False) ####################################### ####################################### # STEP 3: Attach volume to running VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) def test_03_attached_volume_reboot_VM(self): '''Attach volume to running VM, then reboot.''' TestVolumes._start_vm(self.virtual_machine) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) ####################################### ####################################### # STEP 1: Attach volume to running VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) ####################################### ####################################### # STEP 2: Reboot VM with attached vol # ####################################### ####################################### TestVolumes._reboot_vm(self.virtual_machine) vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) def test_04_detach_volume_reboot(self): '''Detach volume from a running VM, then reboot.''' TestVolumes._start_vm(self.virtual_machine) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) ####################################### ####################################### # STEP 1: Attach volume to running VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) ######################################### ######################################### # STEP 2: Detach volume from running VM # ######################################### ######################################### self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, vm.hostid, False) ####################################### ####################################### # STEP 3: Reboot VM with detached vol # ####################################### ####################################### self.virtual_machine.reboot(self.apiClient) vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, vm.hostid, False) def test_05_detach_vol_stopped_VM_start(self): '''Detach volume from a stopped VM, then start.''' TestVolumes._start_vm(self.virtual_machine) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) ####################################### ####################################### # STEP 1: Attach volume to running VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) hostid = vm.hostid ######################################### ######################################### # STEP 2: Detach volume from stopped VM # ######################################### ######################################### self.virtual_machine.stop(self.apiClient) self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'stopped', TestVolumes._vm_not_in_stopped_state_err_msg ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, hostid, False) ####################################### ####################################### # STEP 3: Start VM with detached vol # ####################################### ####################################### TestVolumes._start_vm(self.virtual_machine) vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, vm.hostid, False) def test_06_attach_volume_to_stopped_VM(self): '''Attach a volume to a stopped virtual machine, then start VM''' self.virtual_machine.stop(self.apiClient) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) ####################################### ####################################### # STEP 1: Attach volume to stopped VM # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'stopped', TestVolumes._vm_not_in_stopped_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) if TestData.hypervisor_type == TestData.kvm: self._check_host_side(sf_iscsi_name, None, False) elif TestData.hypervisor_type == TestData.xenServer: self._check_host_side(sf_iscsi_name) TestVolumes._start_vm(self.virtual_machine) vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) def test_07_destroy_expunge_VM_with_volume(self): '''Destroy and expunge VM with attached volume''' ####################################### ####################################### # STEP 1: Create VM and attach volume # ####################################### ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine2], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestVolumes._start_vm(test_virtual_machine) self.volume = test_virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestVolumes._vm_not_in_running_state_err_msg ) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) hostid = vm.hostid ####################################### ####################################### # STEP 2: Destroy and Expunge VM # ####################################### ####################################### test_virtual_machine.delete(self.apiClient, True) self.attached = False vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) self.assertEqual( vol.virtualmachineid, None, "Check if attached to virtual machine" ) self.assertEqual( vol.vmname, None, "Check if VM was expunged" ) list_virtual_machine_response = list_virtual_machines( self.apiClient, id=test_virtual_machine.id ) self.assertEqual( list_virtual_machine_response, None, "Check if VM was actually expunged" ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, hostid, False) def test_08_delete_volume_was_attached(self): '''Delete volume that was attached to a VM and is detached now''' TestVolumes._start_vm(self.virtual_machine) ####################################### ####################################### # STEP 1: Create vol and attach to VM # ####################################### ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_2], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) volume_to_delete_later = new_volume sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) new_volume = self.virtual_machine.attach_volume( self.apiClient, new_volume ) vol = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, new_volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_iscsi_name = sf_util.get_iqn(self.cs_api, new_volume, self) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_util.check_vag(sf_volume, sf_vag_id, self) self._check_host_side(sf_iscsi_name, vm.hostid) ####################################### ####################################### # STEP 2: Detach and delete volume # ####################################### ####################################### new_volume = self.virtual_machine.detach_volume( self.apiClient, new_volume ) vol = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, None, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) self.assertEqual( len(sf_volume.volume_access_groups), 0, TestVolumes._volume_should_not_be_in_a_vag ) self._check_host_side(sf_iscsi_name, vm.hostid, False) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self, False) def test_09_attach_volumes_multiple_accounts(self): '''Attach a data disk to a VM in one account and attach another data disk to a VM in another account''' TestVolumes._start_vm(self.virtual_machine) ####################################### ####################################### # STEP 1: Create account, VM, and vol # ####################################### ####################################### test_account = Account.create( self.apiClient, self.testdata[TestData.testAccount], admin=1 ) self.cleanup.append(test_account) test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine2], accountid=test_account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestVolumes._start_vm(test_virtual_machine) test_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_2], zoneid=self.zone.id, account=test_account.name, domainid=self.domain.id, diskofferingid=self.disk_offering.id ) sf_util.check_and_get_cs_volume(self, test_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) ####################################### ####################################### # STEP 2: Attach volumes to VMs # ####################################### ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vol.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) test_volume = test_virtual_machine.attach_volume( self.apiClient, test_volume ) test_vol = sf_util.check_and_get_cs_volume(self, test_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self) test_vm = self._get_vm(test_virtual_machine.id) self.assertEqual( test_vol.virtualmachineid, test_vm.id, "Check if attached to virtual machine of other acct" ) self.assertEqual( test_vm.state.lower(), 'running', str(test_vm.state) ) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, vol, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) self._check_host_side(sf_iscsi_name, vm.hostid) sf_util.check_vag(sf_volume, sf_vag_id, self) sf_test_account_id = sf_util.get_sf_account_id(self.cs_api, test_account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_test_volumes = self._get_active_sf_volumes(sf_test_account_id) sf_test_volume = sf_util.check_and_get_sf_volume(sf_test_volumes, test_vol.name, self) sf_test_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, test_vol, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_test_volume_size) sf_util.check_size_and_iops(sf_test_volume, test_vol, sf_test_volume_size, self) sf_test_iscsi_name = sf_util.get_iqn(self.cs_api, test_volume, self) self._check_host_side(sf_test_iscsi_name, test_vm.hostid) sf_util.check_vag(sf_test_volume, sf_vag_id, self) def test_10_attach_more_than_one_disk_to_VM(self): '''Attach more than one disk to a VM''' TestVolumes._start_vm(self.virtual_machine) volume_2 = Volume.create( self.apiClient, self.testdata[TestData.volume_2], zoneid=self.zone.id, account=self.account.name, domainid=self.domain.id, diskofferingid=self.disk_offering.id ) self.cleanup.append(volume_2) sf_util.check_and_get_cs_volume(self, volume_2.id, self.testdata[TestData.volume_2][TestData.diskName], self) ####################################### ####################################### # Step 1: Attach volumes to VM # ####################################### ####################################### self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self) self.virtual_machine.attach_volume( self.apiClient, volume_2 ) vol_2 = sf_util.check_and_get_cs_volume(self, volume_2.id, self.testdata[TestData.volume_2][TestData.diskName], self) sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size) sf_volume_2_size = sf_util.get_volume_size_with_hsr(self.cs_api, volume_2, self) self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_2_size) sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self) sf_volumes = self._get_active_sf_volumes(sf_account_id) sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self) sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self) sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self) vm = self._get_vm(self.virtual_machine.id) self._check_host_side(sf_iscsi_name, vm.hostid) sf_util.check_vag(sf_volume, sf_vag_id, self) sf_volume_2 = sf_util.check_and_get_sf_volume(sf_volumes, vol_2.name, self) sf_util.check_size_and_iops(sf_volume_2, vol_2, sf_volume_2_size, self) sf_iscsi_name_2 = sf_util.get_iqn(self.cs_api, volume_2, self) self._check_host_side(sf_iscsi_name_2, vm.hostid) sf_util.check_vag(sf_volume_2, sf_vag_id, self) self.virtual_machine.detach_volume(self.apiClient, volume_2) def test_11_template_from_volume(self): if TestData.hypervisor_type != TestData.kvm: return list_volumes_response = list_volumes( self.apiClient, virtualmachineid=self.virtual_machine.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) vm_1_root_volume = list_volumes_response[0] services = {"displaytext": "Template-1", "name": "Template-1-name", "ostypeid": self.template.ostypeid, "ispublic": "true"} try: Template.create_from_volume(self.apiClient, vm_1_root_volume, services) raise Exception(TestVolumes._template_creation_did_not_fail_err_msg) except Exception as e: if TestVolumes._volume_attached_to_non_stopped_vm_err_msg not in e.errorMsg: raise self.virtual_machine.stop(self.apiClient) template = Template.create_from_volume(self.apiClient, vm_1_root_volume, services) self.cleanup.append(template) vol_snap = Snapshot.create(self.apiClient, volume_id=vm_1_root_volume.id) self.cleanup.append(vol_snap) TestVolumes._start_vm(self.virtual_machine) self._create_vm_using_template_and_destroy_vm(template) services = {"diskname": "Vol-1", "zoneid": self.testdata[TestData.zoneId], "ispublic": True} volume_created_from_snapshot = Volume.create_from_snapshot(self.apiClient, vol_snap.id, services, account=self.account.name, domainid=self.domain.id) self.cleanup.append(volume_created_from_snapshot) services = {"displaytext": "Template-2", "name": "Template-2-name", "ostypeid": self.template.ostypeid, "ispublic": "true"} template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services) self.cleanup.append(template) self._create_vm_using_template_and_destroy_vm(template) volume_created_from_snapshot = self.virtual_machine.attach_volume( self.apiClient, volume_created_from_snapshot ) services = {"displaytext": "Template-3", "name": "Template-3-name", "ostypeid": self.template.ostypeid, "ispublic": "true"} try: Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services) raise Exception(TestVolumes._template_creation_did_not_fail_err_msg) except Exception as e: if TestVolumes._volume_attached_to_non_stopped_vm_err_msg not in e.errorMsg: raise self.virtual_machine.stop(self.apiClient) template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services) self.cleanup.append(template) volume_created_from_snapshot = self.virtual_machine.detach_volume( self.apiClient, volume_created_from_snapshot ) TestVolumes._start_vm(self.virtual_machine) self._create_vm_using_template_and_destroy_vm(template) services = {"displaytext": "Template-4", "name": "Template-4-name", "ostypeid": self.template.ostypeid, "ispublic": "true"} template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services) self.cleanup.append(template) self._create_vm_using_template_and_destroy_vm(template) def test_12_resize_volume_on_running_vm(self): if TestData.hypervisor_type != TestData.kvm: return test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine2], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=True ) self.cleanup.append(test_virtual_machine) list_volumes_response = list_volumes( self.apiClient, virtualmachineid=test_virtual_machine.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) test_vm_root_volume = Volume(list_volumes_response[0].__dict__) self._handle_root_volume_with_started_vm(test_vm_root_volume) volume_name = { TestData.diskName: "test-volume-a", } test_vm_data_volume = Volume.create( self.apiClient, volume_name, account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) self.cleanup.append(test_vm_data_volume) test_vm_data_volume = test_virtual_machine.attach_volume( self.apiClient, test_vm_data_volume ) test_vm_data_volume = Volume(test_vm_data_volume.__dict__) self._handle_data_volume_with_started_vm(test_vm_data_volume) custom_disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Custom)", "customized": True, "customizediops": True, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } custom_disk_offering = DiskOffering.create( self.apiClient, custom_disk_offering, custom=True ) self.cleanup.append(custom_disk_offering) services = { TestData.diskName: "test-volume-custom-a", "customdisksize": 100, "customminiops": 1000, "custommaxiops": 2000, "zoneid": self.testdata[TestData.zoneId] } test_vm_data_volume = Volume.create_custom_disk( self.apiClient, services, account=self.account.name, domainid=self.domain.id, diskofferingid=custom_disk_offering.id ) self.cleanup.append(test_vm_data_volume) test_vm_data_volume = test_virtual_machine.attach_volume( self.apiClient, test_vm_data_volume ) test_vm_data_volume = Volume(test_vm_data_volume.__dict__) self._handle_custom_data_volume_with_started_vm(test_vm_data_volume) def test_13_resize_volume_on_stopped_vm(self): if TestData.hypervisor_type != TestData.kvm: return test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine2], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) self.cleanup.append(test_virtual_machine) list_volumes_response = list_volumes( self.apiClient, virtualmachineid=test_virtual_machine.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) test_vm_root_volume = Volume(list_volumes_response[0].__dict__) err_msg = "Check if SF volume was created in correct account" try: # This should fail because there should not be an equivalent SolidFire volume in the cluster yet. self._verify_volume(test_vm_root_volume) raise Exception("The volume verification did not fail (as expected).") except Exception as e: if err_msg not in str(e): raise # Starting the up the should create its root disk on the SolidFire cluster. test_virtual_machine.start(self.apiClient) test_virtual_machine.stop(self.apiClient) self._handle_root_volume_with_stopped_vm(test_vm_root_volume) volume_name = { TestData.diskName: "test-volume-a", } test_vm_data_volume = Volume.create( self.apiClient, volume_name, account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) self.cleanup.append(test_vm_data_volume) test_vm_data_volume = test_virtual_machine.attach_volume( self.apiClient, test_vm_data_volume ) test_vm_data_volume = Volume(test_vm_data_volume.__dict__) self._handle_data_volume_with_stopped_vm(test_vm_data_volume) custom_disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Custom)", "customized": True, "customizediops": True, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } custom_disk_offering = DiskOffering.create( self.apiClient, custom_disk_offering, custom=True ) self.cleanup.append(custom_disk_offering) services = { TestData.diskName: "test-volume-custom-a", "customdisksize": 100, "customminiops": 1000, "custommaxiops": 2000, "zoneid": self.testdata[TestData.zoneId] } test_vm_data_volume = Volume.create_custom_disk( self.apiClient, services, account=self.account.name, domainid=self.domain.id, diskofferingid=custom_disk_offering.id ) self.cleanup.append(test_vm_data_volume) test_vm_data_volume = test_virtual_machine.attach_volume( self.apiClient, test_vm_data_volume ) test_vm_data_volume = Volume(test_vm_data_volume.__dict__) self._handle_custom_data_volume_with_stopped_vm(test_vm_data_volume) test_vm_data_volume = test_virtual_machine.detach_volume( self.apiClient, test_vm_data_volume ) test_vm_data_volume = Volume(test_vm_data_volume.__dict__) self._handle_custom_data_volume_with_stopped_vm(test_vm_data_volume) def _handle_root_volume_with_started_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 sf_volume_size = self._get_sf_volume(volume.name).total_size try: # Try to change the size and IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the size of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": new_size, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size and IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._only_data_volumes_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_B", "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")", "disksize": new_size, "customizediops": False, "miniops": volume.miniops, "maxiops": volume.maxiops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size of a volume attached to a running VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._only_data_volumes_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_C", "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": volume_size_in_GB, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_min_iops = new_min_iops + 10 new_max_iops = new_max_iops + 20 volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _handle_root_volume_with_stopped_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_size = new_size + 10 volume.resize(self.apiClient, size=new_size) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_min_iops = new_min_iops + 100 new_max_iops = new_max_iops + 200 volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_size = new_size + 10 new_min_iops = new_min_iops + 100 new_max_iops = new_max_iops + 200 disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": new_size, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size and IOPS of a volume attached to a stopped VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._only_data_volumes_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_B", "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")", "disksize": new_size, "customizediops": False, "miniops": volume.miniops, "maxiops": volume.maxiops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size of a volume attached to a stopped VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._only_data_volumes_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) disk_offering = { "name": "SF_DO_C", "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": volume_size_in_GB, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _handle_data_volume_with_started_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 sf_volume_size = self._get_sf_volume(volume.name).total_size try: # Try to change the size and IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._to_change_volume_size_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the size of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._to_change_volume_size_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": new_size, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size and IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_B", "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")", "disksize": new_size, "customizediops": False, "miniops": volume.miniops, "maxiops": volume.maxiops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) try: # Try to change the size of a volume attached to a running VM (should fail). volume.resize(self.apiClient, diskofferingid=disk_offering.id) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._min_iops_err_msg not in str(e): raise disk_offering.delete(self.apiClient) self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_C", "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": volume_size_in_GB, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _handle_data_volume_with_stopped_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 sf_volume_size = self._get_sf_volume(volume.name).total_size try: # Try to change the size and IOPS of a volume attached to a stopped VM (should fail). volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._to_change_volume_size_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the size of a volume attached to a stopped VM (should fail). volume.resize(self.apiClient, size=new_size) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._to_change_volume_size_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the IOPS of a volume attached to a stopped VM (should fail). volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._min_iops_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) disk_offering = { "name": "SF_DO_A", "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": new_size, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_size = new_size + 10 disk_offering = { "name": "SF_DO_B", "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")", "disksize": new_size, "customizediops": False, "miniops": volume.miniops, "maxiops": volume.maxiops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_min_iops = new_min_iops + 100 new_max_iops = new_max_iops + 200 disk_offering = { "name": "SF_DO_C", "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")", "disksize": volume_size_in_GB, "customizediops": False, "miniops": new_min_iops, "maxiops": new_max_iops, "hypervisorsnapshotreserve": 200, TestData.tags: TestData.storageTag, "storagetype": "shared" } disk_offering = DiskOffering.create( self.apiClient, disk_offering ) volume.resize(self.apiClient, diskofferingid=disk_offering.id) disk_offering.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _handle_custom_data_volume_with_started_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 sf_volume_size = self._get_sf_volume(volume.name).total_size try: # Try to change the size and IOPS of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) try: # Try to change the size of a volume attached to a running VM (should fail). volume.resize(self.apiClient, size=new_size) raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg) except Exception as e: if TestVolumes._this_kind_of_disk_err_msg not in str(e): raise self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _handle_custom_data_volume_with_stopped_vm(self, volume): self._verify_volume(volume) volume_size_in_GB = volume.size / TestData.one_GB_in_bytes self.assertTrue( type(volume_size_in_GB) == int, TestVolumes._volume_size_not_an_int ) new_size = volume_size_in_GB + 10 new_min_iops = volume.miniops + 100 new_max_iops = volume.maxiops + 200 volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) new_size = new_size + 10 volume.resize(self.apiClient, size=new_size) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) sf_volume_size = self._get_sf_volume(volume.name).total_size self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size) new_min_iops = volume.miniops + 10 new_max_iops = volume.maxiops + 20 volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops) list_volumes_response = list_volumes( self.apiClient, id=volume.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) volume = Volume(list_volumes_response[0].__dict__) self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size) def _get_sf_volume(self, volume_name): sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, "The SolidFire account ID should be a non-zero integer.") sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id) self.assertNotEqual( len(sf_volumes), 0, "The length of the response for the SolidFire-volume query should not be zero." ) return sf_util.check_and_get_sf_volume(sf_volumes, volume_name, self) def _verify_volume(self, cs_volume, expected_min_iops=None, expected_max_iops=None, expected_size=None): sf_volume = self._get_sf_volume(cs_volume.name) sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, cs_volume, self) sf_util.check_size_and_iops(sf_volume, cs_volume, sf_volume_size, self) if expected_min_iops is not None: self.assertEqual( cs_volume.miniops, expected_min_iops, "Unexpected Min IOPS value (CloudStack volume has " + str(cs_volume.miniops) + "; expected " + str(expected_min_iops) + ")" ) if expected_max_iops is not None: self.assertEqual( cs_volume.maxiops, expected_max_iops, "Unexpected Max IOPS value (CloudStack volume has " + str(cs_volume.maxiops) + "; expected " + str(expected_max_iops) + ")" ) if expected_size is not None: self.assertEqual( sf_volume_size, expected_size, "Unexpected size value (CloudStack volume (with HSR) has " + str(sf_volume_size) + "; expected " + str(expected_size) + ")" ) def _create_vm_using_template_and_destroy_vm(self, template): vm_name = "VM-%d" % random.randint(0, 100) virtual_machine_dict = {"name": vm_name, "displayname": vm_name} virtual_machine = VirtualMachine.create( self.apiClient, virtual_machine_dict, accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=template.id, domainid=self.domain.id, startvm=True ) list_volumes_response = list_volumes( self.apiClient, virtualmachineid=virtual_machine.id, listall=True ) sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg) vm_root_volume = list_volumes_response[0] vm_root_volume_name = vm_root_volume.name sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg) # Get volume information from SolidFire cluster sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id) sf_util.get_sf_volume_by_name(self, sf_volumes, vm_root_volume_name) virtual_machine.delete(self.apiClient, True) def _verify_hsr(self, cs_volume_size_in_gb, hsr, sf_volume_size_in_bytes): cs_volume_size_including_hsr_in_bytes = self._get_cs_volume_size_including_hsr_in_bytes(cs_volume_size_in_gb, hsr) self.assertTrue( cs_volume_size_including_hsr_in_bytes == sf_volume_size_in_bytes, "HSR does not add up correctly." ) def _get_cs_volume_size_including_hsr_in_bytes(self, cs_volume_size_in_gb, hsr): if TestData.hypervisor_type == TestData.kvm: return self._get_bytes_from_gb(cs_volume_size_in_gb) lowest_hsr = 10 if hsr < lowest_hsr: hsr = lowest_hsr return self._get_bytes_from_gb(cs_volume_size_in_gb + (cs_volume_size_in_gb * (hsr / 100))) def _get_bytes_from_gb(self, number_in_gb): return number_in_gb * 1024 * 1024 * 1024 def _get_vm(self, vm_id): list_vms_response = list_virtual_machines(self.apiClient, id=vm_id) sf_util.check_list(list_vms_response, 1, self, TestVolumes._should_only_be_one_vm_in_list_err_msg) return list_vms_response[0] def _check_xen_sr(self, xen_sr_name, should_exist=True): sf_util.check_xen_sr(xen_sr_name, self.xen_session, self, should_exist) def _get_active_sf_volumes(self, sf_account_id=None): sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id) self.assertNotEqual( len(sf_volumes), 0, TestVolumes._volume_response_should_not_be_zero_err_msg ) return sf_volumes def _get_template_cache_name(self): if TestData.hypervisor_type == TestData.kvm: return TestData.templateCacheNameKvm elif TestData.hypervisor_type == TestData.xenServer: return TestData.templateCacheNameXenServer self.assert_(False, "Invalid hypervisor type") def _get_modified_iscsi_name(self, sf_iscsi_name): sf_iscsi_name = sf_iscsi_name.replace("/", "") return sf_iscsi_name[:-1] def _check_host_side(self, sf_iscsi_name, vm_hostid=None, should_exist=True): if TestData.hypervisor_type == TestData.kvm: self._check_kvm_host_side(self._get_modified_iscsi_name(sf_iscsi_name), vm_hostid, should_exist) elif TestData.hypervisor_type == TestData.xenServer: self._check_xen_sr(sf_iscsi_name, should_exist) def _check_kvm_host_side(self, sf_iscsi_name, vm_hostid, should_exist=True): if vm_hostid is None: list_hosts_response = list_hosts( self.apiClient, type="Routing" ) else: list_hosts_response = list_hosts( self.apiClient, id=vm_hostid ) sf_util.check_list(list_hosts_response, 1, self, TestVolumes._should_only_be_one_host_in_list_err_msg) kvm_login = self.testdata[TestData.kvm] for cs_host in list_hosts_response: ssh_connection = sf_util.get_ssh_connection(cs_host.ipaddress, kvm_login[TestData.username], kvm_login[TestData.password]) stdout = ssh_connection.exec_command("ls /dev/disk/by-path | grep " + sf_iscsi_name)[1] result = stdout.read() ssh_connection.close() if should_exist: self.assertFalse(result is None, "Unable to locate 'by-path' field on the KVM host (None)") self.assertFalse(len(result.strip()) <= len(sf_iscsi_name), "Unable to locate the 'by-path' field on the KVM host (Zero-length string)") else: self.assertTrue(result is None or len(result.strip()) == 0, "Found the 'by-path' field on the KVM host, but did not expect to") @classmethod def _start_vm(cls, vm): vm_for_check = list_virtual_machines( cls.apiClient, id=vm.id )[0] if vm_for_check.state == VirtualMachine.STOPPED: vm.start(cls.apiClient) # Libvirt appears to have an issue detaching a volume from a VM while the VM is booting up. # The XML sent to update the VM seems correct, but it doesn't appear to update the XML that describes the VM. # For KVM, just give it 90 seconds to boot up. if TestData.hypervisor_type == TestData.kvm: time.sleep(90) @classmethod def _reboot_vm(cls, vm): vm.reboot(cls.apiClient) # Libvirt appears to have an issue detaching a volume from a VM while the VM is booting up. # The XML sent to update the VM seems correct, but it doesn't appear to update the XML that describes the VM. # For KVM, just give it 90 seconds to boot up. if TestData.hypervisor_type == TestData.kvm: time.sleep(90) @classmethod def _handle_supports_cloning(cls): if TestData.hypervisor_type == TestData.kvm: cls._supports_cloning = True elif TestData.hypervisor_type == TestData.xenServer: # For XenServer, it is OK to set this to True or False depending on what you'd like tested cls._supports_cloning = True sf_util.set_supports_resign(cls._supports_cloning, cls.dbConnection) @classmethod def _connect_to_hypervisor(cls): if TestData.hypervisor_type == TestData.kvm: pass elif TestData.hypervisor_type == TestData.xenServer: host_ip = "https://" + \ list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId], name=TestData.xen_server_hostname)[0].ipaddress cls.xen_session = XenAPI.Session(host_ip) xen_server = cls.testdata[TestData.xenServer] cls.xen_session.xenapi.login_with_password(xen_server[TestData.username], xen_server[TestData.password])