# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. import logging import random import time # All tests inherit from cloudstackTestCase from marvin.cloudstackTestCase import cloudstackTestCase # Import Integration Libraries # base - contains all resources as entities and defines create, delete, list operations on them from marvin.lib.base import Account, DiskOffering, ServiceOffering, Snapshot, StoragePool, Template, User, VirtualMachine, Volume # common - commonly used methods for all tests are listed here from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_virtual_machines, \ list_volumes # utils - utility classes for common cleanup, external library wrappers, etc. from marvin.lib.utils import cleanup_resources, validateList from marvin.codes import PASS from nose.plugins.attrib import attr # Prerequisites: # Only one zone # Only one pod # Only one cluster # # One PowerFlex/ScaleIO storage pool for basic tests # Only KVM hypervisor is supported for PowerFlex/ScaleIO storage pool # KVM host(s) with ScaleIO Data Client (SDC) installed and connected to Metadata Manager (MDM) # # For volume migration tests, additional storage pool(s) are required # One PowerFlex/ScaleIO storage pool on the same ScaleIO storage cluster/instance # One PowerFlex/ScaleIO storage pool on different ScaleIO storage cluster/instance # class TestData(): # constants account = "account" clusterId = "clusterId" computeOffering = "computeoffering" diskName = "diskname" diskOffering = "diskoffering" diskOfferingSameInstance = "diskOfferingSameInstance" diskOfferingDistinctInstance = "diskOfferingDistinctInstance" domainId = "domainId" hypervisor = "hypervisor" kvm = "kvm" login = "login" gatewayip = "gatewayip" one_GB_in_bytes = 1073741824 password = "password" port = "port" primaryStorage = "primarystorage" primaryStorageSameInstance = "primaryStorageSameInstance" primaryStorageDistinctInstance = "primaryStorageDistinctInstance" provider = "provider" scope = "scope" powerFlex = "powerflex" storageTag = "pflex" storageTagSameInstance = "pflexsame" storageTagDistinctInstance = "pflexdiff" tags = "tags" templateCacheNameKvm = "centos55-x86-64" testAccount = "testaccount" url = "url" user = "user" username = "username" virtualMachine = "virtualmachine" virtualMachine2 = "virtualmachine2" virtualMachine3 = "virtualmachine3" virtualMachine4 = "virtualmachine4" volume_1 = "volume_1" volume_2 = "volume_2" volume_3 = "volume_3" volume_4 = "volume_4" kvm = "kvm" zoneId = "zoneId" migrationTests = "migrationTests" # hypervisor type to test hypervisor_type = kvm def __init__(self): self.testdata = { TestData.kvm: { TestData.username: "root", TestData.password: "P@ssword123" }, TestData.account: { "email": "test1@test1.com", "firstname": "John", "lastname": "Doe", "username": "test1", "password": "test" }, TestData.testAccount: { "email": "test2@test2.com", "firstname": "Jane", "lastname": "Doe", "username": "test2", "password": "test" }, TestData.user: { "email": "user@test1.com", "firstname": "Jane", "lastname": "Doe", "username": "test1user", "password": "password" }, TestData.primaryStorage: { "name": "PowerFlexPool-%d" % random.randint(0, 100), TestData.scope: "ZONE", "url": "powerflex://admin:P%40ssword123@10.10.4.141/cspool01", TestData.provider: "PowerFlex", TestData.tags: TestData.storageTag + "," + TestData.storageTagSameInstance + "," + TestData.storageTagDistinctInstance, TestData.hypervisor: "KVM" }, TestData.virtualMachine: { "name": "TestVM1", "displayname": "Test VM 1" }, TestData.virtualMachine2: { "name": "TestVM2", "displayname": "Test VM 2" }, TestData.virtualMachine3: { "name": "TestVM3", "displayname": "Test VM 3" }, TestData.virtualMachine4: { "name": "TestVM4", "displayname": "Test VM 4" }, TestData.computeOffering: { "name": "PowerFlex_Compute", "displaytext": "PowerFlex_Compute", "cpunumber": 1, "cpuspeed": 500, "memory": 512, "storagetype": "shared", TestData.tags: TestData.storageTag }, TestData.diskOffering: { "name": "PowerFlex_Disk", "displaytext": "PowerFlex_Disk", "disksize": 8, TestData.tags: TestData.storageTag, "storagetype": "shared" }, TestData.volume_1: { TestData.diskName: "test-volume-1", }, TestData.volume_2: { TestData.diskName: "test-volume-2", }, TestData.volume_3: { TestData.diskName: "test-volume-3", }, TestData.volume_4: { TestData.diskName: "test-volume-4", }, TestData.zoneId: 1, TestData.clusterId: 1, TestData.domainId: 1, TestData.url: "10.10.3.226", # for volume migration tests TestData.migrationTests: True, # PowerFlex/ScaleIO storage pool on the same ScaleIO storage instance TestData.primaryStorageSameInstance: { "name": "PowerFlexPool-%d" % random.randint(0, 100), TestData.scope: "ZONE", "url": "powerflex://admin:P%40ssword123@10.10.4.141/cspool02", TestData.provider: "PowerFlex", TestData.tags: TestData.storageTag + "," + TestData.storageTagSameInstance, TestData.hypervisor: "KVM" }, # PowerFlex/ScaleIO storage pool on different ScaleIO storage instance TestData.primaryStorageDistinctInstance: { "name": "PowerFlexPool-%d" % random.randint(0, 100), TestData.scope: "ZONE", "url": "powerflex://admin:P%40ssword123@10.10.4.194/cloudstackpool", TestData.provider: "PowerFlex", TestData.tags: TestData.storageTag + "," + TestData.storageTagDistinctInstance, TestData.hypervisor: "KVM" }, TestData.diskOfferingSameInstance: { "name": "PowerFlex_Disk_Same_Inst", "displaytext": "PowerFlex_Disk_Same_Inst", "disksize": 8, TestData.tags: TestData.storageTagSameInstance, "storagetype": "shared" }, TestData.diskOfferingDistinctInstance: { "name": "PowerFlex_Disk_Diff_Inst", "displaytext": "PowerFlex_Disk_Diff_Inst", "disksize": 8, TestData.tags: TestData.storageTagDistinctInstance, "storagetype": "shared" }, } class TestScaleIOVolumes(cloudstackTestCase): _volume_vm_id_and_vm_id_do_not_match_err_msg = "The volume's VM ID and the VM's ID do not match." _vm_not_in_running_state_err_msg = "The VM is not in the 'Running' state." _vm_not_in_stopped_state_err_msg = "The VM is not in the 'Stopped' state." @classmethod def setUpClass(cls): # Set up API client testclient = super(TestScaleIOVolumes, cls).getClsTestClient() cls.apiClient = testclient.getApiClient() cls.configData = testclient.getParsedTestDataConfig() cls.dbConnection = testclient.getDbConnection() cls.testdata = TestData().testdata # Get Resources from Cloud Infrastructure cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId]) cls.cluster = list_clusters(cls.apiClient)[0] cls.template = get_template(cls.apiClient, cls.zone.id, hypervisor=TestData.hypervisor_type) cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId]) # Create test account cls.account = Account.create( cls.apiClient, cls.testdata["account"], admin=1 ) # Set up connection to make customized API calls cls.user = User.create( cls.apiClient, cls.testdata["user"], account=cls.account.name, domainid=cls.domain.id ) url = cls.testdata[TestData.url] primarystorage = cls.testdata[TestData.primaryStorage] cls.primary_storage = StoragePool.create( cls.apiClient, primarystorage, scope=primarystorage[TestData.scope], zoneid=cls.zone.id, provider=primarystorage[TestData.provider], tags=primarystorage[TestData.tags], hypervisor=primarystorage[TestData.hypervisor] ) cls.compute_offering = ServiceOffering.create( cls.apiClient, cls.testdata[TestData.computeOffering] ) cls.disk_offering = DiskOffering.create( cls.apiClient, cls.testdata[TestData.diskOffering] ) if TestData.migrationTests: primarystorage_sameinst = cls.testdata[TestData.primaryStorageSameInstance] cls.primary_storage_same_inst = StoragePool.create( cls.apiClient, primarystorage_sameinst, scope=primarystorage_sameinst[TestData.scope], zoneid=cls.zone.id, provider=primarystorage_sameinst[TestData.provider], tags=primarystorage_sameinst[TestData.tags], hypervisor=primarystorage_sameinst[TestData.hypervisor] ) primarystorage_distinctinst = cls.testdata[TestData.primaryStorageDistinctInstance] cls.primary_storage_distinct_inst = StoragePool.create( cls.apiClient, primarystorage_distinctinst, scope=primarystorage_distinctinst[TestData.scope], zoneid=cls.zone.id, provider=primarystorage_distinctinst[TestData.provider], tags=primarystorage_distinctinst[TestData.tags], hypervisor=primarystorage_distinctinst[TestData.hypervisor] ) cls.disk_offering_same_inst = DiskOffering.create( cls.apiClient, cls.testdata[TestData.diskOfferingSameInstance] ) cls.disk_offering_distinct_inst = DiskOffering.create( cls.apiClient, cls.testdata[TestData.diskOfferingDistinctInstance] ) # Create VM and volume for tests cls.virtual_machine = VirtualMachine.create( cls.apiClient, cls.testdata[TestData.virtualMachine], accountid=cls.account.name, zoneid=cls.zone.id, serviceofferingid=cls.compute_offering.id, templateid=cls.template.id, domainid=cls.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(cls.virtual_machine) cls.volume = Volume.create( cls.apiClient, cls.testdata[TestData.volume_1], account=cls.account.name, domainid=cls.domain.id, zoneid=cls.zone.id, diskofferingid=cls.disk_offering.id ) # Resources that are to be destroyed cls._cleanup = [ cls.volume, cls.virtual_machine, cls.compute_offering, cls.disk_offering, cls.user, cls.account ] @classmethod def tearDownClass(cls): try: if TestData.migrationTests: cls._cleanup.append(cls.disk_offering_same_inst) cls._cleanup.append(cls.disk_offering_distinct_inst) cleanup_resources(cls.apiClient, cls._cleanup) cls.primary_storage.delete(cls.apiClient) if TestData.migrationTests: cls.primary_storage_same_inst.delete(cls.apiClient) cls.primary_storage_distinct_inst.delete(cls.apiClient) except Exception as e: logging.debug("Exception in tearDownClass(cls): %s" % e) def setUp(self): self.attached = False self.cleanup = [] def tearDown(self): if self.attached: self.virtual_machine.detach_volume(self.apiClient, self.volume) cleanup_resources(self.apiClient, self.cleanup) @attr(tags=['basic'], required_hardware=False) def test_01_create_vm_with_volume(self): '''Create VM with attached volume and expunge VM''' ####################################### # STEP 1: Create VM and attach volume # ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine2], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(test_virtual_machine) self.volume = test_virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(test_virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ####################################### # STEP 2: Destroy and Expunge VM # ####################################### test_virtual_machine.delete(self.apiClient, True) self.attached = False vol = self._get_volume(self.volume.id) self.assertEqual( vol.virtualmachineid, None, "Check if attached to virtual machine" ) self.assertEqual( vol.vmname, None, "Check if VM was expunged" ) list_virtual_machine_response = list_virtual_machines( self.apiClient, id=test_virtual_machine.id ) self.assertEqual( list_virtual_machine_response, None, "Check if VM was actually expunged" ) @attr(tags=['basic'], required_hardware=False) def test_02_attach_new_volume_to_stopped_vm(self): '''Attach a volume to a stopped virtual machine, then start VM''' self.virtual_machine.stop(self.apiClient) new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_2], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) self.cleanup.append(new_volume) new_volume = self.virtual_machine.attach_volume( self.apiClient, new_volume ) TestScaleIOVolumes._start_vm(self.virtual_machine) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vm.state.lower(), "running", TestScaleIOVolumes._vm_not_in_running_state_err_msg ) # Detach volume new_volume = self.virtual_machine.detach_volume( self.apiClient, new_volume ) self.assertEqual( new_volume.virtualmachineid, None, "The volume should not be attached to a VM." ) @attr(tags=['basic'], required_hardware=False) def test_03_attach_detach_attach_volume_to_vm(self): '''Attach, detach, and attach volume to a running VM''' TestScaleIOVolumes._start_vm(self.virtual_machine) ####################################### # STEP 1: Attach volume to running VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ######################################### # STEP 2: Detach volume from running VM # ######################################### self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 3: Attach volume to running VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) @attr(tags=['basic'], required_hardware=False) def test_04_detach_vol_stopped_vm_start(self): '''Detach volume from a stopped VM, then start.''' TestScaleIOVolumes._start_vm(self.virtual_machine) ####################################### # STEP 1: Attach volume to running VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ######################################### # STEP 2: Detach volume from stopped VM # ######################################### self.virtual_machine.stop(self.apiClient) self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'stopped', TestScaleIOVolumes._vm_not_in_stopped_state_err_msg ) ####################################### # STEP 3: Start VM with detached vol # ####################################### TestScaleIOVolumes._start_vm(self.virtual_machine) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) @attr(tags=['basic'], required_hardware=False) def test_05_attach_volume_to_stopped_vm(self): '''Attach a volume to a stopped virtual machine, then start VM''' self.virtual_machine.stop(self.apiClient) ####################################### # STEP 1: Attach volume to stopped VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'stopped', TestScaleIOVolumes._vm_not_in_stopped_state_err_msg ) TestScaleIOVolumes._start_vm(self.virtual_machine) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) @attr(tags=['basic'], required_hardware=False) def test_06_attached_volume_reboot_vm(self): '''Attach volume to running VM, then reboot.''' TestScaleIOVolumes._start_vm(self.virtual_machine) ####################################### # STEP 1: Attach volume to running VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ####################################### # STEP 2: Reboot VM with attached vol # ####################################### TestScaleIOVolumes._reboot_vm(self.virtual_machine) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) @attr(tags=['basic'], required_hardware=False) def test_07_detach_volume_reboot_vm(self): '''Detach volume from a running VM, then reboot.''' TestScaleIOVolumes._start_vm(self.virtual_machine) ####################################### # STEP 1: Attach volume to running VM # ####################################### self.volume = self.virtual_machine.attach_volume( self.apiClient, self.volume ) self.attached = True vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, vm.id, TestScaleIOVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ######################################### # STEP 2: Detach volume from running VM # ######################################### self.volume = self.virtual_machine.detach_volume( self.apiClient, self.volume ) self.attached = False vm = self._get_vm(self.virtual_machine.id) self.assertEqual( self.volume.virtualmachineid, None, "The volume should not be attached to a VM." ) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) ####################################### # STEP 3: Reboot VM with detached vol # ####################################### self.virtual_machine.reboot(self.apiClient) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( vm.state.lower(), 'running', TestScaleIOVolumes._vm_not_in_running_state_err_msg ) @attr(tags=['basic'], required_hardware=False) def test_08_delete_volume_was_attached(self): '''Delete volume that was attached to a VM and is detached now''' TestScaleIOVolumes._start_vm(self.virtual_machine) ####################################### # STEP 1: Create vol and attach to VM # ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_2], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering.id ) volume_to_delete_later = new_volume new_volume = self.virtual_machine.attach_volume( self.apiClient, new_volume ) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 2: Detach and delete volume # ####################################### new_volume = self.virtual_machine.detach_volume( self.apiClient, new_volume ) vm = self._get_vm(self.virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, None, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) @attr(tags=['advanced', 'migration'], required_hardware=False) def test_09_migrate_volume_to_same_instance_pool(self): '''Migrate volume to the same instance pool''' if not TestData.migrationTests: self.skipTest("Volume migration tests not enabled, skipping test") ####################################### # STEP 1: Create VM and Start VM # ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine3], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(test_virtual_machine) ####################################### # STEP 2: Create vol and attach to VM # ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_3], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering_same_inst.id ) volume_to_delete_later = new_volume new_volume = test_virtual_machine.attach_volume( self.apiClient, new_volume ) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 3: Stop VM and Migrate volume # ####################################### test_virtual_machine.stop(self.apiClient) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( vm.state.lower(), 'stopped', str(vm.state) ) pools = StoragePool.listForMigration( self.apiClient, id=new_volume.id ) if not pools: self.skipTest("No suitable storage pools found for volume migration, skipping test") self.assertEqual( validateList(pools)[0], PASS, "Invalid pool response from findStoragePoolsForMigration API" ) pool = pools[0] self.debug("Migrating Volume-ID: %s to Same Instance Pool: %s" % (new_volume.id, pool.id)) try: Volume.migrate( self.apiClient, volumeid=new_volume.id, storageid=pool.id ) except Exception as e: self.fail("Volume migration failed with error %s" % e) ####################################### # STEP 4: Detach and delete volume # ####################################### new_volume = test_virtual_machine.detach_volume( self.apiClient, new_volume ) self.assertEqual( new_volume.virtualmachineid, None, "Check if attached to virtual machine" ) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) ####################################### # STEP 4: Delete VM # ####################################### test_virtual_machine.delete(self.apiClient, True) @attr(tags=['advanced', 'migration'], required_hardware=False) def test_10_migrate_volume_to_distinct_instance_pool(self): '''Migrate volume to distinct instance pool''' if not TestData.migrationTests: self.skipTest("Volume migration tests not enabled, skipping test") ####################################### # STEP 1: Create VM and Start VM # ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine4], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(test_virtual_machine) ####################################### # STEP 2: Create vol and attach to VM # ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_4], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering_distinct_inst.id ) volume_to_delete_later = new_volume new_volume = test_virtual_machine.attach_volume( self.apiClient, new_volume ) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 3: Stop VM and Migrate volume # ####################################### test_virtual_machine.stop(self.apiClient) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( vm.state.lower(), 'stopped', str(vm.state) ) pools = StoragePool.listForMigration( self.apiClient, id=new_volume.id ) if not pools: self.skipTest("No suitable storage pools found for volume migration, skipping test") self.assertEqual( validateList(pools)[0], PASS, "Invalid pool response from findStoragePoolsForMigration API" ) pool = pools[0] self.debug("Migrating Volume-ID: %s to Distinct Instance Pool: %s" % (new_volume.id, pool.id)) try: Volume.migrate( self.apiClient, volumeid=new_volume.id, storageid=pool.id ) except Exception as e: self.fail("Volume migration failed with error %s" % e) ####################################### # STEP 4: Detach and delete volume # ####################################### new_volume = test_virtual_machine.detach_volume( self.apiClient, new_volume ) self.assertEqual( new_volume.virtualmachineid, None, "Check if attached to virtual machine" ) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) ####################################### # STEP 4: Delete VM # ####################################### test_virtual_machine.delete(self.apiClient, True) @attr(tags=['advanced', 'migration'], required_hardware=False) def test_11_live_migrate_volume_to_same_instance_pool(self): '''Migrate volume to the same instance pool''' if not TestData.migrationTests: self.skipTest("Volume migration tests not enabled, skipping test") ####################################### # STEP 1: Create VM and Start VM # ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine3], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(test_virtual_machine) ####################################### # STEP 2: Create vol and attach to VM # ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_3], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering_same_inst.id ) volume_to_delete_later = new_volume new_volume = test_virtual_machine.attach_volume( self.apiClient, new_volume ) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 3: Migrate volume # ####################################### pools = StoragePool.listForMigration( self.apiClient, id=new_volume.id ) if not pools: self.skipTest("No suitable storage pools found for volume migration, skipping test") self.assertEqual( validateList(pools)[0], PASS, "Invalid pool response from findStoragePoolsForMigration API" ) pool = pools[0] self.debug("Migrating Volume-ID: %s to Same Instance Pool: %s" % (new_volume.id, pool.id)) try: Volume.migrate( self.apiClient, volumeid=new_volume.id, storageid=pool.id ) except Exception as e: self.fail("Volume migration failed with error %s" % e) ####################################### # STEP 4: Detach and delete volume # ####################################### new_volume = test_virtual_machine.detach_volume( self.apiClient, new_volume ) self.assertEqual( new_volume.virtualmachineid, None, "Check if attached to virtual machine" ) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) ####################################### # STEP 4: Delete VM # ####################################### test_virtual_machine.delete(self.apiClient, True) @attr(tags=['advanced', 'migration'], required_hardware=False) def test_12_migrate_volume_to_distinct_instance_pool(self): '''Migrate volume to distinct instance pool''' if not TestData.migrationTests: self.skipTest("Volume migration tests not enabled, skipping test") ####################################### # STEP 1: Create VM and Start VM # ####################################### test_virtual_machine = VirtualMachine.create( self.apiClient, self.testdata[TestData.virtualMachine4], accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=self.template.id, domainid=self.domain.id, startvm=False ) TestScaleIOVolumes._start_vm(test_virtual_machine) ####################################### # STEP 2: Create vol and attach to VM # ####################################### new_volume = Volume.create( self.apiClient, self.testdata[TestData.volume_4], account=self.account.name, domainid=self.domain.id, zoneid=self.zone.id, diskofferingid=self.disk_offering_distinct_inst.id ) volume_to_delete_later = new_volume new_volume = test_virtual_machine.attach_volume( self.apiClient, new_volume ) vm = self._get_vm(test_virtual_machine.id) self.assertEqual( new_volume.virtualmachineid, vm.id, "Check if attached to virtual machine" ) self.assertEqual( vm.state.lower(), 'running', str(vm.state) ) ####################################### # STEP 3: Migrate volume # ####################################### pools = StoragePool.listForMigration( self.apiClient, id=new_volume.id ) if not pools: self.skipTest("No suitable storage pools found for volume migration, skipping test") self.assertEqual( validateList(pools)[0], PASS, "Invalid pool response from findStoragePoolsForMigration API" ) pool = pools[0] self.debug("Migrating Volume-ID: %s to Distinct Instance Pool: %s" % (new_volume.id, pool.id)) try: Volume.migrate( self.apiClient, volumeid=new_volume.id, storageid=pool.id ) except Exception as e: self.fail("Volume migration failed with error %s" % e) ####################################### # STEP 4: Detach and delete volume # ####################################### new_volume = test_virtual_machine.detach_volume( self.apiClient, new_volume ) self.assertEqual( new_volume.virtualmachineid, None, "Check if attached to virtual machine" ) volume_to_delete_later.delete(self.apiClient) list_volumes_response = list_volumes( self.apiClient, id=new_volume.id ) self.assertEqual( list_volumes_response, None, "Check volume was deleted" ) ####################################### # STEP 4: Delete VM # ####################################### test_virtual_machine.delete(self.apiClient, True) def _create_vm_using_template_and_destroy_vm(self, template): vm_name = "VM-%d" % random.randint(0, 100) virtual_machine_dict = {"name": vm_name, "displayname": vm_name} virtual_machine = VirtualMachine.create( self.apiClient, virtual_machine_dict, accountid=self.account.name, zoneid=self.zone.id, serviceofferingid=self.compute_offering.id, templateid=template.id, domainid=self.domain.id, startvm=True ) list_volumes_response = list_volumes( self.apiClient, virtualmachineid=virtual_machine.id, listall=True ) vm_root_volume = list_volumes_response[0] vm_root_volume_name = vm_root_volume.name virtual_machine.delete(self.apiClient, True) def _get_bytes_from_gb(self, number_in_gb): return number_in_gb * 1024 * 1024 * 1024 def _get_volume(self, volume_id): list_vols_response = list_volumes(self.apiClient, id=volume_id) return list_vols_response[0] def _get_vm(self, vm_id): list_vms_response = list_virtual_machines(self.apiClient, id=vm_id) return list_vms_response[0] def _get_template_cache_name(self): if TestData.hypervisor_type == TestData.kvm: return TestData.templateCacheNameKvm self.assert_(False, "Invalid hypervisor type") @classmethod def _start_vm(cls, vm): vm_for_check = list_virtual_machines( cls.apiClient, id=vm.id )[0] if vm_for_check.state == VirtualMachine.STOPPED: vm.start(cls.apiClient) # For KVM, just give it 90 seconds to boot up. if TestData.hypervisor_type == TestData.kvm: time.sleep(90) @classmethod def _reboot_vm(cls, vm): vm.reboot(cls.apiClient) # For KVM, just give it 90 seconds to boot up. if TestData.hypervisor_type == TestData.kvm: time.sleep(90)