mirror of
				https://github.com/apache/cloudstack.git
				synced 2025-10-26 08:42:29 +01:00 
			
		
		
		
	
		
			
				
	
	
		
			708 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			708 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # Licensed to the Apache Software Foundation (ASF) under one
 | |
| # or more contributor license agreements.  See the NOTICE file
 | |
| # distributed with this work for additional information
 | |
| # regarding copyright ownership.  The ASF licenses this file
 | |
| # to you under the Apache License, Version 2.0 (the
 | |
| # "License"); you may not use this file except in compliance
 | |
| # with the License.  You may obtain a copy of the License at
 | |
| #
 | |
| #   http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing,
 | |
| # software distributed under the License is distributed on an
 | |
| # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 | |
| # KIND, either express or implied.  See the License for the
 | |
| # specific language governing permissions and limitations
 | |
| # under the License.
 | |
| 
 | |
| import logging
 | |
| import random
 | |
| import SignedAPICall
 | |
| import XenAPI
 | |
| 
 | |
| from solidfire.factory import ElementFactory
 | |
| 
 | |
| from util import sf_util
 | |
| 
 | |
| # All tests inherit from cloudstackTestCase
 | |
| from marvin.cloudstackTestCase import cloudstackTestCase
 | |
| 
 | |
| # base - contains all resources as entities and defines create, delete, list operations on them
 | |
| from marvin.lib.base import Account, DiskOffering, ServiceOffering, StoragePool, User, VirtualMachine, Volume
 | |
| 
 | |
| # common - commonly used methods for all tests are listed here
 | |
| from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_volumes
 | |
| 
 | |
| # utils - utility classes for common cleanup, external library wrappers, etc.
 | |
| from marvin.lib.utils import cleanup_resources
 | |
| 
 | |
| # Prerequisites:
 | |
| #  Only one zone
 | |
| #  Only one pod
 | |
| #  Two clusters (have system VMs (including the VR) running on local or NFS storage)
 | |
| #
 | |
| # Running the tests:
 | |
| #  Verify the "xen_server_hostname_src" and "xen_server_hostname_dest" variables are correct.
 | |
| 
 | |
| 
 | |
| class TestData():
 | |
|     account = "account"
 | |
|     capacityBytes = "capacitybytes"
 | |
|     capacityIops = "capacityiops"
 | |
|     clusterId1 = "clusterId1"
 | |
|     clusterId2 = "clusterId2"
 | |
|     computeOffering1 = "computeoffering1"
 | |
|     computeOffering2 = "computeoffering2"
 | |
|     computeOffering3 = "computeoffering3"
 | |
|     diskName = "diskname"
 | |
|     diskOffering1 = "diskoffering1"
 | |
|     diskOffering2 = "diskoffering2"
 | |
|     domainId = "domainid"
 | |
|     hypervisor = "hypervisor"
 | |
|     login = "login"
 | |
|     mvip = "mvip"
 | |
|     name = "name"
 | |
|     password = "password"
 | |
|     podId = "podid"
 | |
|     port = "port"
 | |
|     primaryStorage = "primarystorage"
 | |
|     primaryStorage2 = "primarystorage2"
 | |
|     provider = "provider"
 | |
|     scope = "scope"
 | |
|     solidFire = "solidfire"
 | |
|     storageTag = "SolidFire_SAN_1"
 | |
|     storageTag2 = "SolidFire_Volume_1"
 | |
|     tags = "tags"
 | |
|     templateCacheName = "centos56-x86-64-xen"
 | |
|     testAccount = "testaccount"
 | |
|     url = "url"
 | |
|     user = "user"
 | |
|     username = "username"
 | |
|     virtualMachine = "virtualmachine"
 | |
|     virtualMachine2 = "virtualmachine2"
 | |
|     volume_1 = "volume_1"
 | |
|     xenServer = "xenserver"
 | |
|     zoneId = "zoneid"
 | |
| 
 | |
|     xen_server_hostname_src = "XenServer-6.5-1"
 | |
|     xen_server_hostname_dest = "XenServer-6.5-3"
 | |
| 
 | |
|     def __init__(self):
 | |
|         self.testdata = {
 | |
|             TestData.solidFire: {
 | |
|                 TestData.mvip: "10.117.40.120",
 | |
|                 TestData.username: "admin",
 | |
|                 TestData.password: "admin",
 | |
|                 TestData.port: 443,
 | |
|                 TestData.url: "https://10.117.40.120:443"
 | |
|             },
 | |
|             TestData.xenServer: {
 | |
|                 TestData.username: "root",
 | |
|                 TestData.password: "solidfire"
 | |
|             },
 | |
|             TestData.account: {
 | |
|                 "email": "test@test.com",
 | |
|                 "firstname": "John",
 | |
|                 "lastname": "Doe",
 | |
|                 "username": "test",
 | |
|                 "password": "test"
 | |
|             },
 | |
|             TestData.testAccount: {
 | |
|                 "email": "test2@test2.com",
 | |
|                 "firstname": "Jane",
 | |
|                 "lastname": "Doe",
 | |
|                 "username": "test2",
 | |
|                 "password": "test"
 | |
|             },
 | |
|             TestData.user: {
 | |
|                 "email": "user@test.com",
 | |
|                 "firstname": "Jane",
 | |
|                 "lastname": "Doe",
 | |
|                 "username": "testuser",
 | |
|                 "password": "password"
 | |
|             },
 | |
|             TestData.primaryStorage: {
 | |
|                 TestData.name: "SolidFire-%d" % random.randint(0, 100),
 | |
|                 TestData.scope: "ZONE",
 | |
|                 TestData.url: "MVIP=10.117.40.120;SVIP=10.117.41.120;" +
 | |
|                        "clusterAdminUsername=admin;clusterAdminPassword=admin;" +
 | |
|                        "clusterDefaultMinIops=10000;clusterDefaultMaxIops=15000;" +
 | |
|                        "clusterDefaultBurstIopsPercentOfMaxIops=1.5;",
 | |
|                 TestData.provider: "SolidFire",
 | |
|                 TestData.tags: TestData.storageTag,
 | |
|                 TestData.capacityIops: 4500000,
 | |
|                 TestData.capacityBytes: 2251799813685248,
 | |
|                 TestData.hypervisor: "Any",
 | |
|                 TestData.zoneId: 1
 | |
|             },
 | |
|             TestData.primaryStorage2: {
 | |
|                 TestData.name: "SolidFireShared-%d" % random.randint(0, 100),
 | |
|                 TestData.scope: "CLUSTER",
 | |
|                 TestData.url: "MVIP=10.117.40.120;SVIP=10.117.41.120;" +
 | |
|                         "clusterAdminUsername=admin;clusterAdminPassword=admin;" +
 | |
|                         "minIops=5000;maxIops=50000;burstIops=75000",
 | |
|                 TestData.provider: "SolidFireShared",
 | |
|                 TestData.tags: TestData.storageTag2,
 | |
|                 TestData.capacityIops: 5000,
 | |
|                 TestData.capacityBytes: 1099511627776,
 | |
|                 TestData.hypervisor: "XenServer",
 | |
|                 TestData.podId: 1,
 | |
|                 TestData.zoneId: 1
 | |
|             },
 | |
|             TestData.virtualMachine: {
 | |
|                 "name": "TestVM",
 | |
|                 "displayname": "Test VM"
 | |
|             },
 | |
|             TestData.computeOffering1: {
 | |
|                 "name": "SF_CO_1",
 | |
|                 "displaytext": "SF_CO_1 (Min IOPS = 1,000; Max IOPS = 2,000)",
 | |
|                 "cpunumber": 1,
 | |
|                 "cpuspeed": 100,
 | |
|                 "memory": 128,
 | |
|                 "storagetype": "shared",
 | |
|                 "customizediops": False,
 | |
|                 "miniops": 1000,
 | |
|                 "maxiops": 2000,
 | |
|                 "hypervisorsnapshotreserve": 125,
 | |
|                 TestData.tags: TestData.storageTag,
 | |
|             },
 | |
|             TestData.computeOffering2: {
 | |
|                 "name": "SF_CO_2",
 | |
|                 "displaytext": "SF_CO_2 (Min IOPS = 1,000; Max IOPS = 2,000)",
 | |
|                 "cpunumber": 1,
 | |
|                 "cpuspeed": 100,
 | |
|                 "memory": 128,
 | |
|                 "storagetype": "shared",
 | |
|                 "customizediops": False,
 | |
|                 "miniops": 1000,
 | |
|                 "maxiops": 2000,
 | |
|                 "hypervisorsnapshotreserve": 100,
 | |
|                 TestData.tags: TestData.storageTag,
 | |
|             },
 | |
|             TestData.computeOffering3: {
 | |
|                 "name": "SF_CO_3",
 | |
|                 "displaytext": "SF_CO_3 Desc",
 | |
|                 "cpunumber": 1,
 | |
|                 "cpuspeed": 100,
 | |
|                 "memory": 128,
 | |
|                 "storagetype": "shared",
 | |
|                 TestData.tags: TestData.storageTag2,
 | |
|             },
 | |
|             TestData.diskOffering1: {
 | |
|                 "name": "SF_DO_1",
 | |
|                 "displaytext": "SF_DO_1 (Min IOPS = 3,000; Max IOPS = 6,000)",
 | |
|                 "disksize": 100,
 | |
|                 "customizediops": False,
 | |
|                 "miniops": 3000,
 | |
|                 "maxiops": 6000,
 | |
|                 "hypervisorsnapshotreserve": 125,
 | |
|                 TestData.tags: TestData.storageTag,
 | |
|                 "storagetype": "shared"
 | |
|             },
 | |
|             TestData.diskOffering2: {
 | |
|                 "name": "SF_DO_2",
 | |
|                 "displaytext": "SF_DO_2 (Min IOPS = 3,000; Max IOPS = 6,000)",
 | |
|                 "disksize": 100,
 | |
|                 "customizediops": False,
 | |
|                 "miniops": 3000,
 | |
|                 "maxiops": 6000,
 | |
|                 "hypervisorsnapshotreserve": 100,
 | |
|                 TestData.tags: TestData.storageTag,
 | |
|                 "storagetype": "shared"
 | |
|             },
 | |
|             TestData.volume_1: {
 | |
|                 TestData.diskName: "test-volume",
 | |
|             },
 | |
|             TestData.zoneId: 1,
 | |
|             TestData.clusterId1: 1,
 | |
|             TestData.clusterId2: 2,
 | |
|             TestData.domainId: 1,
 | |
|             TestData.url: "10.117.40.114"
 | |
|         }
 | |
| 
 | |
| 
 | |
| class TestVMMigrationWithStorage(cloudstackTestCase):
 | |
|     _sf_account_id_should_be_non_zero_int_err_msg = "The SolidFire account ID should be a non-zero integer."
 | |
| 
 | |
|     @classmethod
 | |
|     def setUpClass(cls):
 | |
|         # Set up API client
 | |
|         testclient = super(TestVMMigrationWithStorage, cls).getClsTestClient()
 | |
| 
 | |
|         cls.apiClient = testclient.getApiClient()
 | |
|         cls.configData = testclient.getParsedTestDataConfig()
 | |
|         cls.dbConnection = testclient.getDbConnection()
 | |
| 
 | |
|         cls.testdata = TestData().testdata
 | |
| 
 | |
|         xenserver = cls.testdata[TestData.xenServer]
 | |
| 
 | |
|         # Set up xenAPI connection
 | |
|         host_ip = "https://" + \
 | |
|                   list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId1], name=TestData.xen_server_hostname_src)[0].ipaddress
 | |
| 
 | |
|         # Set up XenAPI connection
 | |
|         cls.xen_session_1 = XenAPI.Session(host_ip)
 | |
| 
 | |
|         cls.xen_session_1.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
 | |
| 
 | |
|         # Set up xenAPI connection
 | |
|         host_ip = "https://" + \
 | |
|                   list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId2], name=TestData.xen_server_hostname_dest)[0].ipaddress
 | |
| 
 | |
|         # Set up XenAPI connection
 | |
|         cls.xen_session_2 = XenAPI.Session(host_ip)
 | |
| 
 | |
|         cls.xen_session_2.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
 | |
| 
 | |
|         # Set up SolidFire connection
 | |
|         solidfire = cls.testdata[TestData.solidFire]
 | |
| 
 | |
|         cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
 | |
| 
 | |
|         # Get Resources from Cloud Infrastructure
 | |
|         cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
 | |
|         cls.cluster_1 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId1])[0]
 | |
|         cls.cluster_2 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId2])[0]
 | |
|         cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
 | |
|         cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
 | |
| 
 | |
|         # Create test account
 | |
|         cls.account = Account.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata["account"],
 | |
|             admin=1
 | |
|         )
 | |
| 
 | |
|         # Set up connection to make customized API calls
 | |
|         cls.user = User.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata["user"],
 | |
|             account=cls.account.name,
 | |
|             domainid=cls.domain.id
 | |
|         )
 | |
| 
 | |
|         url = cls.testdata[TestData.url]
 | |
| 
 | |
|         api_url = "http://" + url + ":8080/client/api"
 | |
|         userkeys = User.registerUserKeys(cls.apiClient, cls.user.id)
 | |
| 
 | |
|         cls.cs_api = SignedAPICall.CloudStack(api_url, userkeys.apikey, userkeys.secretkey)
 | |
| 
 | |
|         primarystorage = cls.testdata[TestData.primaryStorage]
 | |
| 
 | |
|         cls.primary_storage = StoragePool.create(
 | |
|             cls.apiClient,
 | |
|             primarystorage
 | |
|         )
 | |
| 
 | |
|         cls.compute_offering_1 = ServiceOffering.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata[TestData.computeOffering1]
 | |
|         )
 | |
| 
 | |
|         cls.compute_offering_2 = ServiceOffering.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata[TestData.computeOffering2]
 | |
|         )
 | |
| 
 | |
|         cls.compute_offering_3 = ServiceOffering.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata[TestData.computeOffering3]
 | |
|         )
 | |
| 
 | |
|         cls.disk_offering_1 = DiskOffering.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata[TestData.diskOffering1]
 | |
|         )
 | |
| 
 | |
|         cls.disk_offering_2 = DiskOffering.create(
 | |
|             cls.apiClient,
 | |
|             cls.testdata[TestData.diskOffering2]
 | |
|         )
 | |
| 
 | |
|         # Resources that are to be destroyed
 | |
|         cls._cleanup = [
 | |
|             cls.compute_offering_1,
 | |
|             cls.compute_offering_2,
 | |
|             cls.compute_offering_3,
 | |
|             cls.disk_offering_1,
 | |
|             cls.disk_offering_2,
 | |
|             cls.user,
 | |
|             cls.account
 | |
|         ]
 | |
| 
 | |
|     @classmethod
 | |
|     def tearDownClass(cls):
 | |
|         try:
 | |
|             cleanup_resources(cls.apiClient, cls._cleanup)
 | |
| 
 | |
|             cls.primary_storage.delete(cls.apiClient)
 | |
| 
 | |
|             sf_util.purge_solidfire_volumes(cls.sfe)
 | |
|         except Exception as e:
 | |
|             logging.debug("Exception in tearDownClass(cls): %s" % e)
 | |
| 
 | |
|     def setUp(self):
 | |
|         self.cleanup = []
 | |
| 
 | |
|     def tearDown(self):
 | |
|         try:
 | |
|             cleanup_resources(self.apiClient, self.cleanup)
 | |
| 
 | |
|             sf_util.purge_solidfire_volumes(self.sfe)
 | |
|         except Exception as e:
 | |
|             logging.debug("Exception in tearDownClass(self): %s" % e)
 | |
| 
 | |
|     def test_01_storage_migrate_root_and_data_disks(self):
 | |
|         src_host, dest_host = self._get_source_and_dest_hosts()
 | |
| 
 | |
|         virtual_machine = VirtualMachine.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.virtualMachine],
 | |
|             accountid=self.account.name,
 | |
|             zoneid=self.zone.id,
 | |
|             serviceofferingid=self.compute_offering_1.id,
 | |
|             templateid=self.template.id,
 | |
|             domainid=self.domain.id,
 | |
|             hostid=src_host.id,
 | |
|             startvm=True
 | |
|         )
 | |
| 
 | |
|         self.cleanup.append(virtual_machine)
 | |
| 
 | |
|         cs_root_volume = list_volumes(self.apiClient, listall=True, virtualmachineid=virtual_machine.id)[0]
 | |
| 
 | |
|         sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
 | |
|                                                   TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
 | |
| 
 | |
|         cs_data_volume = Volume.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.volume_1],
 | |
|             account=self.account.name,
 | |
|             domainid=self.domain.id,
 | |
|             zoneid=self.zone.id,
 | |
|             diskofferingid=self.disk_offering_1.id
 | |
|         )
 | |
| 
 | |
|         self.cleanup.append(cs_data_volume)
 | |
| 
 | |
|         cs_data_volume = virtual_machine.attach_volume(
 | |
|             self.apiClient,
 | |
|             cs_data_volume
 | |
|         )
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
 | |
| 
 | |
|         sf_root_volume, sf_data_volume = self._migrate_and_verify(virtual_machine, dest_host, cs_root_volume, cs_data_volume, sf_account_id,
 | |
|                                                                   sf_root_volume, sf_data_volume, self.xen_session_1, self.xen_session_2)
 | |
| 
 | |
|         src_host, dest_host = dest_host, src_host
 | |
| 
 | |
|         self._migrate_and_verify(virtual_machine, dest_host, cs_root_volume, cs_data_volume, sf_account_id, sf_root_volume, sf_data_volume,
 | |
|                                  self.xen_session_2, self.xen_session_1)
 | |
| 
 | |
|     def test_02_storage_migrate_root_and_data_disks(self):
 | |
|         primarystorage2 = self.testdata[TestData.primaryStorage2]
 | |
| 
 | |
|         primary_storage_2 = StoragePool.create(
 | |
|             self.apiClient,
 | |
|             primarystorage2,
 | |
|             clusterid=self.cluster_1.id
 | |
|         )
 | |
| 
 | |
|         primary_storage_3 = StoragePool.create(
 | |
|             self.apiClient,
 | |
|             primarystorage2,
 | |
|             clusterid=self.cluster_2.id
 | |
|         )
 | |
| 
 | |
|         src_host, dest_host = self._get_source_and_dest_hosts()
 | |
| 
 | |
|         virtual_machine = VirtualMachine.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.virtualMachine],
 | |
|             accountid=self.account.name,
 | |
|             zoneid=self.zone.id,
 | |
|             serviceofferingid=self.compute_offering_3.id,
 | |
|             templateid=self.template.id,
 | |
|             domainid=self.domain.id,
 | |
|             hostid=src_host.id,
 | |
|             startvm=True
 | |
|         )
 | |
| 
 | |
|         cs_data_volume = Volume.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.volume_1],
 | |
|             account=self.account.name,
 | |
|             domainid=self.domain.id,
 | |
|             zoneid=self.zone.id,
 | |
|             diskofferingid=self.disk_offering_1.id
 | |
|         )
 | |
| 
 | |
|         self.cleanup = [
 | |
|             virtual_machine,
 | |
|             cs_data_volume,
 | |
|             primary_storage_2,
 | |
|             primary_storage_3
 | |
|         ]
 | |
| 
 | |
|         cs_data_volume = virtual_machine.attach_volume(
 | |
|             self.apiClient,
 | |
|             cs_data_volume
 | |
|         )
 | |
| 
 | |
|         sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
 | |
|                                                   TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
 | |
| 
 | |
|         sf_data_volume = self._migrate_and_verify_one_disk_only(virtual_machine, dest_host, cs_data_volume, sf_account_id,
 | |
|                                                                 sf_data_volume, self.xen_session_1, self.xen_session_2)
 | |
| 
 | |
|         src_host, dest_host = dest_host, src_host
 | |
| 
 | |
|         self._migrate_and_verify_one_disk_only(virtual_machine, dest_host, cs_data_volume, sf_account_id, sf_data_volume,
 | |
|                                                self.xen_session_2, self.xen_session_1)
 | |
| 
 | |
|     # The hypervisor snapshot reserve isn't large enough for either the compute or disk offering.
 | |
|     def test_03_storage_migrate_root_and_data_disks_fail(self):
 | |
|         self._execute_migration_failure(self.compute_offering_2.id, self.disk_offering_2.id)
 | |
| 
 | |
|     # The hypervisor snapshot reserve isn't large enough for the compute offering.
 | |
|     def test_04_storage_migrate_root_disk_fails(self):
 | |
|         self._execute_migration_failure(self.compute_offering_2.id, self.disk_offering_1.id)
 | |
| 
 | |
|     # The hypervisor snapshot reserve isn't large enough for the disk offering.
 | |
|     def test_05_storage_migrate_data_disk_fails(self):
 | |
|         self._execute_migration_failure(self.compute_offering_1.id, self.disk_offering_2.id)
 | |
| 
 | |
|     def _execute_migration_failure(self, compute_offering_id, disk_offering_id):
 | |
|         src_host, dest_host = self._get_source_and_dest_hosts()
 | |
| 
 | |
|         virtual_machine = VirtualMachine.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.virtualMachine],
 | |
|             accountid=self.account.name,
 | |
|             zoneid=self.zone.id,
 | |
|             serviceofferingid=compute_offering_id,
 | |
|             templateid=self.template.id,
 | |
|             domainid=self.domain.id,
 | |
|             hostid=src_host.id,
 | |
|             startvm=True
 | |
|         )
 | |
| 
 | |
|         self.cleanup.append(virtual_machine)
 | |
| 
 | |
|         cs_root_volume = list_volumes(self.apiClient, listall=True, virtualmachineid=virtual_machine.id)[0]
 | |
| 
 | |
|         sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
 | |
|                                                   TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
 | |
| 
 | |
|         cs_data_volume = Volume.create(
 | |
|             self.apiClient,
 | |
|             self.testdata[TestData.volume_1],
 | |
|             account=self.account.name,
 | |
|             domainid=self.domain.id,
 | |
|             zoneid=self.zone.id,
 | |
|             diskofferingid=disk_offering_id
 | |
|         )
 | |
| 
 | |
|         self.cleanup.append(cs_data_volume)
 | |
| 
 | |
|         cs_data_volume = virtual_machine.attach_volume(
 | |
|             self.apiClient,
 | |
|             cs_data_volume
 | |
|         )
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
 | |
| 
 | |
|         self._fail_migrate_and_verify(virtual_machine, dest_host, cs_root_volume, cs_data_volume, sf_account_id,
 | |
|                                       sf_root_volume, sf_data_volume, self.xen_session_1, self.xen_session_2)
 | |
| 
 | |
|     def _get_source_and_dest_hosts(self):
 | |
|         hosts = list_hosts(self.apiClient)
 | |
| 
 | |
|         for host in hosts:
 | |
|             if host.name == TestData.xen_server_hostname_src:
 | |
|                 src_host = host
 | |
|             elif host.name == TestData.xen_server_hostname_dest:
 | |
|                 dest_host = host
 | |
| 
 | |
|         self.assertIsNotNone(src_host, "Could not locate the source host")
 | |
| 
 | |
|         self.assertIsNotNone(dest_host, "Could not locate the destination host")
 | |
| 
 | |
|         return src_host, dest_host
 | |
| 
 | |
|     def _migrate_and_verify(self, virtual_machine, dest_host, cs_root_volume, cs_data_volume, sf_account_id, src_sf_root_volume, src_sf_data_volume,
 | |
|                             src_xen_session, dest_xen_session):
 | |
|         self._verifyFields(cs_root_volume, src_sf_root_volume)
 | |
|         self._verifyFields(cs_data_volume, src_sf_data_volume)
 | |
| 
 | |
|         virtual_machine.migrate_vm_with_volume(self.apiClient, dest_host.id)
 | |
| 
 | |
|         cs_root_volume = self._get_updated_cs_volume(cs_root_volume.id)
 | |
|         cs_data_volume = self._get_updated_cs_volume(cs_data_volume.id)
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
 | |
|         dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
 | |
| 
 | |
|         self._verifyFields(cs_root_volume, dest_sf_root_volume)
 | |
|         self._verifyFields(cs_data_volume, dest_sf_data_volume)
 | |
| 
 | |
|         self._verify_no_basic_volume_details()
 | |
| 
 | |
|         self._verify_different_volume_access_groups(src_sf_root_volume, dest_sf_root_volume)
 | |
|         self._verify_different_volume_access_groups(src_sf_data_volume, dest_sf_data_volume)
 | |
| 
 | |
|         self._verify_same_account(src_sf_root_volume, dest_sf_root_volume)
 | |
|         self._verify_same_account(src_sf_data_volume, dest_sf_data_volume)
 | |
| 
 | |
|         self._verifySfVolumeIds(src_sf_root_volume, dest_sf_root_volume)
 | |
|         self._verifySfVolumeIds(src_sf_data_volume, dest_sf_data_volume)
 | |
| 
 | |
|         self._verify_xenserver_state(src_xen_session, src_sf_root_volume, dest_xen_session, dest_sf_root_volume)
 | |
|         self._verify_xenserver_state(src_xen_session, src_sf_data_volume, dest_xen_session, dest_sf_data_volume)
 | |
| 
 | |
|         return dest_sf_root_volume, dest_sf_data_volume
 | |
| 
 | |
|     def _migrate_and_verify_one_disk_only(self, virtual_machine, dest_host, cs_volume, sf_account_id, src_sf_volume, src_xen_session, dest_xen_session):
 | |
|         self._verifyFields(cs_volume, src_sf_volume)
 | |
| 
 | |
|         virtual_machine.migrate_vm_with_volume(self.apiClient, dest_host.id)
 | |
| 
 | |
|         cs_volume = self._get_updated_cs_volume(cs_volume.id)
 | |
| 
 | |
|         sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         dest_sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_volume.name, self)
 | |
| 
 | |
|         self._verifyFields(cs_volume, dest_sf_volume)
 | |
| 
 | |
|         self._verify_no_basic_volume_details()
 | |
| 
 | |
|         self._verify_different_volume_access_groups(src_sf_volume, dest_sf_volume)
 | |
| 
 | |
|         self._verify_same_account(src_sf_volume, dest_sf_volume)
 | |
| 
 | |
|         self._verifySfVolumeIds(src_sf_volume, dest_sf_volume)
 | |
| 
 | |
|         self._verify_xenserver_state(src_xen_session, src_sf_volume, dest_xen_session, dest_sf_volume)
 | |
| 
 | |
|         return dest_sf_volume
 | |
| 
 | |
|     def _fail_migrate_and_verify(self, virtual_machine, dest_host, cs_root_volume, cs_data_volume, sf_account_id, src_sf_root_volume, src_sf_data_volume,
 | |
|                                  src_xen_session, dest_xen_session):
 | |
|         self._verifyFields(cs_root_volume, src_sf_root_volume)
 | |
|         self._verifyFields(cs_data_volume, src_sf_data_volume)
 | |
| 
 | |
|         class MigrationException(Exception):
 | |
|             def __init__(self, *args, **kwargs):
 | |
|                 Exception.__init__(self, *args, **kwargs)
 | |
| 
 | |
|         try:
 | |
|             virtual_machine.migrate_vm_with_volume(self.apiClient, dest_host.id)
 | |
| 
 | |
|             raise MigrationException("The migration did not fail (as expected).")
 | |
|         except MigrationException:
 | |
|             raise
 | |
|         except Exception:
 | |
|             pass
 | |
| 
 | |
|         self._verify_no_basic_volume_details()
 | |
| 
 | |
|         cs_root_volume_refreshed = self._get_updated_cs_volume(cs_root_volume.id)
 | |
|         cs_data_volume_refreshed = self._get_updated_cs_volume(cs_data_volume.id)
 | |
| 
 | |
|         self._verifyFields(cs_root_volume_refreshed, src_sf_root_volume)
 | |
|         self._verifyFields(cs_data_volume_refreshed, src_sf_data_volume)
 | |
| 
 | |
|         sf_volumes = sf_util.get_not_active_sf_volumes(self.sfe, sf_account_id)
 | |
| 
 | |
|         dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
 | |
|         dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
 | |
| 
 | |
|         self._verify_xenserver_state(dest_xen_session, dest_sf_root_volume, src_xen_session, src_sf_root_volume)
 | |
|         self._verify_xenserver_state(dest_xen_session, dest_sf_data_volume, src_xen_session, src_sf_data_volume)
 | |
| 
 | |
|     def _verify_different_volume_access_groups(self, src_sf_volume, dest_sf_volume):
 | |
|         src_vags = src_sf_volume.volume_access_groups
 | |
| 
 | |
|         sf_util.check_list(src_vags, 1, self, "'src_vags' should be a list with only one element in it.")
 | |
| 
 | |
|         dest_vags = dest_sf_volume.volume_access_groups
 | |
| 
 | |
|         sf_util.check_list(dest_vags, 1, self, "'dest_vags' should be a list with only one element in it.")
 | |
| 
 | |
|         self.assertNotEqual(src_vags[0], dest_vags[0], "The source and destination volumes should not be in the same volume access group.")
 | |
| 
 | |
|     def _get_updated_cs_volume(self, cs_volume_id):
 | |
|         return list_volumes(self.apiClient, listall=True, id=cs_volume_id)[0]
 | |
| 
 | |
|     def _verify_same_account(self, src_sf_volume, dest_sf_volume):
 | |
|         self.assertEqual(src_sf_volume.account_id, dest_sf_volume.account_id, "The source and destination volumes should be in the same SolidFire account.")
 | |
| 
 | |
|     def _verifySfVolumeIds(self, src_sf_volume, dest_sf_volume):
 | |
|         self.assert_(src_sf_volume.volume_id < dest_sf_volume.volume_id,
 | |
|                      "The destination SolidFire root volume's ID should be greater than the id of the source one.")
 | |
| 
 | |
|     # verify the name, folder, and iscsi_name
 | |
|     def _verifyFields(self, cs_volume, sf_volume):
 | |
|         self.assert_(cs_volume.name == sf_volume.name, "The CloudStack volume name does not match the SolidFire volume name.")
 | |
| 
 | |
|         cs_volume_folder = self._get_cs_volume_folder(cs_volume.id)
 | |
| 
 | |
|         self.assert_(int(cs_volume_folder) == sf_volume.volume_id, "The CloudStack folder name does not match the SolidFire volume ID.")
 | |
| 
 | |
|         cs_volume_iscsi_name = self._get_cs_volume_iscsi_name(cs_volume.id)
 | |
| 
 | |
|         self.assert_(cs_volume_iscsi_name == sf_util.format_iqn(sf_volume.iqn), "The CloudStack volume iscsi_name does not match the SolidFire volume IQN.")
 | |
| 
 | |
|     def _get_cs_volume_property(self, cs_volume_id, volume_property):
 | |
|         sql_query = "Select " + volume_property + " From volumes Where uuid = '" + cs_volume_id + "'"
 | |
| 
 | |
|         # make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
 | |
|         sql_result = self.dbConnection.execute(sql_query)
 | |
| 
 | |
|         return sql_result[0][0]
 | |
| 
 | |
|     def _get_cs_volume_folder(self, cs_volume_id):
 | |
|         return self._get_cs_volume_property(cs_volume_id, "folder")
 | |
| 
 | |
|     def _get_cs_volume_iscsi_name(self, cs_volume_id):
 | |
|         return self._get_cs_volume_property(cs_volume_id, "iscsi_name")
 | |
| 
 | |
|     def _verify_no_basic_volume_details(self):
 | |
|         sql_query = "Select id From volume_details Where name like 'basic_'"
 | |
| 
 | |
|         # make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
 | |
|         sql_result = self.dbConnection.execute(sql_query)
 | |
| 
 | |
|         sf_util.check_list(sql_result, 0, self, "The cloud.volume_details table should not have any name fields that start with 'basic_'.")
 | |
| 
 | |
|     def _verify_xenserver_state(self, xen_session_1, sf_volume_1, xen_session_2, sf_volume_2):
 | |
|         sr_name = sf_util.format_iqn(sf_volume_1.iqn)
 | |
| 
 | |
|         sf_util.check_xen_sr(sr_name, xen_session_1, self, False)
 | |
| 
 | |
|         sr_name = sf_util.format_iqn(sf_volume_2.iqn)
 | |
| 
 | |
|         sf_util.check_xen_sr(sr_name, xen_session_2, self)
 |