mirror of
				https://github.com/apache/cloudstack.git
				synced 2025-11-04 00:02:37 +01:00 
			
		
		
		
	Allowed zone-wide primary storage based on a custom plug-in to be added via the GUI in a KVM-only environment (previously this only worked for XenServer and VMware) Added support for root disks on managed storage with KVM Added support for volume snapshots with managed storage on KVM Enable creating a template directly from a volume (i.e. without having to go through a volume snapshot) on KVM with managed storage Only allow the resizing of a volume for managed storage on KVM if the volume in question is either not attached to a VM or is attached to a VM in the Stopped state. Included support for Reinstall VM on KVM with managed storage Enabled offline migration on KVM from non-managed storage to managed storage and vice versa Included support for online storage migration on KVM with managed storage (NFS and Ceph to managed storage) Added support to download (extract) a managed-storage volume to a QCOW2 file When uploading a file from outside of CloudStack to CloudStack, set the min and max IOPS, if applicable. Included support for the KVM auto-convergence feature The compression flag was actually added in version 1.0.3 (1000003) as opposed to version 1.3.0 (1003000) (changed this to reflect the correct version) On KVM when using iSCSI-based managed storage, if the user shuts a VM down from the guest OS (as opposed to doing so from CloudStack), we need to pass to the KVM agent a list of applicable iSCSI volumes that need to be disconnected. Added a new Global Setting: kvm.storage.live.migration.wait For XenServer, added a check to enforce that only volumes from zone-wide managed storage can be storage motioned from a host in one cluster to a host in another cluster (cannot do so at the time being with volumes from cluster-scoped managed storage) Don’t allow Storage XenMotion on a VM that has any managed-storage volume with one or more snapshots. Enabled for managed storage with VMware: Template caching, create snapshot, delete snapshot, create volume from snapshot, and create template from snapshot Added an SIOC API plug-in to support VMware SIOC When starting a VM that uses managed storage in a cluster other than the one it last was running in, we need to remove the reference to the iSCSI volume from the original cluster. Added the ability to revert a volume to a snapshot Enabled cluster-scoped managed storage Added support for VMware dynamic discovery
		
			
				
	
	
		
			2712 lines
		
	
	
		
			93 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			2712 lines
		
	
	
		
			93 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# Licensed to the Apache Software Foundation (ASF) under one
 | 
						|
# or more contributor license agreements.  See the NOTICE file
 | 
						|
# distributed with this work for additional information
 | 
						|
# regarding copyright ownership.  The ASF licenses this file
 | 
						|
# to you under the Apache License, Version 2.0 (the
 | 
						|
# "License"); you may not use this file except in compliance
 | 
						|
# with the License.  You may obtain a copy of the License at
 | 
						|
#
 | 
						|
#   http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
#
 | 
						|
# Unless required by applicable law or agreed to in writing,
 | 
						|
# software distributed under the License is distributed on an
 | 
						|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 | 
						|
# KIND, either express or implied.  See the License for the
 | 
						|
# specific language governing permissions and limitations
 | 
						|
# under the License.
 | 
						|
 | 
						|
import logging
 | 
						|
import random
 | 
						|
import SignedAPICall
 | 
						|
import time
 | 
						|
import XenAPI
 | 
						|
 | 
						|
from solidfire.factory import ElementFactory
 | 
						|
 | 
						|
from util import sf_util
 | 
						|
 | 
						|
# All tests inherit from cloudstackTestCase
 | 
						|
from marvin.cloudstackTestCase import cloudstackTestCase
 | 
						|
 | 
						|
# Import Integration Libraries
 | 
						|
 | 
						|
# base - contains all resources as entities and defines create, delete, list operations on them
 | 
						|
from marvin.lib.base import Account, DiskOffering, ServiceOffering, Snapshot, StoragePool, Template, User, VirtualMachine, Volume
 | 
						|
 | 
						|
# common - commonly used methods for all tests are listed here
 | 
						|
from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_virtual_machines, \
 | 
						|
    list_volumes
 | 
						|
 | 
						|
# utils - utility classes for common cleanup, external library wrappers, etc.
 | 
						|
from marvin.lib.utils import cleanup_resources
 | 
						|
 | 
						|
# Prerequisites:
 | 
						|
#  Only one zone
 | 
						|
#  Only one pod
 | 
						|
#  Only one cluster
 | 
						|
#
 | 
						|
# Running the tests:
 | 
						|
#  Change the "hypervisor_type" variable to control which hypervisor type to test.
 | 
						|
#  If using XenServer, verify the "xen_server_hostname" variable is correct.
 | 
						|
#  If using XenServer, change the "supports_cloning" variable to True or False as desired.
 | 
						|
#
 | 
						|
# Note:
 | 
						|
#  If you do have more than one cluster, you might need to change this line: cls.cluster = list_clusters(cls.apiClient)[0]
 | 
						|
 | 
						|
 | 
						|
class TestData():
 | 
						|
    # constants
 | 
						|
    account = "account"
 | 
						|
    capacityBytes = "capacitybytes"
 | 
						|
    capacityIops = "capacityiops"
 | 
						|
    clusterId = "clusterId"
 | 
						|
    computeOffering = "computeoffering"
 | 
						|
    diskName = "diskname"
 | 
						|
    diskOffering = "diskoffering"
 | 
						|
    domainId = "domainId"
 | 
						|
    hypervisor = "hypervisor"
 | 
						|
    kvm = "kvm"
 | 
						|
    login = "login"
 | 
						|
    mvip = "mvip"
 | 
						|
    one_GB_in_bytes = 1073741824
 | 
						|
    password = "password"
 | 
						|
    port = "port"
 | 
						|
    primaryStorage = "primarystorage"
 | 
						|
    provider = "provider"
 | 
						|
    scope = "scope"
 | 
						|
    solidFire = "solidfire"
 | 
						|
    storageTag = "SolidFire_SAN_1"
 | 
						|
    tags = "tags"
 | 
						|
    templateCacheNameKvm = "centos55-x86-64"
 | 
						|
    templateCacheNameXenServer = "centos56-x86-64-xen"
 | 
						|
    testAccount = "testaccount"
 | 
						|
    url = "url"
 | 
						|
    user = "user"
 | 
						|
    username = "username"
 | 
						|
    virtualMachine = "virtualmachine"
 | 
						|
    virtualMachine2 = "virtualmachine2"
 | 
						|
    volume_1 = "volume_1"
 | 
						|
    volume_2 = "volume_2"
 | 
						|
    xenServer = "xenserver"
 | 
						|
    zoneId = "zoneId"
 | 
						|
 | 
						|
    # modify to control which hypervisor type to test
 | 
						|
    hypervisor_type = kvm
 | 
						|
    xen_server_hostname = "XenServer-6.5-1"
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        self.testdata = {
 | 
						|
            TestData.solidFire: {
 | 
						|
                TestData.mvip: "10.117.40.120",
 | 
						|
                TestData.username: "admin",
 | 
						|
                TestData.password: "admin",
 | 
						|
                TestData.port: 443,
 | 
						|
                TestData.url: "https://10.117.40.120:443"
 | 
						|
            },
 | 
						|
            TestData.kvm: {
 | 
						|
                TestData.username: "root",
 | 
						|
                TestData.password: "solidfire"
 | 
						|
            },
 | 
						|
            TestData.xenServer: {
 | 
						|
                TestData.username: "root",
 | 
						|
                TestData.password: "solidfire"
 | 
						|
            },
 | 
						|
            TestData.account: {
 | 
						|
                "email": "test@test.com",
 | 
						|
                "firstname": "John",
 | 
						|
                "lastname": "Doe",
 | 
						|
                "username": "test",
 | 
						|
                "password": "test"
 | 
						|
            },
 | 
						|
            TestData.testAccount: {
 | 
						|
                "email": "test2@test2.com",
 | 
						|
                "firstname": "Jane",
 | 
						|
                "lastname": "Doe",
 | 
						|
                "username": "test2",
 | 
						|
                "password": "test"
 | 
						|
            },
 | 
						|
            TestData.user: {
 | 
						|
                "email": "user@test.com",
 | 
						|
                "firstname": "Jane",
 | 
						|
                "lastname": "Doe",
 | 
						|
                "username": "testuser",
 | 
						|
                "password": "password"
 | 
						|
            },
 | 
						|
            TestData.primaryStorage: {
 | 
						|
                "name": "SolidFire-%d" % random.randint(0, 100),
 | 
						|
                TestData.scope: "ZONE",
 | 
						|
                "url": "MVIP=10.117.40.120;SVIP=10.117.41.120;" +
 | 
						|
                       "clusterAdminUsername=admin;clusterAdminPassword=admin;" +
 | 
						|
                       "clusterDefaultMinIops=10000;clusterDefaultMaxIops=15000;" +
 | 
						|
                       "clusterDefaultBurstIopsPercentOfMaxIops=1.5;",
 | 
						|
                TestData.provider: "SolidFire",
 | 
						|
                TestData.tags: TestData.storageTag,
 | 
						|
                TestData.capacityIops: 4500000,
 | 
						|
                TestData.capacityBytes: 2251799813685248,
 | 
						|
                TestData.hypervisor: "Any"
 | 
						|
            },
 | 
						|
            TestData.virtualMachine: {
 | 
						|
                "name": "TestVM",
 | 
						|
                "displayname": "Test VM"
 | 
						|
            },
 | 
						|
            TestData.virtualMachine2: {
 | 
						|
                "name": "TestVM2",
 | 
						|
                "displayname": "Test VM 2"
 | 
						|
            },
 | 
						|
            TestData.computeOffering: {
 | 
						|
                "name": "SF_CO_1",
 | 
						|
                "displaytext": "SF_CO_1 (Min IOPS = 10,000; Max IOPS = 15,000)",
 | 
						|
                "cpunumber": 1,
 | 
						|
                "cpuspeed": 100,
 | 
						|
                "memory": 128,
 | 
						|
                "storagetype": "shared",
 | 
						|
                "customizediops": False,
 | 
						|
                "miniops": "10000",
 | 
						|
                "maxiops": "15000",
 | 
						|
                "hypervisorsnapshotreserve": 200,
 | 
						|
                TestData.tags: TestData.storageTag
 | 
						|
            },
 | 
						|
            TestData.diskOffering: {
 | 
						|
                "name": "SF_DO_1",
 | 
						|
                "displaytext": "SF_DO_1 (Min IOPS = 300; Max IOPS = 500)",
 | 
						|
                "disksize": 128,
 | 
						|
                "customizediops": False,
 | 
						|
                "miniops": 300,
 | 
						|
                "maxiops": 500,
 | 
						|
                "hypervisorsnapshotreserve": 200,
 | 
						|
                TestData.tags: TestData.storageTag,
 | 
						|
                "storagetype": "shared"
 | 
						|
            },
 | 
						|
            TestData.volume_1: {
 | 
						|
                TestData.diskName: "test-volume",
 | 
						|
            },
 | 
						|
            TestData.volume_2: {
 | 
						|
                TestData.diskName: "test-volume-2",
 | 
						|
            },
 | 
						|
            TestData.zoneId: 1,
 | 
						|
            TestData.clusterId: 1,
 | 
						|
            TestData.domainId: 1,
 | 
						|
            TestData.url: "10.117.40.114"
 | 
						|
        }
 | 
						|
 | 
						|
 | 
						|
class TestVolumes(cloudstackTestCase):
 | 
						|
    _should_only_be_one_vm_in_list_err_msg = "There should only be one VM in this list."
 | 
						|
    _should_only_be_one_volume_in_list_err_msg = "There should only be one volume in this list."
 | 
						|
    _should_only_be_one_host_in_list_err_msg = "There should only be one host in this list."
 | 
						|
    _should_only_be_two_volumes_in_list_err_msg = "There should only be two volumes in this list."
 | 
						|
    _sf_account_id_should_be_non_zero_int_err_msg = "The SolidFire account ID should be a non-zero integer."
 | 
						|
    _volume_size_should_be_non_zero_int_err_msg = "The SolidFire volume size should be a non-zero integer."
 | 
						|
    _volume_vm_id_and_vm_id_do_not_match_err_msg = "The volume's VM ID and the VM's ID do not match."
 | 
						|
    _vm_not_in_running_state_err_msg = "The VM is not in the 'Running' state."
 | 
						|
    _vm_not_in_stopped_state_err_msg = "The VM is not in the 'Stopped' state."
 | 
						|
    _volume_response_should_not_be_zero_err_msg = "The length of the response for the SolidFire-volume query should not be zero."
 | 
						|
    _volume_should_not_be_in_a_vag = "The volume should not be in a volume access group."
 | 
						|
    _volume_size_not_an_int = "'volume_size_in_GB' is not of type 'int'"
 | 
						|
    _only_data_volumes_err_msg = "Only data volumes can be resized via a new disk offering."
 | 
						|
    _to_change_volume_size_err_msg = "To change a volume's size without providing a new disk offering, its current " \
 | 
						|
        "disk offering must be customizable or it must be a root volume (if providing a disk offering, make sure it is " \
 | 
						|
        "different from the current disk offering)."
 | 
						|
    _min_iops_err_msg = "The current disk offering does not support customization of the 'Min IOPS' parameter."
 | 
						|
    _this_kind_of_disk_err_msg = "This kind of KVM disk cannot be resized while it is connected to a VM that's not in the Stopped state."
 | 
						|
    _template_creation_did_not_fail_err_msg = "The template creation did not fail (as expected)."
 | 
						|
    _volume_resize_did_not_fail_err_msg = "The volume resize did not fail (as expected)."
 | 
						|
    _volume_attached_to_non_stopped_vm_err_msg = "volume is attached to a non-stopped VM"
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def setUpClass(cls):
 | 
						|
        # Set up API client
 | 
						|
        testclient = super(TestVolumes, cls).getClsTestClient()
 | 
						|
 | 
						|
        cls.apiClient = testclient.getApiClient()
 | 
						|
        cls.configData = testclient.getParsedTestDataConfig()
 | 
						|
        cls.dbConnection = testclient.getDbConnection()
 | 
						|
 | 
						|
        cls.testdata = TestData().testdata
 | 
						|
 | 
						|
        cls._handle_supports_cloning()
 | 
						|
 | 
						|
        cls._connect_to_hypervisor()
 | 
						|
 | 
						|
        # Set up SolidFire connection
 | 
						|
        solidfire = cls.testdata[TestData.solidFire]
 | 
						|
 | 
						|
        cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
 | 
						|
 | 
						|
        # Get Resources from Cloud Infrastructure
 | 
						|
        cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
 | 
						|
        cls.cluster = list_clusters(cls.apiClient)[0]
 | 
						|
        cls.template = get_template(cls.apiClient, cls.zone.id, hypervisor=TestData.hypervisor_type)
 | 
						|
        cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
 | 
						|
 | 
						|
        # Create test account
 | 
						|
        cls.account = Account.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata["account"],
 | 
						|
            admin=1
 | 
						|
        )
 | 
						|
 | 
						|
        # Set up connection to make customized API calls
 | 
						|
        cls.user = User.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata["user"],
 | 
						|
            account=cls.account.name,
 | 
						|
            domainid=cls.domain.id
 | 
						|
        )
 | 
						|
 | 
						|
        url = cls.testdata[TestData.url]
 | 
						|
 | 
						|
        api_url = "http://" + url + ":8080/client/api"
 | 
						|
        userkeys = User.registerUserKeys(cls.apiClient, cls.user.id)
 | 
						|
 | 
						|
        cls.cs_api = SignedAPICall.CloudStack(api_url, userkeys.apikey, userkeys.secretkey)
 | 
						|
 | 
						|
        primarystorage = cls.testdata[TestData.primaryStorage]
 | 
						|
 | 
						|
        cls.primary_storage = StoragePool.create(
 | 
						|
            cls.apiClient,
 | 
						|
            primarystorage,
 | 
						|
            scope=primarystorage[TestData.scope],
 | 
						|
            zoneid=cls.zone.id,
 | 
						|
            provider=primarystorage[TestData.provider],
 | 
						|
            tags=primarystorage[TestData.tags],
 | 
						|
            capacityiops=primarystorage[TestData.capacityIops],
 | 
						|
            capacitybytes=primarystorage[TestData.capacityBytes],
 | 
						|
            hypervisor=primarystorage[TestData.hypervisor]
 | 
						|
        )
 | 
						|
 | 
						|
        cls.compute_offering = ServiceOffering.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata[TestData.computeOffering]
 | 
						|
        )
 | 
						|
 | 
						|
        cls.disk_offering = DiskOffering.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata[TestData.diskOffering]
 | 
						|
        )
 | 
						|
 | 
						|
        # Create VM and volume for tests
 | 
						|
        cls.virtual_machine = VirtualMachine.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata[TestData.virtualMachine],
 | 
						|
            accountid=cls.account.name,
 | 
						|
            zoneid=cls.zone.id,
 | 
						|
            serviceofferingid=cls.compute_offering.id,
 | 
						|
            templateid=cls.template.id,
 | 
						|
            domainid=cls.domain.id,
 | 
						|
            startvm=False
 | 
						|
        )
 | 
						|
 | 
						|
        TestVolumes._start_vm(cls.virtual_machine)
 | 
						|
 | 
						|
        cls.volume = Volume.create(
 | 
						|
            cls.apiClient,
 | 
						|
            cls.testdata[TestData.volume_1],
 | 
						|
            account=cls.account.name,
 | 
						|
            domainid=cls.domain.id,
 | 
						|
            zoneid=cls.zone.id,
 | 
						|
            diskofferingid=cls.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        # Resources that are to be destroyed
 | 
						|
        cls._cleanup = [
 | 
						|
            cls.volume,
 | 
						|
            cls.virtual_machine,
 | 
						|
            cls.compute_offering,
 | 
						|
            cls.disk_offering,
 | 
						|
            cls.user,
 | 
						|
            cls.account
 | 
						|
        ]
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def tearDownClass(cls):
 | 
						|
        try:
 | 
						|
            cleanup_resources(cls.apiClient, cls._cleanup)
 | 
						|
 | 
						|
            cls.primary_storage.delete(cls.apiClient)
 | 
						|
 | 
						|
            sf_util.purge_solidfire_volumes(cls.sfe)
 | 
						|
        except Exception as e:
 | 
						|
            logging.debug("Exception in tearDownClass(cls): %s" % e)
 | 
						|
 | 
						|
    def setUp(self):
 | 
						|
        self.attached = False
 | 
						|
        self.cleanup = []
 | 
						|
 | 
						|
    def tearDown(self):
 | 
						|
        if self.attached:
 | 
						|
            self.virtual_machine.detach_volume(self.apiClient, self.volume)
 | 
						|
 | 
						|
        cleanup_resources(self.apiClient, self.cleanup)
 | 
						|
 | 
						|
    def test_00_check_template_cache(self):
 | 
						|
        if self._supports_cloning == False:
 | 
						|
            return
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes()
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, self._get_template_cache_name(), self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            "The volume should not be in a VAG."
 | 
						|
        )
 | 
						|
 | 
						|
        sf_account_id = sf_volume.account_id
 | 
						|
 | 
						|
        sf_account = self.sfe.get_account_by_id(sf_account_id).account
 | 
						|
 | 
						|
        sf_account_name = sf_account.username
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            sf_account_name.endswith("_1"),
 | 
						|
            True,
 | 
						|
            "The template cache volume's account does not end with '_1'."
 | 
						|
        )
 | 
						|
 | 
						|
    def test_01_attach_new_volume_to_stopped_VM(self):
 | 
						|
        '''Attach a volume to a stopped virtual machine, then start VM'''
 | 
						|
 | 
						|
        self.virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        new_volume = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.volume_2],
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(new_volume)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        new_volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            new_volume
 | 
						|
        )
 | 
						|
 | 
						|
        newvolume = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            newvolume.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            "running",
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, new_volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, new_volume, self)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, newvolume.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, newvolume, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        # Detach volume
 | 
						|
        new_volume = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            new_volume
 | 
						|
        )
 | 
						|
 | 
						|
    def test_02_attach_detach_attach_volume(self):
 | 
						|
        '''Attach, detach, and attach volume to a running VM'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Attach volume to running VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
        # STEP 2: Detach volume from running VM #
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = False
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            None,
 | 
						|
            "The volume should not be attached to a VM."
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            str(vm.state)
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            "The volume should not be in a VAG."
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid, False)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 3: Attach volume to running VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
    def test_03_attached_volume_reboot_VM(self):
 | 
						|
        '''Attach volume to running VM, then reboot.'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Attach volume to running VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 2: Reboot VM with attached vol #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        TestVolumes._reboot_vm(self.virtual_machine)
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
    def test_04_detach_volume_reboot(self):
 | 
						|
        '''Detach volume from a running VM, then reboot.'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Attach volume to running VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
        # STEP 2: Detach volume from running VM #
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = False
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            None,
 | 
						|
            "The volume should not be attached to a VM."
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid, False)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 3: Reboot VM with detached vol #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.virtual_machine.reboot(self.apiClient)
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid, False)
 | 
						|
 | 
						|
    def test_05_detach_vol_stopped_VM_start(self):
 | 
						|
        '''Detach volume from a stopped VM, then start.'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Attach volume to running VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        hostid = vm.hostid
 | 
						|
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
        # STEP 2: Detach volume from stopped VM #
 | 
						|
        #########################################
 | 
						|
        #########################################
 | 
						|
 | 
						|
        self.virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = False
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            None,
 | 
						|
            "The volume should not be attached to a VM."
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'stopped',
 | 
						|
            TestVolumes._vm_not_in_stopped_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, hostid, False)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 3: Start VM with detached vol  #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid, False)
 | 
						|
 | 
						|
    def test_06_attach_volume_to_stopped_VM(self):
 | 
						|
        '''Attach a volume to a stopped virtual machine, then start VM'''
 | 
						|
 | 
						|
        self.virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Attach volume to stopped VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'stopped',
 | 
						|
            TestVolumes._vm_not_in_stopped_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            self._check_host_side(sf_iscsi_name, None, False)
 | 
						|
        elif TestData.hypervisor_type == TestData.xenServer:
 | 
						|
            self._check_host_side(sf_iscsi_name)
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
    def test_07_destroy_expunge_VM_with_volume(self):
 | 
						|
        '''Destroy and expunge VM with attached volume'''
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Create VM and attach volume #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        test_virtual_machine = VirtualMachine.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.virtualMachine2],
 | 
						|
            accountid=self.account.name,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            serviceofferingid=self.compute_offering.id,
 | 
						|
            templateid=self.template.id,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            startvm=False
 | 
						|
        )
 | 
						|
 | 
						|
        TestVolumes._start_vm(test_virtual_machine)
 | 
						|
 | 
						|
        self.volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(test_virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            TestVolumes._volume_vm_id_and_vm_id_do_not_match_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            TestVolumes._vm_not_in_running_state_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        hostid = vm.hostid
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        #   STEP 2: Destroy and Expunge VM    #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        test_virtual_machine.delete(self.apiClient, True)
 | 
						|
 | 
						|
        self.attached = False
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            None,
 | 
						|
            "Check if attached to virtual machine"
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.vmname,
 | 
						|
            None,
 | 
						|
            "Check if VM was expunged"
 | 
						|
        )
 | 
						|
 | 
						|
        list_virtual_machine_response = list_virtual_machines(
 | 
						|
            self.apiClient,
 | 
						|
            id=test_virtual_machine.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            list_virtual_machine_response,
 | 
						|
            None,
 | 
						|
            "Check if VM was actually expunged"
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, hostid, False)
 | 
						|
 | 
						|
    def test_08_delete_volume_was_attached(self):
 | 
						|
        '''Delete volume that was attached to a VM and is detached now'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Create vol and attach to VM #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        new_volume = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.volume_2],
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        volume_to_delete_later = new_volume
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        new_volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            new_volume
 | 
						|
        )
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            "Check if attached to virtual machine"
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            str(vm.state)
 | 
						|
        )
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, new_volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, new_volume, self)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        #  STEP 2: Detach and delete volume   #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        new_volume = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            new_volume
 | 
						|
        )
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, new_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            None,
 | 
						|
            "Check if attached to virtual machine"
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            str(vm.state)
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            len(sf_volume.volume_access_groups),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_should_not_be_in_a_vag
 | 
						|
        )
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid, False)
 | 
						|
 | 
						|
        volume_to_delete_later.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=new_volume.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            list_volumes_response,
 | 
						|
            None,
 | 
						|
            "Check volume was deleted"
 | 
						|
        )
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self, False)
 | 
						|
 | 
						|
    def test_09_attach_volumes_multiple_accounts(self):
 | 
						|
        '''Attach a data disk to a VM in one account and attach another data disk to a VM in another account'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        # STEP 1: Create account, VM, and vol #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        test_account = Account.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.testAccount],
 | 
						|
            admin=1
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_account)
 | 
						|
 | 
						|
        test_virtual_machine = VirtualMachine.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.virtualMachine2],
 | 
						|
            accountid=test_account.name,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            serviceofferingid=self.compute_offering.id,
 | 
						|
            templateid=self.template.id,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            startvm=False
 | 
						|
        )
 | 
						|
 | 
						|
        TestVolumes._start_vm(test_virtual_machine)
 | 
						|
 | 
						|
        test_volume = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.volume_2],
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            account=test_account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, test_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        #    STEP 2: Attach volumes to VMs    #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.volume = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vol.virtualmachineid,
 | 
						|
            vm.id,
 | 
						|
            "Check if attached to virtual machine"
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            str(vm.state)
 | 
						|
        )
 | 
						|
 | 
						|
        test_volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vol = sf_util.check_and_get_cs_volume(self, test_volume.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        test_vm = self._get_vm(test_virtual_machine.id)
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            test_vol.virtualmachineid,
 | 
						|
            test_vm.id,
 | 
						|
            "Check if attached to virtual machine of other acct"
 | 
						|
        )
 | 
						|
 | 
						|
        self.assertEqual(
 | 
						|
            test_vm.state.lower(),
 | 
						|
            'running',
 | 
						|
            str(test_vm.state)
 | 
						|
        )
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, vol, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        sf_test_account_id = sf_util.get_sf_account_id(self.cs_api, test_account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_test_volumes = self._get_active_sf_volumes(sf_test_account_id)
 | 
						|
 | 
						|
        sf_test_volume = sf_util.check_and_get_sf_volume(sf_test_volumes, test_vol.name, self)
 | 
						|
 | 
						|
        sf_test_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, test_vol, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_test_volume_size)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_test_volume, test_vol, sf_test_volume_size, self)
 | 
						|
 | 
						|
        sf_test_iscsi_name = sf_util.get_iqn(self.cs_api, test_volume, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_test_iscsi_name, test_vm.hostid)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_test_volume, sf_vag_id, self)
 | 
						|
 | 
						|
    def test_10_attach_more_than_one_disk_to_VM(self):
 | 
						|
        '''Attach more than one disk to a VM'''
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        volume_2 = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.volume_2],
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(volume_2)
 | 
						|
 | 
						|
        sf_util.check_and_get_cs_volume(self, volume_2.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
        #    Step 1: Attach volumes to VM     #
 | 
						|
        #######################################
 | 
						|
        #######################################
 | 
						|
 | 
						|
        self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            self.volume
 | 
						|
        )
 | 
						|
 | 
						|
        self.attached = True
 | 
						|
 | 
						|
        vol = sf_util.check_and_get_cs_volume(self, self.volume.id, self.testdata[TestData.volume_1][TestData.diskName], self)
 | 
						|
 | 
						|
        self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            volume_2
 | 
						|
        )
 | 
						|
 | 
						|
        vol_2 = sf_util.check_and_get_cs_volume(self, volume_2.id, self.testdata[TestData.volume_2][TestData.diskName], self)
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_size)
 | 
						|
 | 
						|
        sf_volume_2_size = sf_util.get_volume_size_with_hsr(self.cs_api, volume_2, self)
 | 
						|
 | 
						|
        self._verify_hsr(self.disk_offering.disksize, self.disk_offering.hypervisorsnapshotreserve, sf_volume_2_size)
 | 
						|
 | 
						|
        sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, self.primary_storage.id, self)
 | 
						|
 | 
						|
        sf_volumes = self._get_active_sf_volumes(sf_account_id)
 | 
						|
 | 
						|
        sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
 | 
						|
 | 
						|
        sf_iscsi_name = sf_util.get_iqn(self.cs_api, self.volume, self)
 | 
						|
 | 
						|
        vm = self._get_vm(self.virtual_machine.id)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name, vm.hostid)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume, sf_vag_id, self)
 | 
						|
 | 
						|
        sf_volume_2 = sf_util.check_and_get_sf_volume(sf_volumes, vol_2.name, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume_2, vol_2, sf_volume_2_size, self)
 | 
						|
 | 
						|
        sf_iscsi_name_2 = sf_util.get_iqn(self.cs_api, volume_2, self)
 | 
						|
 | 
						|
        self._check_host_side(sf_iscsi_name_2, vm.hostid)
 | 
						|
 | 
						|
        sf_util.check_vag(sf_volume_2, sf_vag_id, self)
 | 
						|
 | 
						|
        self.virtual_machine.detach_volume(self.apiClient, volume_2)
 | 
						|
 | 
						|
    def test_11_template_from_volume(self):
 | 
						|
        if TestData.hypervisor_type != TestData.kvm:
 | 
						|
            return
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            virtualmachineid=self.virtual_machine.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        vm_1_root_volume = list_volumes_response[0]
 | 
						|
 | 
						|
        services = {"displaytext": "Template-1", "name": "Template-1-name", "ostypeid": self.template.ostypeid, "ispublic": "true"}
 | 
						|
 | 
						|
        try:
 | 
						|
            Template.create_from_volume(self.apiClient, vm_1_root_volume, services)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._template_creation_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._volume_attached_to_non_stopped_vm_err_msg not in e.errorMsg:
 | 
						|
                raise
 | 
						|
 | 
						|
        self.virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        template = Template.create_from_volume(self.apiClient, vm_1_root_volume, services)
 | 
						|
 | 
						|
        self.cleanup.append(template)
 | 
						|
 | 
						|
        vol_snap = Snapshot.create(self.apiClient, volume_id=vm_1_root_volume.id)
 | 
						|
 | 
						|
        self.cleanup.append(vol_snap)
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        self._create_vm_using_template_and_destroy_vm(template)
 | 
						|
 | 
						|
        services = {"diskname": "Vol-1", "zoneid": self.testdata[TestData.zoneId], "ispublic": True}
 | 
						|
 | 
						|
        volume_created_from_snapshot = Volume.create_from_snapshot(self.apiClient, vol_snap.id, services, account=self.account.name, domainid=self.domain.id)
 | 
						|
 | 
						|
        self.cleanup.append(volume_created_from_snapshot)
 | 
						|
 | 
						|
        services = {"displaytext": "Template-2", "name": "Template-2-name", "ostypeid": self.template.ostypeid, "ispublic": "true"}
 | 
						|
 | 
						|
        template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services)
 | 
						|
 | 
						|
        self.cleanup.append(template)
 | 
						|
 | 
						|
        self._create_vm_using_template_and_destroy_vm(template)
 | 
						|
 | 
						|
        volume_created_from_snapshot = self.virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            volume_created_from_snapshot
 | 
						|
        )
 | 
						|
 | 
						|
        services = {"displaytext": "Template-3", "name": "Template-3-name", "ostypeid": self.template.ostypeid, "ispublic": "true"}
 | 
						|
 | 
						|
        try:
 | 
						|
            Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._template_creation_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._volume_attached_to_non_stopped_vm_err_msg not in e.errorMsg:
 | 
						|
                raise
 | 
						|
 | 
						|
        self.virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services)
 | 
						|
 | 
						|
        self.cleanup.append(template)
 | 
						|
 | 
						|
        volume_created_from_snapshot = self.virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            volume_created_from_snapshot
 | 
						|
        )
 | 
						|
 | 
						|
        TestVolumes._start_vm(self.virtual_machine)
 | 
						|
 | 
						|
        self._create_vm_using_template_and_destroy_vm(template)
 | 
						|
 | 
						|
        services = {"displaytext": "Template-4", "name": "Template-4-name", "ostypeid": self.template.ostypeid, "ispublic": "true"}
 | 
						|
 | 
						|
        template = Template.create_from_volume(self.apiClient, volume_created_from_snapshot, services)
 | 
						|
 | 
						|
        self.cleanup.append(template)
 | 
						|
 | 
						|
        self._create_vm_using_template_and_destroy_vm(template)
 | 
						|
 | 
						|
    def test_12_resize_volume_on_running_vm(self):
 | 
						|
        if TestData.hypervisor_type != TestData.kvm:
 | 
						|
            return
 | 
						|
 | 
						|
        test_virtual_machine = VirtualMachine.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.virtualMachine2],
 | 
						|
            accountid=self.account.name,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            serviceofferingid=self.compute_offering.id,
 | 
						|
            templateid=self.template.id,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            startvm=True
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_virtual_machine)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            virtualmachineid=test_virtual_machine.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        test_vm_root_volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._handle_root_volume_with_started_vm(test_vm_root_volume)
 | 
						|
 | 
						|
        volume_name = {
 | 
						|
            TestData.diskName: "test-volume-a",
 | 
						|
        }
 | 
						|
 | 
						|
        test_vm_data_volume = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            volume_name,
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_vm_data_volume)
 | 
						|
 | 
						|
        test_vm_data_volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_vm_data_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vm_data_volume = Volume(test_vm_data_volume.__dict__)
 | 
						|
 | 
						|
        self._handle_data_volume_with_started_vm(test_vm_data_volume)
 | 
						|
 | 
						|
        custom_disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Custom)",
 | 
						|
            "customized": True,
 | 
						|
            "customizediops": True,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        custom_disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            custom_disk_offering,
 | 
						|
            custom=True
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(custom_disk_offering)
 | 
						|
 | 
						|
        services = {
 | 
						|
            TestData.diskName: "test-volume-custom-a",
 | 
						|
            "customdisksize": 100,
 | 
						|
            "customminiops": 1000,
 | 
						|
            "custommaxiops": 2000,
 | 
						|
            "zoneid": self.testdata[TestData.zoneId]
 | 
						|
        }
 | 
						|
 | 
						|
        test_vm_data_volume = Volume.create_custom_disk(
 | 
						|
            self.apiClient,
 | 
						|
            services,
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            diskofferingid=custom_disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_vm_data_volume)
 | 
						|
 | 
						|
        test_vm_data_volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_vm_data_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vm_data_volume = Volume(test_vm_data_volume.__dict__)
 | 
						|
 | 
						|
        self._handle_custom_data_volume_with_started_vm(test_vm_data_volume)
 | 
						|
 | 
						|
    def test_13_resize_volume_on_stopped_vm(self):
 | 
						|
        if TestData.hypervisor_type != TestData.kvm:
 | 
						|
            return
 | 
						|
 | 
						|
        test_virtual_machine = VirtualMachine.create(
 | 
						|
            self.apiClient,
 | 
						|
            self.testdata[TestData.virtualMachine2],
 | 
						|
            accountid=self.account.name,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            serviceofferingid=self.compute_offering.id,
 | 
						|
            templateid=self.template.id,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            startvm=False
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_virtual_machine)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            virtualmachineid=test_virtual_machine.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        test_vm_root_volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        err_msg = "Check if SF volume was created in correct account"
 | 
						|
 | 
						|
        try:
 | 
						|
            # This should fail because there should not be an equivalent SolidFire volume in the cluster yet.
 | 
						|
            self._verify_volume(test_vm_root_volume)
 | 
						|
 | 
						|
            raise Exception("The volume verification did not fail (as expected).")
 | 
						|
        except Exception as e:
 | 
						|
            if err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        # Starting the up the should create its root disk on the SolidFire cluster.
 | 
						|
        test_virtual_machine.start(self.apiClient)
 | 
						|
 | 
						|
        test_virtual_machine.stop(self.apiClient)
 | 
						|
 | 
						|
        self._handle_root_volume_with_stopped_vm(test_vm_root_volume)
 | 
						|
 | 
						|
        volume_name = {
 | 
						|
            TestData.diskName: "test-volume-a",
 | 
						|
        }
 | 
						|
 | 
						|
        test_vm_data_volume = Volume.create(
 | 
						|
            self.apiClient,
 | 
						|
            volume_name,
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            diskofferingid=self.disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_vm_data_volume)
 | 
						|
 | 
						|
        test_vm_data_volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_vm_data_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vm_data_volume = Volume(test_vm_data_volume.__dict__)
 | 
						|
 | 
						|
        self._handle_data_volume_with_stopped_vm(test_vm_data_volume)
 | 
						|
 | 
						|
        custom_disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Custom)",
 | 
						|
            "customized": True,
 | 
						|
            "customizediops": True,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        custom_disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            custom_disk_offering,
 | 
						|
            custom=True
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(custom_disk_offering)
 | 
						|
 | 
						|
        services = {
 | 
						|
            TestData.diskName: "test-volume-custom-a",
 | 
						|
            "customdisksize": 100,
 | 
						|
            "customminiops": 1000,
 | 
						|
            "custommaxiops": 2000,
 | 
						|
            "zoneid": self.testdata[TestData.zoneId]
 | 
						|
        }
 | 
						|
 | 
						|
        test_vm_data_volume = Volume.create_custom_disk(
 | 
						|
            self.apiClient,
 | 
						|
            services,
 | 
						|
            account=self.account.name,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            diskofferingid=custom_disk_offering.id
 | 
						|
        )
 | 
						|
 | 
						|
        self.cleanup.append(test_vm_data_volume)
 | 
						|
 | 
						|
        test_vm_data_volume = test_virtual_machine.attach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_vm_data_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vm_data_volume = Volume(test_vm_data_volume.__dict__)
 | 
						|
 | 
						|
        self._handle_custom_data_volume_with_stopped_vm(test_vm_data_volume)
 | 
						|
 | 
						|
        test_vm_data_volume = test_virtual_machine.detach_volume(
 | 
						|
            self.apiClient,
 | 
						|
            test_vm_data_volume
 | 
						|
        )
 | 
						|
 | 
						|
        test_vm_data_volume = Volume(test_vm_data_volume.__dict__)
 | 
						|
 | 
						|
        self._handle_custom_data_volume_with_stopped_vm(test_vm_data_volume)
 | 
						|
 | 
						|
    def _handle_root_volume_with_started_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._only_data_volumes_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_B",
 | 
						|
            "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": volume.miniops,
 | 
						|
            "maxiops": volume.maxiops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._only_data_volumes_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_C",
 | 
						|
            "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": volume_size_in_GB,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_min_iops = new_min_iops + 10
 | 
						|
        new_max_iops = new_max_iops + 20
 | 
						|
 | 
						|
        volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _handle_root_volume_with_stopped_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_size = new_size + 10
 | 
						|
 | 
						|
        volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_min_iops = new_min_iops + 100
 | 
						|
        new_max_iops = new_max_iops + 200
 | 
						|
 | 
						|
        volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_size = new_size + 10
 | 
						|
        new_min_iops = new_min_iops + 100
 | 
						|
        new_max_iops = new_max_iops + 200
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a stopped VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._only_data_volumes_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_B",
 | 
						|
            "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": volume.miniops,
 | 
						|
            "maxiops": volume.maxiops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a stopped VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._only_data_volumes_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_C",
 | 
						|
            "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": volume_size_in_GB,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _handle_data_volume_with_started_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._to_change_volume_size_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._to_change_volume_size_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_B",
 | 
						|
            "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": volume.miniops,
 | 
						|
            "maxiops": volume.maxiops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._min_iops_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_C",
 | 
						|
            "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": volume_size_in_GB,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _handle_data_volume_with_stopped_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a stopped VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._to_change_volume_size_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a stopped VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._to_change_volume_size_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the IOPS of a volume attached to a stopped VM (should fail).
 | 
						|
            volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._min_iops_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_A",
 | 
						|
            "displaytext": "SF_DO_A (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_size = new_size + 10
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_B",
 | 
						|
            "displaytext": "SF_DO_B (Min IOPS = " + str(volume.miniops) + "; Max IOPS = " + str(volume.maxiops) + ")",
 | 
						|
            "disksize": new_size,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": volume.miniops,
 | 
						|
            "maxiops": volume.maxiops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_min_iops = new_min_iops + 100
 | 
						|
        new_max_iops = new_max_iops + 200
 | 
						|
 | 
						|
        disk_offering = {
 | 
						|
            "name": "SF_DO_C",
 | 
						|
            "displaytext": "SF_DO_C (Min IOPS = " + str(new_min_iops) + "; Max IOPS = " + str(new_max_iops) + ")",
 | 
						|
            "disksize": volume_size_in_GB,
 | 
						|
            "customizediops": False,
 | 
						|
            "miniops": new_min_iops,
 | 
						|
            "maxiops": new_max_iops,
 | 
						|
            "hypervisorsnapshotreserve": 200,
 | 
						|
            TestData.tags: TestData.storageTag,
 | 
						|
            "storagetype": "shared"
 | 
						|
        }
 | 
						|
 | 
						|
        disk_offering = DiskOffering.create(
 | 
						|
            self.apiClient,
 | 
						|
            disk_offering
 | 
						|
        )
 | 
						|
 | 
						|
        volume.resize(self.apiClient, diskofferingid=disk_offering.id)
 | 
						|
 | 
						|
        disk_offering.delete(self.apiClient)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _handle_custom_data_volume_with_started_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size and IOPS of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        try:
 | 
						|
            # Try to change the size of a volume attached to a running VM (should fail).
 | 
						|
            volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
            raise Exception(TestVolumes._volume_resize_did_not_fail_err_msg)
 | 
						|
        except Exception as e:
 | 
						|
            if TestVolumes._this_kind_of_disk_err_msg not in str(e):
 | 
						|
                raise
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _handle_custom_data_volume_with_stopped_vm(self, volume):
 | 
						|
        self._verify_volume(volume)
 | 
						|
 | 
						|
        volume_size_in_GB = volume.size / TestData.one_GB_in_bytes
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            type(volume_size_in_GB) == int,
 | 
						|
            TestVolumes._volume_size_not_an_int
 | 
						|
        )
 | 
						|
 | 
						|
        new_size = volume_size_in_GB + 10
 | 
						|
        new_min_iops = volume.miniops + 100
 | 
						|
        new_max_iops = volume.maxiops + 200
 | 
						|
 | 
						|
        volume.resize(self.apiClient, size=new_size, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
        new_size = new_size + 10
 | 
						|
 | 
						|
        volume.resize(self.apiClient, size=new_size)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        sf_volume_size = self._get_sf_volume(volume.name).total_size
 | 
						|
 | 
						|
        self._verify_volume(volume, volume.miniops, volume.maxiops, sf_volume_size)
 | 
						|
 | 
						|
        new_min_iops = volume.miniops + 10
 | 
						|
        new_max_iops = volume.maxiops + 20
 | 
						|
 | 
						|
        volume.resize(self.apiClient, miniops=new_min_iops, maxiops=new_max_iops)
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            id=volume.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        volume = Volume(list_volumes_response[0].__dict__)
 | 
						|
 | 
						|
        self._verify_volume(volume, new_min_iops, new_max_iops, sf_volume_size)
 | 
						|
 | 
						|
    def _get_sf_volume(self, volume_name):
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, "The SolidFire account ID should be a non-zero integer.")
 | 
						|
 | 
						|
        sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | 
						|
 | 
						|
        self.assertNotEqual(
 | 
						|
            len(sf_volumes),
 | 
						|
            0,
 | 
						|
            "The length of the response for the SolidFire-volume query should not be zero."
 | 
						|
        )
 | 
						|
 | 
						|
        return sf_util.check_and_get_sf_volume(sf_volumes, volume_name, self)
 | 
						|
 | 
						|
    def _verify_volume(self, cs_volume, expected_min_iops=None, expected_max_iops=None, expected_size=None):
 | 
						|
        sf_volume = self._get_sf_volume(cs_volume.name)
 | 
						|
 | 
						|
        sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, cs_volume, self)
 | 
						|
 | 
						|
        sf_util.check_size_and_iops(sf_volume, cs_volume, sf_volume_size, self)
 | 
						|
 | 
						|
        if expected_min_iops is not None:
 | 
						|
            self.assertEqual(
 | 
						|
                cs_volume.miniops,
 | 
						|
                expected_min_iops,
 | 
						|
                "Unexpected Min IOPS value (CloudStack volume has " + str(cs_volume.miniops) + "; expected " + str(expected_min_iops) + ")"
 | 
						|
            )
 | 
						|
 | 
						|
        if expected_max_iops is not None:
 | 
						|
            self.assertEqual(
 | 
						|
                cs_volume.maxiops,
 | 
						|
                expected_max_iops,
 | 
						|
                "Unexpected Max IOPS value (CloudStack volume has " + str(cs_volume.maxiops) + "; expected " + str(expected_max_iops) + ")"
 | 
						|
            )
 | 
						|
 | 
						|
        if expected_size is not None:
 | 
						|
            self.assertEqual(
 | 
						|
                sf_volume_size,
 | 
						|
                expected_size,
 | 
						|
                "Unexpected size value (CloudStack volume (with HSR) has " + str(sf_volume_size) + "; expected " + str(expected_size) + ")"
 | 
						|
            )
 | 
						|
 | 
						|
    def _create_vm_using_template_and_destroy_vm(self, template):
 | 
						|
        vm_name = "VM-%d" % random.randint(0, 100)
 | 
						|
 | 
						|
        virtual_machine_dict = {"name": vm_name, "displayname": vm_name}
 | 
						|
 | 
						|
        virtual_machine = VirtualMachine.create(
 | 
						|
            self.apiClient,
 | 
						|
            virtual_machine_dict,
 | 
						|
            accountid=self.account.name,
 | 
						|
            zoneid=self.zone.id,
 | 
						|
            serviceofferingid=self.compute_offering.id,
 | 
						|
            templateid=template.id,
 | 
						|
            domainid=self.domain.id,
 | 
						|
            startvm=True
 | 
						|
        )
 | 
						|
 | 
						|
        list_volumes_response = list_volumes(
 | 
						|
            self.apiClient,
 | 
						|
            virtualmachineid=virtual_machine.id,
 | 
						|
            listall=True
 | 
						|
        )
 | 
						|
 | 
						|
        sf_util.check_list(list_volumes_response, 1, self, TestVolumes._should_only_be_one_volume_in_list_err_msg)
 | 
						|
 | 
						|
        vm_root_volume = list_volumes_response[0]
 | 
						|
        vm_root_volume_name = vm_root_volume.name
 | 
						|
 | 
						|
        sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestVolumes._sf_account_id_should_be_non_zero_int_err_msg)
 | 
						|
 | 
						|
        # Get volume information from SolidFire cluster
 | 
						|
        sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | 
						|
 | 
						|
        sf_util.get_sf_volume_by_name(self, sf_volumes, vm_root_volume_name)
 | 
						|
 | 
						|
        virtual_machine.delete(self.apiClient, True)
 | 
						|
 | 
						|
    def _verify_hsr(self, cs_volume_size_in_gb, hsr, sf_volume_size_in_bytes):
 | 
						|
        cs_volume_size_including_hsr_in_bytes = self._get_cs_volume_size_including_hsr_in_bytes(cs_volume_size_in_gb, hsr)
 | 
						|
 | 
						|
        self.assertTrue(
 | 
						|
            cs_volume_size_including_hsr_in_bytes == sf_volume_size_in_bytes,
 | 
						|
            "HSR does not add up correctly."
 | 
						|
        )
 | 
						|
 | 
						|
    def _get_cs_volume_size_including_hsr_in_bytes(self, cs_volume_size_in_gb, hsr):
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            return self._get_bytes_from_gb(cs_volume_size_in_gb)
 | 
						|
 | 
						|
        lowest_hsr = 10
 | 
						|
 | 
						|
        if hsr < lowest_hsr:
 | 
						|
            hsr = lowest_hsr
 | 
						|
 | 
						|
        return self._get_bytes_from_gb(cs_volume_size_in_gb + (cs_volume_size_in_gb * (hsr / 100)))
 | 
						|
 | 
						|
    def _get_bytes_from_gb(self, number_in_gb):
 | 
						|
        return number_in_gb * 1024 * 1024 * 1024
 | 
						|
 | 
						|
    def _get_vm(self, vm_id):
 | 
						|
        list_vms_response = list_virtual_machines(self.apiClient, id=vm_id)
 | 
						|
 | 
						|
        sf_util.check_list(list_vms_response, 1, self, TestVolumes._should_only_be_one_vm_in_list_err_msg)
 | 
						|
 | 
						|
        return list_vms_response[0]
 | 
						|
 | 
						|
    def _check_xen_sr(self, xen_sr_name, should_exist=True):
 | 
						|
        sf_util.check_xen_sr(xen_sr_name, self.xen_session, self, should_exist)
 | 
						|
 | 
						|
    def _get_active_sf_volumes(self, sf_account_id=None):
 | 
						|
        sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
 | 
						|
 | 
						|
        self.assertNotEqual(
 | 
						|
            len(sf_volumes),
 | 
						|
            0,
 | 
						|
            TestVolumes._volume_response_should_not_be_zero_err_msg
 | 
						|
        )
 | 
						|
 | 
						|
        return sf_volumes
 | 
						|
 | 
						|
    def _get_template_cache_name(self):
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            return TestData.templateCacheNameKvm
 | 
						|
        elif TestData.hypervisor_type == TestData.xenServer:
 | 
						|
            return TestData.templateCacheNameXenServer
 | 
						|
 | 
						|
        self.assert_(False, "Invalid hypervisor type")
 | 
						|
 | 
						|
    def _get_modified_iscsi_name(self, sf_iscsi_name):
 | 
						|
        sf_iscsi_name = sf_iscsi_name.replace("/", "")
 | 
						|
 | 
						|
        return sf_iscsi_name[:-1]
 | 
						|
 | 
						|
    def _check_host_side(self, sf_iscsi_name, vm_hostid=None, should_exist=True):
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            self._check_kvm_host_side(self._get_modified_iscsi_name(sf_iscsi_name), vm_hostid, should_exist)
 | 
						|
        elif TestData.hypervisor_type == TestData.xenServer:
 | 
						|
            self._check_xen_sr(sf_iscsi_name, should_exist)
 | 
						|
 | 
						|
    def _check_kvm_host_side(self, sf_iscsi_name, vm_hostid, should_exist=True):
 | 
						|
        if vm_hostid is None:
 | 
						|
            list_hosts_response = list_hosts(
 | 
						|
                self.apiClient,
 | 
						|
                type="Routing"
 | 
						|
            )
 | 
						|
        else:
 | 
						|
            list_hosts_response = list_hosts(
 | 
						|
                self.apiClient,
 | 
						|
                id=vm_hostid
 | 
						|
            )
 | 
						|
 | 
						|
            sf_util.check_list(list_hosts_response, 1, self, TestVolumes._should_only_be_one_host_in_list_err_msg)
 | 
						|
 | 
						|
        kvm_login = self.testdata[TestData.kvm]
 | 
						|
 | 
						|
        for cs_host in list_hosts_response:
 | 
						|
            ssh_connection = sf_util.get_ssh_connection(cs_host.ipaddress, kvm_login[TestData.username], kvm_login[TestData.password])
 | 
						|
 | 
						|
            stdout = ssh_connection.exec_command("ls /dev/disk/by-path | grep " + sf_iscsi_name)[1]
 | 
						|
 | 
						|
            result = stdout.read()
 | 
						|
 | 
						|
            ssh_connection.close()
 | 
						|
 | 
						|
            if should_exist:
 | 
						|
                self.assertFalse(result is None, "Unable to locate 'by-path' field on the KVM host (None)")
 | 
						|
                self.assertFalse(len(result.strip()) <= len(sf_iscsi_name), "Unable to locate the 'by-path' field on the KVM host (Zero-length string)")
 | 
						|
            else:
 | 
						|
                self.assertTrue(result is None or len(result.strip()) == 0, "Found the 'by-path' field on the KVM host, but did not expect to")
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def _start_vm(cls, vm):
 | 
						|
        vm_for_check = list_virtual_machines(
 | 
						|
            cls.apiClient,
 | 
						|
            id=vm.id
 | 
						|
        )[0]
 | 
						|
 | 
						|
        if vm_for_check.state == VirtualMachine.STOPPED:
 | 
						|
            vm.start(cls.apiClient)
 | 
						|
 | 
						|
            # Libvirt appears to have an issue detaching a volume from a VM while the VM is booting up.
 | 
						|
            # The XML sent to update the VM seems correct, but it doesn't appear to update the XML that describes the VM.
 | 
						|
            # For KVM, just give it 90 seconds to boot up.
 | 
						|
            if TestData.hypervisor_type == TestData.kvm:
 | 
						|
                time.sleep(90)
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def _reboot_vm(cls, vm):
 | 
						|
        vm.reboot(cls.apiClient)
 | 
						|
 | 
						|
        # Libvirt appears to have an issue detaching a volume from a VM while the VM is booting up.
 | 
						|
        # The XML sent to update the VM seems correct, but it doesn't appear to update the XML that describes the VM.
 | 
						|
        # For KVM, just give it 90 seconds to boot up.
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            time.sleep(90)
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def _handle_supports_cloning(cls):
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            cls._supports_cloning = True
 | 
						|
        elif TestData.hypervisor_type == TestData.xenServer:
 | 
						|
            # For XenServer, it is OK to set this to True or False depending on what you'd like tested
 | 
						|
            cls._supports_cloning = True
 | 
						|
 | 
						|
            sf_util.set_supports_resign(cls._supports_cloning, cls.dbConnection)
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def _connect_to_hypervisor(cls):
 | 
						|
        if TestData.hypervisor_type == TestData.kvm:
 | 
						|
            pass
 | 
						|
        elif TestData.hypervisor_type == TestData.xenServer:
 | 
						|
            host_ip = "https://" + \
 | 
						|
                  list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId], name=TestData.xen_server_hostname)[0].ipaddress
 | 
						|
 | 
						|
            cls.xen_session = XenAPI.Session(host_ip)
 | 
						|
 | 
						|
            xen_server = cls.testdata[TestData.xenServer]
 | 
						|
 | 
						|
            cls.xen_session.xenapi.login_with_password(xen_server[TestData.username], xen_server[TestData.password])
 | 
						|
 |