mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 08:42:29 +01:00
* cleanup plan * more robust cleanup within member method * ss_domain_limits fixed * egress 'icmp' instead of 'all' and cleanup * configdrive syntax, fixed to be up to par with py2 * cleanup fixed the lb secondary ip tests? * deal with different base64 encoding in py3 for userdata * cleanup of multiple_ips_per_nic * cleanup and reformat of test_volumes * cleanup and fixes for test_ps_domain_limits.py * cleanup and fix of test_ps_limits.py * fix occasional match of float against int * cleanup and fix test_ps_resize_volume.py * cleanup and fix test_snapshots * cleanup ss_max_limits and fix for float vs int problem in API * mere cleanup of test_volume_destroy_recover * add missing command creation * cleanup of test_vpc_on_host_maintenance * cleanup test_vpc_network * cleanup, comments and logging in test_vpc_network_lbrules * cleanup of test_vpc_network_pfrules * cleanup and format code for test_vpc_network_staticnatrule * f string instead of conversion specifiers * check http and ssh fix and cleanup (for vpc pfrules tests) * generalise create network method * make ip optional in creating webserver * remove unused code and add rules to enable ssh * more cleanup * remove unused code and cleanup * small cleanup, mostly precarous run environment required * cleanup and removed unused code * advancedsg only, cleanup, pulled in services * reformat/cleanup * log result of error after verify rule * add nw_off_no_services * tags=["TODO"] for escalations_networks * tags=["TODO"] for organization_states * tags=["TODO"] for browse_templates * tags=["TODO"] for configdrive * tags=["TODO"] for vpc_vms_deployment * add remove network cleanup and fixes * move tests that fail on all platforms out of the way Co-authored-by: Daan Hoogland <dahn@onecht.net>
1544 lines
53 KiB
Python
1544 lines
53 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
""" P1 tests for Volumes
|
|
"""
|
|
# Import Local Modules
|
|
from nose.plugins.attrib import attr
|
|
from marvin.cloudstackTestCase import cloudstackTestCase
|
|
from marvin.cloudstackAPI import (listHypervisorCapabilities,
|
|
attachIso,
|
|
deleteVolume
|
|
)
|
|
from marvin.lib.utils import cleanup_resources, validateList
|
|
from marvin.lib.base import (Account,
|
|
ServiceOffering,
|
|
VirtualMachine,
|
|
Volume,
|
|
Host,
|
|
Iso,
|
|
Configurations,
|
|
DiskOffering,
|
|
Domain,
|
|
StoragePool)
|
|
from marvin.lib.common import (get_domain,
|
|
get_zone,
|
|
get_template,
|
|
get_pod,
|
|
find_storage_pool_type,
|
|
update_resource_limit)
|
|
from marvin.codes import PASS
|
|
# Import System modules
|
|
import time
|
|
|
|
|
|
class TestAttachVolume(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestAttachVolume, cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.pod = get_pod(cls.api_client, cls.zone.id)
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"]
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
cls.template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
# get max data volumes limit based on the hypervisor type and version
|
|
listHost = Host.list(
|
|
cls.api_client,
|
|
type='Routing',
|
|
zoneid=cls.zone.id,
|
|
podid=cls.pod.id,
|
|
)
|
|
ver = listHost[0].hypervisorversion
|
|
hv = listHost[0].hypervisor
|
|
cmd = listHypervisorCapabilities.listHypervisorCapabilitiesCmd()
|
|
cmd.hypervisor = hv
|
|
res = cls.api_client.listHypervisorCapabilities(cmd)
|
|
cls.debug('Hypervisor Capabilities: {}'.format(res))
|
|
for i in range(len(res)):
|
|
if res[i].hypervisorversion == ver:
|
|
break
|
|
cls.max_data_volumes = int(res[i].maxdatavolumeslimit)
|
|
if 'kvm' in cls.hypervisor:
|
|
cls.max_data_volumes = 24
|
|
cls.debug('max data volumes:{}'.format(cls.max_data_volumes))
|
|
cls.testdata["volume"]["max"] = cls.max_data_volumes
|
|
# Create VMs, NAT Rules etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
update_resource_limit(
|
|
cls.api_client,
|
|
2, # Instance
|
|
account=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
max=cls.max_data_volumes + 1
|
|
)
|
|
|
|
cls.debug('max data volumes:{}'.format(cls.max_data_volumes))
|
|
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
cls.virtual_machine = VirtualMachine.create(
|
|
cls.api_client,
|
|
cls.testdata["virtual_machine"],
|
|
accountid=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
serviceofferingid=cls.service_offering.id,
|
|
templateid=cls.template.id,
|
|
zoneid=cls.zone.id
|
|
)
|
|
cls._cleanup.append(cls.virtual_machine)
|
|
|
|
def setUp(self):
|
|
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("Skipping because of unsupported storage type")
|
|
|
|
def tearDown(self):
|
|
super(TestAttachVolume, self).tearDown()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestAttachVolume, cls).tearDownClass()
|
|
|
|
@attr(tags=["advanced", "advancedns", "needle"])
|
|
def test_01_volume_attach(self):
|
|
"""Test Attach volumes (max capacity)
|
|
"""
|
|
# Validate the following
|
|
# 1. Deploy a vm and create 5 data disk
|
|
# 2. Attach all the created Volume to the vm.
|
|
# 3. Reboot the VM. VM should be successfully rebooted
|
|
# 4. Stop the VM. Stop VM should be successful
|
|
# 5. Start The VM. Start VM should be successful
|
|
|
|
# Create 5 volumes and attach to VM
|
|
|
|
for i in range(self.max_data_volumes):
|
|
volume = Volume.create(
|
|
self.apiclient,
|
|
self.testdata["volume"],
|
|
zoneid=self.zone.id,
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
# Check List Volume response for newly created volume
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
# Attach volume to VM
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
volume
|
|
)
|
|
# Check all volumes attached to same VM
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='DATADISK',
|
|
listall=True
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes")
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list")
|
|
self.assertEqual(
|
|
len(list_volume_response),
|
|
self.max_data_volumes,
|
|
"Volumes attached to the VM %s. Expected %s" %
|
|
(len(list_volume_response),
|
|
self.max_data_volumes))
|
|
self.debug("Rebooting the VM: %s" % self.virtual_machine.id)
|
|
# Reboot VM
|
|
self.virtual_machine.reboot(self.apiclient)
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertNotEqual(
|
|
len(vm_response),
|
|
0,
|
|
"Check VMs available in List VMs response"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(vm_response, list),
|
|
True,
|
|
"Check list VM response for valid list"
|
|
)
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Running',
|
|
"Check the state of VM"
|
|
)
|
|
|
|
# Stop VM
|
|
self.virtual_machine.stop(self.apiclient)
|
|
|
|
# Start VM
|
|
self.virtual_machine.start(self.apiclient)
|
|
# Sleep to ensure that VM is in ready state
|
|
time.sleep(self.testdata["sleep"])
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
self.assertEqual(
|
|
isinstance(vm_response, list),
|
|
True,
|
|
"Check list VM response for valid list"
|
|
)
|
|
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertNotEqual(
|
|
len(vm_response),
|
|
0,
|
|
"Check VMs available in List VMs response"
|
|
)
|
|
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Running',
|
|
"Check the state of VM"
|
|
)
|
|
return
|
|
|
|
@attr(tags=["advanced", "advancedns"])
|
|
def test_02_volume_attach_max(self):
|
|
"""Test attach volumes (more than max) to an instance
|
|
|
|
# Validate the following
|
|
# 1. Attach one more data volume to VM (Already 5 attached)
|
|
# 2. Attach volume should fail
|
|
"""
|
|
|
|
# Create a volume and attach to VM
|
|
|
|
# Update limit so that account could create one more volume
|
|
if 'kvm' in self.hypervisor:
|
|
update_resource_limit(
|
|
self.api_client,
|
|
2, # Instance
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
max=32
|
|
)
|
|
# Attach volume to VM
|
|
with self.assertRaises(Exception):
|
|
for i in range(self.max_data_volumes):
|
|
volume = Volume.create(
|
|
self.apiclient,
|
|
self.testdata["volume"],
|
|
zoneid=self.zone.id,
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
# Check List Volume response for newly created volume
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
# Attach volume to VM
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
volume
|
|
)
|
|
return
|
|
|
|
|
|
class TestAttachDetachVolume(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestAttachDetachVolume, cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.pod = get_pod(cls.api_client, cls.zone.id)
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"]
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
cls.testdata["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["template"] = template.id
|
|
# get max data volumes limit based on the hypervisor type and version
|
|
listHost = Host.list(
|
|
cls.api_client,
|
|
type='Routing',
|
|
zoneid=cls.zone.id,
|
|
podid=cls.pod.id,
|
|
)
|
|
ver = listHost[0].hypervisorversion
|
|
hv = listHost[0].hypervisor
|
|
cmd = listHypervisorCapabilities.listHypervisorCapabilitiesCmd()
|
|
cmd.hypervisor = hv # cls.services["virtual_machine"]["hypervisor"]
|
|
res = cls.api_client.listHypervisorCapabilities(cmd)
|
|
cls.debug('Hypervisor Capabilities: {}'.format(res))
|
|
for i in range(len(res)):
|
|
if res[i].hypervisorversion == ver:
|
|
break
|
|
cls.max_data_volumes = int(res[i].maxdatavolumeslimit)
|
|
if 'kvm' in cls.hypervisor:
|
|
cls.max_data_volumes = 24
|
|
cls.debug('max data volumes:{}'.format(cls.max_data_volumes))
|
|
cls.testdata["volume"]["max"] = cls.max_data_volumes
|
|
# Create VMs, NAT Rules etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
update_resource_limit(
|
|
cls.api_client,
|
|
2, # Instance
|
|
account=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
max=cls.max_data_volumes + 1
|
|
)
|
|
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
cls.virtual_machine = VirtualMachine.create(
|
|
cls.api_client,
|
|
cls.testdata["virtual_machine"],
|
|
accountid=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
serviceofferingid=cls.service_offering.id,
|
|
)
|
|
cls._cleanup.append(cls.virtual_machine)
|
|
|
|
def setUp(self):
|
|
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("RBD storage type is required for data volumes for LXC")
|
|
|
|
def tearDown(self):
|
|
super(TestAttachDetachVolume, self).tearDown()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestAttachDetachVolume, cls).tearDownClass()
|
|
|
|
@attr(tags=["advanced", "advancedns"])
|
|
def test_01_volume_attach_detach(self):
|
|
"""Test Volume attach/detach to VM (5 data volumes)
|
|
"""
|
|
|
|
# Validate the following
|
|
# 1. Deploy a vm and create 5 data disk
|
|
# 2. Attach all the created Volume to the vm.
|
|
# 3. Detach all the volumes attached.
|
|
# 4. Reboot the VM. VM should be successfully rebooted
|
|
# 5. Stop the VM. Stop VM should be successful
|
|
# 6. Start The VM. Start VM should be successful
|
|
|
|
volumes = []
|
|
# Create 5 volumes and attach to VM
|
|
for i in range(self.max_data_volumes):
|
|
volume = Volume.create(
|
|
self.apiclient,
|
|
self.testdata["volume"],
|
|
zoneid=self.zone.id,
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
self.cleanup.append(volume)
|
|
volumes.append(volume)
|
|
|
|
# Check List Volume response for newly created volume
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes")
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list")
|
|
# Attach volume to VM
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
volume
|
|
)
|
|
|
|
# Check all volumes attached to same VM
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='DATADISK',
|
|
listall=True
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
self.assertEqual(
|
|
len(list_volume_response),
|
|
self.max_data_volumes,
|
|
"Volumes attached to the VM %s. Expected %s" %
|
|
(len(list_volume_response),
|
|
self.max_data_volumes))
|
|
|
|
# Detach all volumes from VM
|
|
for volume in volumes:
|
|
self.virtual_machine.detach_volume(
|
|
self.apiclient,
|
|
volume
|
|
)
|
|
# Reboot VM
|
|
self.debug("Rebooting the VM: %s" % self.virtual_machine.id)
|
|
self.virtual_machine.reboot(self.apiclient)
|
|
# Sleep to ensure that VM is in ready state
|
|
time.sleep(self.testdata["sleep"])
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertEqual(
|
|
isinstance(vm_response, list),
|
|
True,
|
|
"Check list VM response for valid list"
|
|
)
|
|
|
|
self.assertNotEqual(
|
|
len(vm_response),
|
|
0,
|
|
"Check VMs available in List VMs response"
|
|
)
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Running',
|
|
"Check the state of VM"
|
|
)
|
|
|
|
# Stop VM
|
|
self.virtual_machine.stop(self.apiclient)
|
|
|
|
# Start VM
|
|
self.virtual_machine.start(self.apiclient)
|
|
# Sleep to ensure that VM is in ready state
|
|
time.sleep(self.testdata["sleep"])
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertEqual(
|
|
isinstance(vm_response, list),
|
|
True,
|
|
"Check list VM response for valid list"
|
|
)
|
|
self.assertNotEqual(
|
|
len(vm_response),
|
|
0,
|
|
"Check VMs available in List VMs response"
|
|
)
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Running',
|
|
"Check the state of VM"
|
|
)
|
|
|
|
return
|
|
|
|
@attr(tags=["advanced", "advancedns"], required_hardware="false")
|
|
def test_02_root_volume_attach_detach(self):
|
|
"""Test Root Volume attach/detach to VM
|
|
# Validate the following
|
|
# 1. Deploy a VM
|
|
# 2. Verify that we are testing a supported hypervisor
|
|
# 3. Check for root volume
|
|
# 4. Stop VM
|
|
# 5. Detach root volume
|
|
# 6. Verify root volume detached
|
|
# 7. Attach root volume
|
|
# 8. Start VM
|
|
"""
|
|
|
|
if (self.hypervisor.lower() == 'vmware'
|
|
or self.hypervisor.lower() == 'kvm'
|
|
or self.hypervisor.lower() == 'simulator'
|
|
or self.hypervisor.lower() == 'xenserver'):
|
|
root_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='ROOT',
|
|
listall=True
|
|
)
|
|
|
|
self.assertEqual(
|
|
validateList(root_volume_response)[0],
|
|
PASS,
|
|
"Invalid response returned for root volume list"
|
|
)
|
|
|
|
# Grab the root volume for later use
|
|
root_volume = root_volume_response[0]
|
|
|
|
# Stop VM
|
|
self.debug("Stopping the VM: %s" % self.virtual_machine.id)
|
|
self.virtual_machine.stop(self.apiclient)
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
|
|
# Ensure that vm_response is a valid list
|
|
self.assertEqual(
|
|
validateList(vm_response)[0],
|
|
PASS,
|
|
"Invalid response returned for vm_response list"
|
|
)
|
|
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Stopped',
|
|
"Check the state of VM"
|
|
)
|
|
|
|
# Detach root volume from VM
|
|
self.virtual_machine.detach_volume(
|
|
self.apiclient,
|
|
root_volume
|
|
)
|
|
|
|
# Verify that root disk is gone
|
|
no_root_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='ROOT',
|
|
listall=True
|
|
)
|
|
|
|
self.assertEqual(
|
|
no_root_volume_response,
|
|
None,
|
|
"Check if root volume exists in ListVolumes"
|
|
)
|
|
|
|
# Attach root volume to VM
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
root_volume,
|
|
0
|
|
)
|
|
|
|
# Check for root volume
|
|
new_root_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='ROOT',
|
|
listall=True
|
|
)
|
|
|
|
# Ensure that new_root_volume_response is a valid list
|
|
self.assertEqual(
|
|
validateList(new_root_volume_response)[0],
|
|
PASS,
|
|
"Invalid response returned for new_root_volume_response list"
|
|
)
|
|
|
|
# Start VM
|
|
self.virtual_machine.start(self.apiclient)
|
|
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertEqual(
|
|
validateList(vm_response)[0],
|
|
PASS,
|
|
"Invalid response returned for vm_response list during VM start up"
|
|
)
|
|
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.state,
|
|
'Running',
|
|
"Ensure the state of VM is running"
|
|
)
|
|
|
|
else:
|
|
self.skipTest("Root Volume attach/detach is not supported on %s " % self.hypervisor)
|
|
return
|
|
|
|
|
|
class TestAttachVolumeISO(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestAttachVolumeISO, cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.pod = get_pod(cls.api_client, cls.zone.id)
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"]
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
cls.testdata["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.testdata["iso"]["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["template"] = template.id
|
|
# get max data volumes limit based on the hypervisor type and version
|
|
listHost = Host.list(
|
|
cls.api_client,
|
|
type='Routing',
|
|
zoneid=cls.zone.id,
|
|
podid=cls.pod.id,
|
|
)
|
|
ver = listHost[0].hypervisorversion
|
|
hv = listHost[0].hypervisor
|
|
cmd = listHypervisorCapabilities.listHypervisorCapabilitiesCmd()
|
|
cmd.hypervisor = hv
|
|
res = cls.api_client.listHypervisorCapabilities(cmd)
|
|
cls.debug('Hypervisor Capabilities: {}'.format(res))
|
|
for i in range(len(res)):
|
|
if res[i].hypervisorversion == ver:
|
|
break
|
|
cls.max_data_volumes = int(res[i].maxdatavolumeslimit)
|
|
if 'kvm' in cls.hypervisor:
|
|
cls.max_data_volumes = 24
|
|
cls.debug('max data volumes:{}'.format(cls.max_data_volumes))
|
|
cls.testdata["volume"]["max"] = cls.max_data_volumes
|
|
# Create VMs, NAT Rules etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
update_resource_limit(
|
|
cls.api_client,
|
|
2, # Instance
|
|
account=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
max=cls.max_data_volumes + 1
|
|
)
|
|
cls.debug('max data volumes:{}'.format(cls.max_data_volumes))
|
|
cls.testdata["volume"]["max"] = cls.max_data_volumes
|
|
# Create VMs, NAT Rules etc
|
|
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
cls.virtual_machine = VirtualMachine.create(
|
|
cls.api_client,
|
|
cls.testdata["virtual_machine"],
|
|
accountid=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
serviceofferingid=cls.service_offering.id,
|
|
)
|
|
cls._cleanup.append(cls.virtual_machine)
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestAttachVolumeISO, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("RBD storage type is required for data volumes for LXC")
|
|
|
|
def tearDown(self):
|
|
super(TestAttachVolumeISO, self).tearDown()
|
|
|
|
@attr(tags=["advanced", "advancedns"], required_hardware="true")
|
|
def test_01_volume_iso_attach(self):
|
|
"""Test Volumes and ISO attach
|
|
# Validate the following
|
|
# 1. Create and attach 5 data volumes to VM
|
|
# 2. Create an ISO. Attach it to VM instance
|
|
# 3. Verify that attach ISO is successful
|
|
"""
|
|
|
|
# Create 5 volumes and attach to VM
|
|
if self.hypervisor.lower() in ["lxc"]:
|
|
self.skipTest("attach ISO is not supported on LXC")
|
|
for i in range(self.max_data_volumes):
|
|
volume = Volume.create(
|
|
self.apiclient,
|
|
self.testdata["volume"],
|
|
zoneid=self.zone.id,
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
self.cleanup.append(volume)
|
|
self.debug("Created volume: %s for account: %s" % (
|
|
volume.id,
|
|
self.account.name
|
|
))
|
|
# Check List Volume response for newly created volume
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
# Attach volume to VM
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
volume
|
|
)
|
|
self.cleanup.remove(volume)
|
|
|
|
# Check all volumes attached to same VM
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='DATADISK',
|
|
listall=True
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
self.assertEqual(
|
|
len(list_volume_response),
|
|
self.max_data_volumes,
|
|
"Volumes attached to the VM %s. Expected %s" %
|
|
(len(list_volume_response),
|
|
self.max_data_volumes))
|
|
# Create an ISO and attach it to VM
|
|
iso = Iso.create(
|
|
self.apiclient,
|
|
self.testdata["iso"],
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
)
|
|
self.cleanup.append(iso)
|
|
self.debug("Created ISO with ID: %s for account: %s" % (
|
|
iso.id,
|
|
self.account.name
|
|
))
|
|
|
|
try:
|
|
self.debug("Downloading ISO with ID: %s" % iso.id)
|
|
iso.download(self.apiclient)
|
|
except Exception as e:
|
|
self.fail("Exception while downloading ISO %s: %s"
|
|
% (iso.id, e))
|
|
|
|
# Attach ISO to virtual machine
|
|
self.debug("Attach ISO ID: %s to VM: %s" % (
|
|
iso.id,
|
|
self.virtual_machine.id
|
|
))
|
|
cmd = attachIso.attachIsoCmd()
|
|
cmd.id = iso.id
|
|
cmd.virtualmachineid = self.virtual_machine.id
|
|
self.apiclient.attachIso(cmd)
|
|
self.cleanup.remove(iso)
|
|
|
|
# Verify ISO is attached to VM
|
|
vm_response = VirtualMachine.list(
|
|
self.apiclient,
|
|
id=self.virtual_machine.id,
|
|
)
|
|
# Verify VM response to check whether VM deployment was successful
|
|
self.assertEqual(
|
|
isinstance(vm_response, list),
|
|
True,
|
|
"Check list VM response for valid list"
|
|
)
|
|
|
|
self.assertNotEqual(
|
|
len(vm_response),
|
|
0,
|
|
"Check VMs available in List VMs response"
|
|
)
|
|
vm = vm_response[0]
|
|
self.assertEqual(
|
|
vm.isoid,
|
|
iso.id,
|
|
"Check ISO is attached to VM or not"
|
|
)
|
|
return
|
|
|
|
|
|
class TestVolumes(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestVolumes, cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"]
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
cls.testdata["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["template"] = template.id
|
|
cls.testdata["virtual_machine"][
|
|
"diskofferingid"] = cls.disk_offering.id
|
|
|
|
# Create VMs, VMs etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
cls.virtual_machine = VirtualMachine.create(
|
|
cls.api_client,
|
|
cls.testdata["virtual_machine"],
|
|
accountid=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
serviceofferingid=cls.service_offering.id,
|
|
startvm=False
|
|
)
|
|
cls._cleanup.append(cls.virtual_machine)
|
|
|
|
cls.volume = Volume.create(
|
|
cls.api_client,
|
|
cls.testdata["volume"],
|
|
zoneid=cls.zone.id,
|
|
account=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
diskofferingid=cls.disk_offering.id
|
|
)
|
|
cls._cleanup.append(cls.volume)
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestVolumes, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("RBD storage type is required for data volumes for LXC")
|
|
|
|
def tearDown(self):
|
|
super(TestVolumes, self).tearDown()
|
|
|
|
@attr(tags=["advanced", "advancedns"], required_hardware="false")
|
|
def test_01_attach_volume(self):
|
|
"""Attach a created Volume to a Running VM
|
|
# Validate the following
|
|
# Create a data volume. (this is being done in setup)
|
|
# 2. List Volumes should not have vmname and virtualmachineid fields in
|
|
# response before volume attach (to VM)
|
|
# 3. Attch volume to VM. Attach volume should be successful.
|
|
# 4. List Volumes should have vmname and virtualmachineid fields in
|
|
# response before volume attach (to VM)
|
|
FIXME: This test (method) creates a precondition for `test_02_detach_volume()`:
|
|
an attached `self.volume` to `self.virtual_machine`
|
|
"""
|
|
|
|
# Check the list volumes response for vmname and virtualmachineid
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=self.volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
volume = list_volume_response[0]
|
|
|
|
self.assertEqual(
|
|
volume.type,
|
|
'DATADISK',
|
|
"Check volume type from list volume response"
|
|
)
|
|
|
|
self.assertEqual(
|
|
hasattr(volume, 'vmname'),
|
|
True,
|
|
"Check whether volume has vmname field"
|
|
)
|
|
self.assertEqual(
|
|
hasattr(volume, 'virtualmachineid'),
|
|
True,
|
|
"Check whether volume has virtualmachineid field"
|
|
)
|
|
|
|
# Attach volume to VM
|
|
self.debug("Attach volume: %s to VM: %s" % (
|
|
self.volume.id,
|
|
self.virtual_machine.id
|
|
))
|
|
self.virtual_machine.attach_volume(self.apiclient, self.volume)
|
|
self._cleanup.remove(self.volume)
|
|
|
|
# Check all volumes attached to same VM
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
type='DATADISK',
|
|
listall=True
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
volume = list_volume_response[0]
|
|
self.assertEqual(
|
|
volume.vmname,
|
|
self.virtual_machine.name,
|
|
"Check virtual machine name in list volumes response"
|
|
)
|
|
self.assertEqual(
|
|
volume.virtualmachineid,
|
|
self.virtual_machine.id,
|
|
"Check VM ID in list Volume response"
|
|
)
|
|
return
|
|
|
|
@attr(tags=["advanced", "advancedns"], required_hardware="false")
|
|
def test_02_detach_volume(self):
|
|
"""Detach a Volume attached to a VM
|
|
FIXME: this test method assumes a precondition set in `self.test_01_atttach_volume()`:
|
|
there is expected to be a VM with a disk attached
|
|
# Validate the following
|
|
# 1. Data disk should be detached from instance
|
|
# 2. Listvolumes should not have vmname and virtualmachineid fields for
|
|
# that volume.
|
|
FIXME: this test (method) leaves a precondition for `self.test_03_delete_detached_volume()`:
|
|
|
|
"""
|
|
|
|
self.debug("Detach volume: %s to VM: %s" % (
|
|
self.volume.id,
|
|
self.virtual_machine.id
|
|
))
|
|
self.virtual_machine.detach_volume(self.apiclient, self.volume)
|
|
self._cleanup.append(self.volume)
|
|
|
|
# Sleep to ensure the current state will reflected in other calls
|
|
time.sleep(self.testdata["sleep"])
|
|
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=self.volume.id
|
|
)
|
|
self.assertNotEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Check if volume exists in ListVolumes"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(list_volume_response, list),
|
|
True,
|
|
"Check list volumes response for valid list"
|
|
)
|
|
volume = list_volume_response[0]
|
|
self.assertEqual(
|
|
volume.virtualmachineid,
|
|
None,
|
|
"Check if volume state (detached) is reflected"
|
|
)
|
|
|
|
self.assertEqual(
|
|
volume.vmname,
|
|
None,
|
|
"Check if volume state (detached) is reflected"
|
|
)
|
|
return
|
|
|
|
@attr(tags=["advanced", "advancedns"], required_hardware="false")
|
|
def test_03_delete_detached_volume(self):
|
|
"""Delete a Volume unattached to an VM
|
|
FIXME: this expects `test_02_detach_volume()` to leave `self.volume` free and ready to delete
|
|
# Validate the following
|
|
# 1. volume should be deleted successfully and listVolume should not
|
|
# contain the deleted volume details.
|
|
"""
|
|
|
|
self.debug("Deleting volume: %s" % self.volume.id)
|
|
cmd = deleteVolume.deleteVolumeCmd()
|
|
cmd.id = self.volume.id
|
|
self.apiclient.deleteVolume(cmd)
|
|
self._cleanup.remove(self.volume)
|
|
|
|
# Sleep to ensure the current state will reflected in other calls
|
|
time.sleep(self.testdata["sleep"])
|
|
|
|
list_volume_response = Volume.list(
|
|
self.apiclient,
|
|
id=self.volume.id,
|
|
)
|
|
self.assertEqual(
|
|
list_volume_response,
|
|
None,
|
|
"Volume %s was not deleted" % self.volume.id
|
|
)
|
|
return
|
|
|
|
@attr(
|
|
tags=[
|
|
"advanced",
|
|
"advancedns",
|
|
"simulator",
|
|
"basic",
|
|
"eip",
|
|
"sg"],
|
|
required_hardware="false")
|
|
def test_create_volume_under_domain(self):
|
|
"""Create a volume under a non-root domain as non-root-domain user
|
|
|
|
1. Create a domain under ROOT
|
|
2. Create a user within this domain
|
|
3. As user in step 2. create a volume with standard disk offering
|
|
4. Ensure the volume is created in the domain and available to the
|
|
user in their listVolumes call
|
|
"""
|
|
dom = Domain.create(
|
|
self.apiclient,
|
|
services={},
|
|
name="NROOT",
|
|
parentdomainid=self.domain.id
|
|
)
|
|
self.cleanup.append(dom)
|
|
self.assertTrue(dom is not None, msg="Domain creation failed")
|
|
|
|
domuser = Account.create(
|
|
apiclient=self.apiclient,
|
|
services=self.testdata["account"],
|
|
admin=False,
|
|
domainid=dom.id
|
|
)
|
|
self.cleanup.append(domuser)
|
|
self.assertTrue(domuser is not None)
|
|
|
|
domapiclient = self.testClient.getUserApiClient(
|
|
UserName=domuser.name,
|
|
DomainName=dom.name)
|
|
|
|
diskoffering = DiskOffering.list(self.apiclient)
|
|
self.assertTrue(
|
|
isinstance(
|
|
diskoffering,
|
|
list),
|
|
msg="DiskOffering list is not a list?")
|
|
self.assertTrue(
|
|
len(diskoffering) > 0,
|
|
"no disk offerings in the deployment")
|
|
|
|
vol = Volume.create(
|
|
domapiclient,
|
|
services=self.testdata["volume"],
|
|
zoneid=self.zone.id,
|
|
account=domuser.name,
|
|
domainid=dom.id,
|
|
diskofferingid=[x for x in diskoffering if not x.iscustomized][0].id
|
|
)
|
|
self.cleanup.append(vol)
|
|
self.assertTrue(
|
|
vol is not None, "volume creation fails in domain %s as user %s" %
|
|
(dom.name, domuser.name))
|
|
|
|
listed_vol = Volume.list(domapiclient, id=vol.id)
|
|
self.assertTrue(
|
|
listed_vol is not None and isinstance(
|
|
listed_vol,
|
|
list),
|
|
"invalid response from listVolumes for volume %s" %
|
|
vol.id)
|
|
self.assertTrue(
|
|
listed_vol[0].id == vol.id,
|
|
"Volume returned by list volumes %s not matching with queried\
|
|
volume %s in domain %s" %
|
|
(listed_vol[0].id,
|
|
vol.id,
|
|
dom.name))
|
|
|
|
|
|
class TestDeployVmWithCustomDisk(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(
|
|
TestDeployVmWithCustomDisk,
|
|
cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"],
|
|
custom=True
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
cls.testdata["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["template"] = template.id
|
|
|
|
# Create VMs, NAT Rules etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
|
|
def setUp(self):
|
|
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("RBD storage type is required for data volumes for LXC")
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestDeployVmWithCustomDisk, cls).tearDownClass()
|
|
|
|
def tearDown(self):
|
|
super(TestDeployVmWithCustomDisk, self).tearDown()
|
|
|
|
@attr(tags=["advanced", "configuration", "advancedns", "simulator",
|
|
"api", "basic", "eip", "sg"])
|
|
def test_deployVmWithCustomDisk(self):
|
|
"""Test custom disk sizes beyond range
|
|
# Steps for validation
|
|
# 1. listConfigurations - custom.diskoffering.size.min
|
|
# and custom.diskoffering.size.max
|
|
# 2. deployVm with custom disk offering size < min
|
|
# 3. deployVm with custom disk offering min< size < max
|
|
# 4. deployVm with custom disk offering size > max
|
|
# Validate the following
|
|
# 2. and 4. of deploy VM should fail.
|
|
# Only case 3. should succeed.
|
|
# cleanup all created data disks from the account
|
|
"""
|
|
|
|
config = Configurations.list(
|
|
self.apiclient,
|
|
name="custom.diskoffering.size.min"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(config, list),
|
|
True,
|
|
"custom.diskoffering.size.min should be present in global config"
|
|
)
|
|
# minimum size of custom disk (in GBs)
|
|
min_size = int(config[0].value)
|
|
self.debug("custom.diskoffering.size.min: %s" % min_size)
|
|
|
|
config = Configurations.list(
|
|
self.apiclient,
|
|
name="custom.diskoffering.size.max"
|
|
)
|
|
self.assertEqual(
|
|
isinstance(config, list),
|
|
True,
|
|
"custom.diskoffering.size.min should be present in global config"
|
|
)
|
|
# maximum size of custom disk (in GBs)
|
|
max_size = int(config[0].value)
|
|
self.debug("custom.diskoffering.size.max: %s" % max_size)
|
|
|
|
self.debug("Creating a volume with size less than min cust disk size")
|
|
self.testdata["custom_volume"]["customdisksize"] = (min_size - 1)
|
|
self.testdata["custom_volume"]["zoneid"] = self.zone.id
|
|
with self.assertRaises(Exception):
|
|
vol = Volume.create_custom_disk(
|
|
self.apiclient,
|
|
self.testdata["custom_volume"],
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
self.cleanup.append(vol)
|
|
self.debug("Create volume failed!")
|
|
|
|
self.debug("Creating a volume with size more than max cust disk size")
|
|
self.testdata["custom_volume"]["customdisksize"] = (max_size + 1)
|
|
with self.assertRaises(Exception):
|
|
vol = Volume.create_custom_disk(
|
|
self.apiclient,
|
|
self.testdata["custom_volume"],
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
self.cleanup.append(vol)
|
|
self.debug("Create volume failed!")
|
|
|
|
self.debug("Creating a volume with size more than min cust disk " +
|
|
"but less than max cust disk size"
|
|
)
|
|
self.testdata["custom_volume"]["customdisksize"] = (min_size + 1)
|
|
vol = Volume.create_custom_disk(
|
|
self.apiclient,
|
|
self.testdata["custom_volume"],
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
diskofferingid=self.disk_offering.id
|
|
)
|
|
self.cleanup.append(vol)
|
|
self.debug("Create volume of cust disk size succeeded")
|
|
|
|
return
|
|
|
|
|
|
class TestMigrateVolume(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls.testClient = super(TestMigrateVolume, cls).getClsTestClient()
|
|
cls.api_client = cls.testClient.getApiClient()
|
|
|
|
cls.testdata = cls.testClient.getParsedTestDataConfig()
|
|
# Get Zone, Domain and templates
|
|
cls.domain = get_domain(cls.api_client)
|
|
cls.zone = get_zone(cls.api_client, cls.testClient.getZoneForTests())
|
|
cls.testdata['mode'] = cls.zone.networktype
|
|
cls._cleanup = []
|
|
cls.unsupportedStorageType = False
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
if cls.hypervisor.lower() == 'lxc':
|
|
if not find_storage_pool_type(cls.api_client, storagetype='rbd'):
|
|
cls.unsupportedStorageType = True
|
|
return
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["disk_offering"]
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
template = get_template(
|
|
cls.api_client,
|
|
cls.zone.id,
|
|
cls.testdata["ostype"]
|
|
)
|
|
cls.testdata["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["zoneid"] = cls.zone.id
|
|
cls.testdata["virtual_machine"]["template"] = template.id
|
|
cls.testdata["virtual_machine"][
|
|
"diskofferingid"] = cls.disk_offering.id
|
|
|
|
# Create VMs, VMs etc
|
|
cls.account = Account.create(
|
|
cls.api_client,
|
|
cls.testdata["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.account)
|
|
cls.small_offering = ServiceOffering.create(
|
|
cls.api_client,
|
|
cls.testdata["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.small_offering)
|
|
cls.virtual_machine = VirtualMachine.create(
|
|
cls.api_client,
|
|
cls.testdata["virtual_machine"],
|
|
accountid=cls.account.name,
|
|
domainid=cls.account.domainid,
|
|
serviceofferingid=cls.small_offering.id,
|
|
mode=cls.testdata["mode"]
|
|
)
|
|
cls._cleanup.append(cls.virtual_machine)
|
|
return
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestMigrateVolume, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
|
|
if self.unsupportedStorageType:
|
|
self.skipTest("RBD storage type is required for data volumes for LXC")
|
|
return
|
|
|
|
def tearDown(self):
|
|
super(TestMigrateVolume, self).tearDown()
|
|
|
|
@attr(tags=["advanced", "sg", "advancedsg"], required_hardware='true')
|
|
def test_01_migrateVolume(self):
|
|
"""
|
|
@Desc:Volume is not retaining same uuid when migrating from one
|
|
storage to another.
|
|
Step1:Create a volume/data disk
|
|
Step2:Verify UUID of the volume
|
|
Step3:Migrate the volume to another primary storage within
|
|
the cluster
|
|
Step4:Migrating volume to new primary storage should succeed
|
|
Step5:volume UUID should not change even after migration
|
|
"""
|
|
if self.hypervisor.lower() in ["kvm"]:
|
|
self.skipTest("KVM cannot migrate an attached volume without the VM")
|
|
|
|
vol = Volume.create(
|
|
self.apiclient,
|
|
self.testdata["volume"],
|
|
diskofferingid=self.disk_offering.id,
|
|
zoneid=self.zone.id,
|
|
account=self.account.name,
|
|
domainid=self.account.domainid,
|
|
)
|
|
self.cleanup.append(vol)
|
|
self.assertIsNotNone(vol, "Failed to create volume")
|
|
vol_res = Volume.list(
|
|
self.apiclient,
|
|
id=vol.id
|
|
)
|
|
self.assertEqual(
|
|
validateList(vol_res)[0],
|
|
PASS,
|
|
"Invalid response returned for list volumes")
|
|
vol_uuid = vol_res[0].id
|
|
self.virtual_machine.attach_volume(
|
|
self.apiclient,
|
|
vol
|
|
)
|
|
self.cleanup.remove(vol)
|
|
|
|
pools = StoragePool.listForMigration(
|
|
self.apiclient,
|
|
id=vol.id
|
|
)
|
|
if not pools:
|
|
self.skipTest(
|
|
"No suitable storage pools found for volume migration.\
|
|
Skipping")
|
|
self.assertEqual(
|
|
validateList(pools)[0],
|
|
PASS,
|
|
"invalid pool response from findStoragePoolsForMigration")
|
|
pool = pools[0]
|
|
self.debug("Migrating Volume-ID: %s to Pool: %s" % (vol.id, pool.id))
|
|
Volume.migrate(
|
|
self.apiclient,
|
|
volumeid=vol.id,
|
|
storageid=pool.id,
|
|
livemigrate='true'
|
|
)
|
|
|
|
migrated_vols = Volume.list(
|
|
self.apiclient,
|
|
virtualmachineid=self.virtual_machine.id,
|
|
listall='true',
|
|
type='DATADISK'
|
|
)
|
|
self.assertEqual(
|
|
validateList(migrated_vols)[0],
|
|
PASS,
|
|
"invalid volumes response after migration")
|
|
migrated_vol_uuid = migrated_vols[0].id
|
|
self.assertEqual(
|
|
vol_uuid,
|
|
migrated_vol_uuid,
|
|
"Volume is not retaining same uuid when migrating from one\
|
|
storage to another"
|
|
)
|
|
self.virtual_machine.detach_volume(
|
|
self.apiclient,
|
|
vol
|
|
)
|
|
self.cleanup.append(vol)
|
|
return
|