mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 08:42:29 +01:00
Switched to the official SolidFire SDK for Python
This commit is contained in:
parent
3f6faeb4c3
commit
d4ae1ab6b6
@ -19,3 +19,6 @@
|
|||||||
cloudmonkey
|
cloudmonkey
|
||||||
|
|
||||||
# Marvin dependencies are installed via its bundle
|
# Marvin dependencies are installed via its bundle
|
||||||
|
|
||||||
|
# Install the SolidFire SDK for Python
|
||||||
|
solidfire-sdk-python
|
||||||
@ -21,6 +21,8 @@ import SignedAPICall
|
|||||||
import time
|
import time
|
||||||
import XenAPI
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
from util import sf_util
|
from util import sf_util
|
||||||
|
|
||||||
# All tests inherit from cloudstackTestCase
|
# All tests inherit from cloudstackTestCase
|
||||||
@ -37,8 +39,6 @@ from marvin.lib.common import get_domain, get_template, get_zone, list_hosts, li
|
|||||||
# utils - utility classes for common cleanup, external library wrappers, etc.
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
from marvin.lib.utils import cleanup_resources
|
from marvin.lib.utils import cleanup_resources
|
||||||
|
|
||||||
from solidfire import solidfire_element_api as sf_api
|
|
||||||
|
|
||||||
# Prerequisites:
|
# Prerequisites:
|
||||||
# Only one zone
|
# Only one zone
|
||||||
# Only one pod
|
# Only one pod
|
||||||
@ -59,12 +59,10 @@ class TestData:
|
|||||||
diskSize = "disksize"
|
diskSize = "disksize"
|
||||||
domainId = "domainId"
|
domainId = "domainId"
|
||||||
hypervisor = "hypervisor"
|
hypervisor = "hypervisor"
|
||||||
login = "login"
|
|
||||||
mvip = "mvip"
|
mvip = "mvip"
|
||||||
name = "name"
|
name = "name"
|
||||||
newHost = "newHost"
|
newHost = "newHost"
|
||||||
newHostDisplayName = "newHostDisplayName"
|
newHostDisplayName = "newHostDisplayName"
|
||||||
osType = "ostype"
|
|
||||||
password = "password"
|
password = "password"
|
||||||
podId = "podid"
|
podId = "podid"
|
||||||
port = "port"
|
port = "port"
|
||||||
@ -89,7 +87,7 @@ class TestData:
|
|||||||
self.testdata = {
|
self.testdata = {
|
||||||
TestData.solidFire: {
|
TestData.solidFire: {
|
||||||
TestData.mvip: "192.168.139.112",
|
TestData.mvip: "192.168.139.112",
|
||||||
TestData.login: "admin",
|
TestData.username: "admin",
|
||||||
TestData.password: "admin",
|
TestData.password: "admin",
|
||||||
TestData.port: 443,
|
TestData.port: 443,
|
||||||
TestData.url: "https://192.168.139.112:443"
|
TestData.url: "https://192.168.139.112:443"
|
||||||
@ -170,7 +168,6 @@ class TestData:
|
|||||||
"diskname": "testvolume2",
|
"diskname": "testvolume2",
|
||||||
},
|
},
|
||||||
TestData.newHostDisplayName: "XenServer-6.5-3",
|
TestData.newHostDisplayName: "XenServer-6.5-3",
|
||||||
TestData.osType: "CentOS 5.6(64-bit) no GUI (XenServer)",
|
|
||||||
TestData.zoneId: 1,
|
TestData.zoneId: 1,
|
||||||
TestData.clusterId: 1,
|
TestData.clusterId: 1,
|
||||||
TestData.domainId: 1,
|
TestData.domainId: 1,
|
||||||
@ -186,7 +183,9 @@ class TestAddRemoveHosts(cloudstackTestCase):
|
|||||||
def setUpClass(cls):
|
def setUpClass(cls):
|
||||||
# Set up API client
|
# Set up API client
|
||||||
testclient = super(TestAddRemoveHosts, cls).getClsTestClient()
|
testclient = super(TestAddRemoveHosts, cls).getClsTestClient()
|
||||||
|
|
||||||
cls.apiClient = testclient.getApiClient()
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
cls.dbConnection = testclient.getDbConnection()
|
cls.dbConnection = testclient.getDbConnection()
|
||||||
|
|
||||||
cls.testdata = TestData().testdata
|
cls.testdata = TestData().testdata
|
||||||
@ -203,12 +202,14 @@ class TestAddRemoveHosts(cloudstackTestCase):
|
|||||||
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
# Set up SolidFire connection
|
# Set up SolidFire connection
|
||||||
cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire])
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
# Get Resources from Cloud Infrastructure
|
# Get Resources from Cloud Infrastructure
|
||||||
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
||||||
cls.cluster = list_clusters(cls.apiClient)[0]
|
cls.cluster = list_clusters(cls.apiClient)[0]
|
||||||
cls.template = get_template(cls.apiClient, cls.zone.id, cls.testdata[TestData.osType])
|
cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
# Create test account
|
# Create test account
|
||||||
@ -249,7 +250,7 @@ class TestAddRemoveHosts(cloudstackTestCase):
|
|||||||
try:
|
try:
|
||||||
cleanup_resources(cls.apiClient, cls._cleanup)
|
cleanup_resources(cls.apiClient, cls._cleanup)
|
||||||
|
|
||||||
sf_util.purge_solidfire_volumes(cls.sf_client)
|
sf_util.purge_solidfire_volumes(cls.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
@ -423,8 +424,8 @@ class TestAddRemoveHosts(cloudstackTestCase):
|
|||||||
|
|
||||||
self._perform_add_remove_host(primary_storage_2.id, sf_iscsi_name)
|
self._perform_add_remove_host(primary_storage_2.id, sf_iscsi_name)
|
||||||
|
|
||||||
def _perform_add_remove_host(self, primary_storage_id, sf_iscsi_name):
|
def _perform_add_remove_host(self, primary_storage_id, sr_name):
|
||||||
xen_sr = self.xen_session.xenapi.SR.get_by_name_label(sf_iscsi_name)[0]
|
xen_sr = self.xen_session.xenapi.SR.get_by_name_label(sr_name)[0]
|
||||||
|
|
||||||
pbds = self.xen_session.xenapi.SR.get_PBDs(xen_sr)
|
pbds = self.xen_session.xenapi.SR.get_PBDs(xen_sr)
|
||||||
|
|
||||||
@ -651,10 +652,10 @@ class TestAddRemoveHosts(cloudstackTestCase):
|
|||||||
return sf_vag_id
|
return sf_vag_id
|
||||||
|
|
||||||
def _get_sf_vag(self, sf_vag_id):
|
def _get_sf_vag(self, sf_vag_id):
|
||||||
return self.sf_client.list_volume_access_groups(sf_vag_id, 1)["volumeAccessGroups"][0]
|
return self.sfe.list_volume_access_groups(sf_vag_id, 1).volume_access_groups[0]
|
||||||
|
|
||||||
def _get_sf_vag_initiators(self, sf_vag):
|
def _get_sf_vag_initiators(self, sf_vag):
|
||||||
return sf_vag["initiators"]
|
return sf_vag.initiators
|
||||||
|
|
||||||
def _verifyVag(self, host_iscsi_iqns, sf_vag_initiators):
|
def _verifyVag(self, host_iscsi_iqns, sf_vag_initiators):
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
|
|||||||
588
test/integration/plugins/solidfire/TestManagedSystemVMs.py
Normal file
588
test/integration/plugins/solidfire/TestManagedSystemVMs.py
Normal file
@ -0,0 +1,588 @@
|
|||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing,
|
||||||
|
# software distributed under the License is distributed on an
|
||||||
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
# KIND, either express or implied. See the License for the
|
||||||
|
# specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import random
|
||||||
|
import SignedAPICall
|
||||||
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
|
from util import sf_util
|
||||||
|
|
||||||
|
from marvin.cloudstackAPI import destroySystemVm
|
||||||
|
|
||||||
|
# All tests inherit from cloudstackTestCase
|
||||||
|
from marvin.cloudstackTestCase import cloudstackTestCase
|
||||||
|
|
||||||
|
from nose.plugins.attrib import attr
|
||||||
|
|
||||||
|
# Import Integration Libraries
|
||||||
|
|
||||||
|
# base - contains all resources as entities and defines create, delete, list operations on them
|
||||||
|
from marvin.lib.base import Account, Router, ServiceOffering, StoragePool, User, VirtualMachine, Zone
|
||||||
|
|
||||||
|
# common - commonly used methods for all tests are listed here
|
||||||
|
from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_ssvms, list_routers
|
||||||
|
|
||||||
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
|
from marvin.lib.utils import cleanup_resources, wait_until
|
||||||
|
|
||||||
|
# Prerequisites:
|
||||||
|
# * Only use one SolidFire cluster for the two primary storages based on the "SolidFire" storage plug-in.
|
||||||
|
# * Do not run other workloads on the SolidFire cluster while running this test as this test checks at a certain
|
||||||
|
# point to make sure no active SolidFire volumes exist.
|
||||||
|
# * Only one zone
|
||||||
|
# * Only one secondary storage VM and one console proxy VM running on NFS (no virtual router or user VMs exist)
|
||||||
|
# * Only one pod
|
||||||
|
# * Only one cluster
|
||||||
|
# * Set storage.cleanup.enabled to true
|
||||||
|
# * Set storage.cleanup.interval to 150
|
||||||
|
# * Set storage.cleanup.delay to 60
|
||||||
|
|
||||||
|
|
||||||
|
class TestData():
|
||||||
|
account = "account"
|
||||||
|
capacityBytes = "capacitybytes"
|
||||||
|
capacityIops = "capacityiops"
|
||||||
|
clusterId = "clusterid"
|
||||||
|
computeOffering = "computeoffering"
|
||||||
|
diskOffering = "diskoffering"
|
||||||
|
domainId = "domainid"
|
||||||
|
email = "email"
|
||||||
|
firstname = "firstname"
|
||||||
|
hypervisor = "hypervisor"
|
||||||
|
lastname = "lastname"
|
||||||
|
max_iops = "maxiops"
|
||||||
|
min_iops = "miniops"
|
||||||
|
mvip = "mvip"
|
||||||
|
name = "name"
|
||||||
|
password = "password"
|
||||||
|
port = "port"
|
||||||
|
primaryStorage = "primarystorage"
|
||||||
|
provider = "provider"
|
||||||
|
scope = "scope"
|
||||||
|
solidFire = "solidfire"
|
||||||
|
storageTag = "SolidFire_SAN_1"
|
||||||
|
systemOffering = "systemoffering"
|
||||||
|
systemOfferingFailure = "systemofferingFailure"
|
||||||
|
tags = "tags"
|
||||||
|
url = "url"
|
||||||
|
user = "user"
|
||||||
|
username = "username"
|
||||||
|
virtualMachine = "virtualmachine"
|
||||||
|
xenServer = "xenserver"
|
||||||
|
zoneId = "zoneid"
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
self.testdata = {
|
||||||
|
TestData.solidFire: {
|
||||||
|
TestData.mvip: "192.168.139.112",
|
||||||
|
TestData.username: "admin",
|
||||||
|
TestData.password: "admin",
|
||||||
|
TestData.port: 443,
|
||||||
|
TestData.url: "https://192.168.139.112:443"
|
||||||
|
},
|
||||||
|
TestData.xenServer: {
|
||||||
|
TestData.username: "root",
|
||||||
|
TestData.password: "solidfire"
|
||||||
|
},
|
||||||
|
TestData.account: {
|
||||||
|
TestData.email: "test@test.com",
|
||||||
|
TestData.firstname: "John",
|
||||||
|
TestData.lastname: "Doe",
|
||||||
|
TestData.username: "test",
|
||||||
|
TestData.password: "test"
|
||||||
|
},
|
||||||
|
TestData.user: {
|
||||||
|
TestData.email: "user@test.com",
|
||||||
|
TestData.firstname: "Jane",
|
||||||
|
TestData.lastname: "Doe",
|
||||||
|
TestData.username: "testuser",
|
||||||
|
TestData.password: "password"
|
||||||
|
},
|
||||||
|
TestData.primaryStorage: {
|
||||||
|
TestData.name: TestData.get_name_for_solidfire_storage(),
|
||||||
|
TestData.scope: "ZONE",
|
||||||
|
TestData.url: "MVIP=192.168.139.112;SVIP=10.10.8.112;" +
|
||||||
|
"clusterAdminUsername=admin;clusterAdminPassword=admin;" +
|
||||||
|
"clusterDefaultMinIops=10000;clusterDefaultMaxIops=15000;" +
|
||||||
|
"clusterDefaultBurstIopsPercentOfMaxIops=1.5;",
|
||||||
|
TestData.provider: "SolidFire",
|
||||||
|
TestData.tags: TestData.storageTag,
|
||||||
|
TestData.capacityIops: 4500000,
|
||||||
|
TestData.capacityBytes: 2251799813685248,
|
||||||
|
TestData.hypervisor: "Any",
|
||||||
|
TestData.zoneId: 1
|
||||||
|
},
|
||||||
|
TestData.virtualMachine: {
|
||||||
|
TestData.name: "TestVM",
|
||||||
|
"displayname": "Test VM"
|
||||||
|
},
|
||||||
|
TestData.computeOffering: {
|
||||||
|
TestData.name: "SF_CO_1",
|
||||||
|
"displaytext": "SF_CO_1 (Min IOPS = 10,000; Max IOPS = 15,000)",
|
||||||
|
"cpunumber": 1,
|
||||||
|
"cpuspeed": 100,
|
||||||
|
"memory": 128,
|
||||||
|
"storagetype": "shared",
|
||||||
|
"customizediops": False,
|
||||||
|
TestData.min_iops: 10000,
|
||||||
|
TestData.max_iops: 15000,
|
||||||
|
"hypervisorsnapshotreserve": 200,
|
||||||
|
TestData.tags: TestData.storageTag
|
||||||
|
},
|
||||||
|
TestData.systemOffering: {
|
||||||
|
TestData.name: "SF_SO_1",
|
||||||
|
"displaytext": "Managed SO (Min IOPS = 4,000; Max IOPS = 8,000)",
|
||||||
|
"cpunumber": 1,
|
||||||
|
"cpuspeed": 100,
|
||||||
|
"memory": 128,
|
||||||
|
"storagetype": "shared",
|
||||||
|
TestData.min_iops: 4000,
|
||||||
|
TestData.max_iops: 8000,
|
||||||
|
TestData.tags: TestData.storageTag,
|
||||||
|
"issystem": True
|
||||||
|
},
|
||||||
|
TestData.systemOfferingFailure: {
|
||||||
|
TestData.name: "SF_SO_2",
|
||||||
|
"displaytext": "Managed SO (Customized IOPS)",
|
||||||
|
"cpunumber": 1,
|
||||||
|
"cpuspeed": 100,
|
||||||
|
"memory": 128,
|
||||||
|
"storagetype": "shared",
|
||||||
|
"customizediops": True,
|
||||||
|
TestData.tags: TestData.storageTag,
|
||||||
|
"issystem": True
|
||||||
|
},
|
||||||
|
TestData.zoneId: 1,
|
||||||
|
TestData.clusterId: 1,
|
||||||
|
TestData.domainId: 1,
|
||||||
|
TestData.url: "192.168.129.50"
|
||||||
|
}
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def get_name_for_solidfire_storage():
|
||||||
|
return "SolidFire-%d" % random.randint(0, 100)
|
||||||
|
|
||||||
|
|
||||||
|
class TestManagedSystemVMs(cloudstackTestCase):
|
||||||
|
_unique_name_suffix = "-Temp"
|
||||||
|
|
||||||
|
_secondary_storage_unique_name = "Cloud.com-SecondaryStorage"
|
||||||
|
_secondary_storage_temp_unique_name = _secondary_storage_unique_name + _unique_name_suffix
|
||||||
|
|
||||||
|
_console_proxy_unique_name = "Cloud.com-ConsoleProxy"
|
||||||
|
_console_proxy_temp_unique_name = _console_proxy_unique_name + _unique_name_suffix
|
||||||
|
|
||||||
|
_virtual_router_unique_name = "Cloud.com-SoftwareRouter"
|
||||||
|
_virtual_router_temp_unique_name = _virtual_router_unique_name + _unique_name_suffix
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def setUpClass(cls):
|
||||||
|
# Set up API client
|
||||||
|
testclient = super(TestManagedSystemVMs, cls).getClsTestClient()
|
||||||
|
|
||||||
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
|
cls.dbConnection = testclient.getDbConnection()
|
||||||
|
|
||||||
|
cls.testdata = TestData().testdata
|
||||||
|
|
||||||
|
# Set up xenAPI connection
|
||||||
|
host_ip = "https://" + \
|
||||||
|
list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId], name="XenServer-6.5-1")[0].ipaddress
|
||||||
|
|
||||||
|
# Set up XenAPI connection
|
||||||
|
cls.xen_session = XenAPI.Session(host_ip)
|
||||||
|
|
||||||
|
xenserver = cls.testdata[TestData.xenServer]
|
||||||
|
|
||||||
|
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
|
# Set up SolidFire connection
|
||||||
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
|
# Get Resources from Cloud Infrastructure
|
||||||
|
cls.zone = Zone(get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId]).__dict__)
|
||||||
|
cls.cluster = list_clusters(cls.apiClient)[0]
|
||||||
|
cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
|
# Create test account
|
||||||
|
cls.account = Account.create(
|
||||||
|
cls.apiClient,
|
||||||
|
cls.testdata["account"],
|
||||||
|
admin=1
|
||||||
|
)
|
||||||
|
|
||||||
|
# Set up connection to make customized API calls
|
||||||
|
cls.user = User.create(
|
||||||
|
cls.apiClient,
|
||||||
|
cls.testdata["user"],
|
||||||
|
account=cls.account.name,
|
||||||
|
domainid=cls.domain.id
|
||||||
|
)
|
||||||
|
|
||||||
|
url = cls.testdata[TestData.url]
|
||||||
|
|
||||||
|
api_url = "http://" + url + ":8080/client/api"
|
||||||
|
userkeys = User.registerUserKeys(cls.apiClient, cls.user.id)
|
||||||
|
|
||||||
|
cls.cs_api = SignedAPICall.CloudStack(api_url, userkeys.apikey, userkeys.secretkey)
|
||||||
|
|
||||||
|
cls.compute_offering = ServiceOffering.create(
|
||||||
|
cls.apiClient,
|
||||||
|
cls.testdata[TestData.computeOffering]
|
||||||
|
)
|
||||||
|
|
||||||
|
systemoffering = cls.testdata[TestData.systemOffering]
|
||||||
|
|
||||||
|
systemoffering[TestData.name] = "Managed SSVM"
|
||||||
|
systemoffering['systemvmtype'] = "secondarystoragevm"
|
||||||
|
|
||||||
|
cls.secondary_storage_offering = ServiceOffering.create(
|
||||||
|
cls.apiClient,
|
||||||
|
systemoffering
|
||||||
|
)
|
||||||
|
|
||||||
|
systemoffering[TestData.name] = "Managed CPVM"
|
||||||
|
systemoffering['systemvmtype'] = "consoleproxy"
|
||||||
|
|
||||||
|
cls.console_proxy_offering = ServiceOffering.create(
|
||||||
|
cls.apiClient,
|
||||||
|
systemoffering
|
||||||
|
)
|
||||||
|
|
||||||
|
systemoffering[TestData.name] = "Managed VR"
|
||||||
|
systemoffering['systemvmtype'] = "domainrouter"
|
||||||
|
|
||||||
|
cls.virtual_router_offering = ServiceOffering.create(
|
||||||
|
cls.apiClient,
|
||||||
|
systemoffering
|
||||||
|
)
|
||||||
|
|
||||||
|
# Resources that are to be destroyed
|
||||||
|
cls._cleanup = [
|
||||||
|
cls.secondary_storage_offering,
|
||||||
|
cls.console_proxy_offering,
|
||||||
|
cls.virtual_router_offering,
|
||||||
|
cls.compute_offering,
|
||||||
|
cls.user,
|
||||||
|
cls.account
|
||||||
|
]
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def tearDownClass(cls):
|
||||||
|
try:
|
||||||
|
cleanup_resources(cls.apiClient, cls._cleanup)
|
||||||
|
except Exception as e:
|
||||||
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
self.cleanup = []
|
||||||
|
|
||||||
|
def tearDown(self):
|
||||||
|
try:
|
||||||
|
cleanup_resources(self.apiClient, self.cleanup)
|
||||||
|
|
||||||
|
sf_util.purge_solidfire_volumes(self.sfe)
|
||||||
|
except Exception as e:
|
||||||
|
logging.debug("Exception in tearDownClass(self): %s" % e)
|
||||||
|
|
||||||
|
@attr(hypervisor='XenServer')
|
||||||
|
def test_01_create_system_vms_on_managed_storage(self):
|
||||||
|
self._disable_zone_and_delete_system_vms(None, False)
|
||||||
|
|
||||||
|
primary_storage = self.testdata[TestData.primaryStorage]
|
||||||
|
|
||||||
|
primary_storage_1 = StoragePool.create(
|
||||||
|
self.apiClient,
|
||||||
|
primary_storage
|
||||||
|
)
|
||||||
|
|
||||||
|
self._prepare_to_use_managed_storage_for_system_vms()
|
||||||
|
|
||||||
|
enabled = "Enabled"
|
||||||
|
|
||||||
|
self.zone.update(self.apiClient, id=self.zone.id, allocationstate=enabled)
|
||||||
|
|
||||||
|
system_vms = self._wait_for_and_get_running_system_vms(2)
|
||||||
|
|
||||||
|
virtual_machine = VirtualMachine.create(
|
||||||
|
self.apiClient,
|
||||||
|
self.testdata[TestData.virtualMachine],
|
||||||
|
accountid=self.account.name,
|
||||||
|
zoneid=self.zone.id,
|
||||||
|
serviceofferingid=self.compute_offering.id,
|
||||||
|
templateid=self.template.id,
|
||||||
|
domainid=self.domain.id,
|
||||||
|
startvm=True
|
||||||
|
)
|
||||||
|
|
||||||
|
# This virtual machine was only created and started so that the virtual router would be created and started.
|
||||||
|
# Just delete this virtual machine once it has been created and started.
|
||||||
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
|
virtual_router = list_routers(self.apiClient, listall=True, state="Running")[0]
|
||||||
|
|
||||||
|
system_vms.append(virtual_router)
|
||||||
|
|
||||||
|
self._check_system_vms(system_vms, primary_storage_1.id)
|
||||||
|
|
||||||
|
primary_storage[TestData.name] = TestData.get_name_for_solidfire_storage()
|
||||||
|
|
||||||
|
primary_storage_2 = StoragePool.create(
|
||||||
|
self.apiClient,
|
||||||
|
primary_storage
|
||||||
|
)
|
||||||
|
|
||||||
|
StoragePool.enableMaintenance(self.apiClient, primary_storage_1.id)
|
||||||
|
|
||||||
|
self._wait_for_storage_cleanup_thread(system_vms)
|
||||||
|
|
||||||
|
sf_util.purge_solidfire_volumes(self.sfe)
|
||||||
|
|
||||||
|
system_vms = self._wait_for_and_get_running_system_vms(2)
|
||||||
|
|
||||||
|
virtual_router = list_routers(self.apiClient, listall=True, state="Running")[0]
|
||||||
|
|
||||||
|
system_vms.append(virtual_router)
|
||||||
|
|
||||||
|
self._check_system_vms(system_vms, primary_storage_2.id)
|
||||||
|
|
||||||
|
StoragePool.cancelMaintenance(self.apiClient, primary_storage_1.id)
|
||||||
|
|
||||||
|
primary_storage_1.delete(self.apiClient)
|
||||||
|
|
||||||
|
self._disable_zone_and_delete_system_vms(virtual_router)
|
||||||
|
|
||||||
|
self._wait_for_storage_cleanup_thread(system_vms)
|
||||||
|
|
||||||
|
sf_util.purge_solidfire_volumes(self.sfe)
|
||||||
|
|
||||||
|
primary_storage_2.delete(self.apiClient)
|
||||||
|
|
||||||
|
self._verify_no_active_solidfire_volumes()
|
||||||
|
|
||||||
|
self._prepare_to_stop_using_managed_storage_for_system_vms()
|
||||||
|
|
||||||
|
self.zone.update(self.apiClient, id=self.zone.id, allocationstate=enabled)
|
||||||
|
|
||||||
|
self._wait_for_and_get_running_system_vms(2)
|
||||||
|
|
||||||
|
@attr(hypervisor='XenServer')
|
||||||
|
def test_02_failure_to_create_service_offering_with_customized_iops(self):
|
||||||
|
try:
|
||||||
|
ServiceOffering.create(
|
||||||
|
self.apiClient,
|
||||||
|
self.testdata[TestData.systemOfferingFailure]
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assert_(True, "The service offering was created, but should not have been.")
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
|
||||||
|
def _prepare_to_use_managed_storage_for_system_vms(self):
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._secondary_storage_unique_name, TestManagedSystemVMs._secondary_storage_temp_unique_name)
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._console_proxy_unique_name, TestManagedSystemVMs._console_proxy_temp_unique_name)
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._virtual_router_unique_name, TestManagedSystemVMs._virtual_router_temp_unique_name)
|
||||||
|
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.secondary_storage_offering.id, TestManagedSystemVMs._secondary_storage_unique_name)
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.console_proxy_offering.id, TestManagedSystemVMs._console_proxy_unique_name)
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.virtual_router_offering.id, TestManagedSystemVMs._virtual_router_unique_name)
|
||||||
|
|
||||||
|
def _prepare_to_stop_using_managed_storage_for_system_vms(self):
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.secondary_storage_offering.id, None)
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.console_proxy_offering.id, None)
|
||||||
|
self._update_system_vm_unique_name_based_on_uuid(self.virtual_router_offering.id, None)
|
||||||
|
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._secondary_storage_temp_unique_name, TestManagedSystemVMs._secondary_storage_unique_name)
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._console_proxy_temp_unique_name, TestManagedSystemVMs._console_proxy_unique_name)
|
||||||
|
self._update_system_vm_unique_name(TestManagedSystemVMs._virtual_router_temp_unique_name, TestManagedSystemVMs._virtual_router_unique_name)
|
||||||
|
|
||||||
|
def _wait_for_storage_cleanup_thread(self, system_vms):
|
||||||
|
retry_interval = 60
|
||||||
|
num_tries = 10
|
||||||
|
|
||||||
|
wait_result, return_val = wait_until(retry_interval, num_tries, self._check_resource_state, system_vms)
|
||||||
|
|
||||||
|
if not wait_result:
|
||||||
|
raise Exception(return_val)
|
||||||
|
|
||||||
|
def _check_resource_state(self, system_vms):
|
||||||
|
try:
|
||||||
|
self._verify_system_vms_deleted(system_vms)
|
||||||
|
|
||||||
|
return True, None
|
||||||
|
except:
|
||||||
|
return False, "The system is not in the necessary state."
|
||||||
|
|
||||||
|
def _verify_system_vms_deleted(self, system_vms):
|
||||||
|
for system_vm in system_vms:
|
||||||
|
cs_root_volume = self._get_root_volume_for_system_vm(system_vm.id, 'Expunged')
|
||||||
|
|
||||||
|
self._verify_managed_system_vm_deleted(cs_root_volume.name)
|
||||||
|
|
||||||
|
def _disable_zone_and_delete_system_vms(self, virtual_router, verify_managed_system_vm_deleted=True):
|
||||||
|
self.zone.update(self.apiClient, id=self.zone.id, allocationstate="Disabled")
|
||||||
|
|
||||||
|
if virtual_router is not None:
|
||||||
|
Router.destroy(self.apiClient, virtual_router.id)
|
||||||
|
|
||||||
|
if verify_managed_system_vm_deleted:
|
||||||
|
cs_root_volume = self._get_root_volume_for_system_vm(virtual_router.id, 'Expunged')
|
||||||
|
|
||||||
|
self._verify_managed_system_vm_deleted(cs_root_volume.name)
|
||||||
|
|
||||||
|
# list_ssvms lists the secondary storage VM and the console proxy VM
|
||||||
|
system_vms = list_ssvms(self.apiClient)
|
||||||
|
|
||||||
|
for system_vm in system_vms:
|
||||||
|
destroy_ssvm_cmd = destroySystemVm.destroySystemVmCmd()
|
||||||
|
|
||||||
|
destroy_ssvm_cmd.id = system_vm.id
|
||||||
|
|
||||||
|
self.apiClient.destroySystemVm(destroy_ssvm_cmd)
|
||||||
|
|
||||||
|
if verify_managed_system_vm_deleted:
|
||||||
|
cs_root_volume = self._get_root_volume_for_system_vm(system_vm.id, 'Expunged')
|
||||||
|
|
||||||
|
self._verify_managed_system_vm_deleted(cs_root_volume.name)
|
||||||
|
|
||||||
|
def _verify_managed_system_vm_deleted(self, cs_root_volume_name):
|
||||||
|
sf_not_active_volumes = sf_util.get_not_active_sf_volumes(self.sfe)
|
||||||
|
|
||||||
|
sf_root_volume = sf_util.check_and_get_sf_volume(sf_not_active_volumes, cs_root_volume_name, self)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
len(sf_root_volume.volume_access_groups),
|
||||||
|
0,
|
||||||
|
"The volume should not be in a volume access group."
|
||||||
|
)
|
||||||
|
|
||||||
|
sr_name = sf_util.format_iqn(sf_root_volume.iqn)
|
||||||
|
|
||||||
|
sf_util.check_xen_sr(sr_name, self.xen_session, self, False)
|
||||||
|
|
||||||
|
def _wait_for_and_get_running_system_vms(self, expected_number_of_system_vms):
|
||||||
|
retry_interval = 60
|
||||||
|
num_tries = 10
|
||||||
|
|
||||||
|
wait_result, return_val = wait_until(retry_interval, num_tries, self._check_number_of_running_system_vms, expected_number_of_system_vms)
|
||||||
|
|
||||||
|
if not wait_result:
|
||||||
|
raise Exception(return_val)
|
||||||
|
|
||||||
|
return return_val
|
||||||
|
|
||||||
|
def _check_number_of_running_system_vms(self, expected_number_of_system_vms):
|
||||||
|
# list_ssvms lists the secondary storage VM and the console proxy VM
|
||||||
|
system_vms = list_ssvms(self.apiClient, state="Running")
|
||||||
|
|
||||||
|
if system_vms is not None and len(system_vms) == expected_number_of_system_vms:
|
||||||
|
return True, system_vms
|
||||||
|
|
||||||
|
return False, "Timed out waiting for running system VMs"
|
||||||
|
|
||||||
|
def _verify_no_active_solidfire_volumes(self):
|
||||||
|
sf_active_volumes = sf_util.get_active_sf_volumes(self.sfe)
|
||||||
|
|
||||||
|
sf_util.check_list(sf_active_volumes, 0, self, "There should be no active SolidFire volumes in the cluster.")
|
||||||
|
|
||||||
|
def _check_system_vms(self, system_vms, primary_storage_id):
|
||||||
|
sf_active_volumes = sf_util.get_active_sf_volumes(self.sfe)
|
||||||
|
|
||||||
|
sf_vag_id = sf_util.get_vag_id(self.cs_api, self.cluster.id, primary_storage_id, self)
|
||||||
|
|
||||||
|
for system_vm in system_vms:
|
||||||
|
cs_root_volume = self._get_root_volume_for_system_vm(system_vm.id, 'Ready')
|
||||||
|
sf_root_volume = sf_util.check_and_get_sf_volume(sf_active_volumes, cs_root_volume.name, self)
|
||||||
|
|
||||||
|
sf_volume_size = sf_util.get_volume_size_with_hsr(self.cs_api, cs_root_volume, self)
|
||||||
|
|
||||||
|
sf_util.check_size_and_iops(sf_root_volume, cs_root_volume, sf_volume_size, self)
|
||||||
|
|
||||||
|
self._check_iops_against_iops_of_system_offering(cs_root_volume, self.testdata[TestData.systemOffering])
|
||||||
|
|
||||||
|
sf_util.check_vag(sf_root_volume, sf_vag_id, self)
|
||||||
|
|
||||||
|
sr_name = sf_util.format_iqn(sf_root_volume.iqn)
|
||||||
|
|
||||||
|
sf_util.check_xen_sr(sr_name, self.xen_session, self)
|
||||||
|
|
||||||
|
def _check_iops_against_iops_of_system_offering(self, cs_volume, system_offering):
|
||||||
|
self.assertEqual(
|
||||||
|
system_offering[TestData.min_iops],
|
||||||
|
cs_volume.miniops,
|
||||||
|
"Check QoS - Min IOPS: of " + cs_volume.name + " should be " + str(system_offering[TestData.min_iops])
|
||||||
|
)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
system_offering[TestData.max_iops],
|
||||||
|
cs_volume.maxiops,
|
||||||
|
"Check QoS - Min IOPS: of " + cs_volume.name + " should be " + str(system_offering[TestData.max_iops])
|
||||||
|
)
|
||||||
|
|
||||||
|
def _get_root_volume_for_system_vm(self, system_vm_id, state):
|
||||||
|
sql_query = "Select id From vm_instance Where uuid = '" + system_vm_id + "'"
|
||||||
|
|
||||||
|
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
||||||
|
sql_result = self.dbConnection.execute(sql_query)
|
||||||
|
|
||||||
|
instance_id = sql_result[0][0]
|
||||||
|
|
||||||
|
sql_query = "Select uuid, name, min_iops, max_iops From volumes Where instance_id = " + str(instance_id) + \
|
||||||
|
" and state = '" + state + "' Order by removed desc"
|
||||||
|
|
||||||
|
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
||||||
|
sql_result = self.dbConnection.execute(sql_query)
|
||||||
|
|
||||||
|
uuid = sql_result[0][0]
|
||||||
|
name = sql_result[0][1]
|
||||||
|
min_iops = sql_result[0][2]
|
||||||
|
max_iops = sql_result[0][3]
|
||||||
|
|
||||||
|
class CloudStackVolume(object):
|
||||||
|
pass
|
||||||
|
|
||||||
|
cs_volume = CloudStackVolume()
|
||||||
|
|
||||||
|
cs_volume.id = uuid
|
||||||
|
cs_volume.name = name
|
||||||
|
cs_volume.miniops = min_iops
|
||||||
|
cs_volume.maxiops = max_iops
|
||||||
|
|
||||||
|
return cs_volume
|
||||||
|
|
||||||
|
def _update_system_vm_unique_name(self, unique_name, new_unique_name):
|
||||||
|
sql_query = "Update disk_offering set unique_name = '" + new_unique_name + "' Where unique_name = '" + unique_name + "'"
|
||||||
|
|
||||||
|
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
||||||
|
self.dbConnection.execute(sql_query)
|
||||||
|
|
||||||
|
def _update_system_vm_unique_name_based_on_uuid(self, uuid, new_unique_name):
|
||||||
|
if (new_unique_name is None):
|
||||||
|
sql_query = "Update disk_offering set unique_name = NULL Where uuid = '" + uuid + "'"
|
||||||
|
else:
|
||||||
|
sql_query = "Update disk_offering set unique_name = '" + new_unique_name + "' Where uuid = '" + uuid + "'"
|
||||||
|
|
||||||
|
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
||||||
|
self.dbConnection.execute(sql_query)
|
||||||
|
|
||||||
@ -21,6 +21,8 @@ import SignedAPICall
|
|||||||
import time
|
import time
|
||||||
import XenAPI
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
from util import sf_util
|
from util import sf_util
|
||||||
|
|
||||||
# All tests inherit from cloudstackTestCase
|
# All tests inherit from cloudstackTestCase
|
||||||
@ -39,8 +41,6 @@ from marvin.lib.common import get_domain, get_template, get_zone, list_clusters,
|
|||||||
# utils - utility classes for common cleanup, external library wrappers, etc.
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
from marvin.lib.utils import cleanup_resources, wait_until
|
from marvin.lib.utils import cleanup_resources, wait_until
|
||||||
|
|
||||||
from solidfire import solidfire_element_api as sf_api
|
|
||||||
|
|
||||||
# Prerequisites:
|
# Prerequisites:
|
||||||
# Only one zone
|
# Only one zone
|
||||||
# Only one pod
|
# Only one pod
|
||||||
@ -57,7 +57,6 @@ class TestData():
|
|||||||
diskOffering = "diskoffering"
|
diskOffering = "diskoffering"
|
||||||
domainId = "domainId"
|
domainId = "domainId"
|
||||||
hypervisor = "hypervisor"
|
hypervisor = "hypervisor"
|
||||||
login = "login"
|
|
||||||
mvip = "mvip"
|
mvip = "mvip"
|
||||||
password = "password"
|
password = "password"
|
||||||
port = "port"
|
port = "port"
|
||||||
@ -67,7 +66,6 @@ class TestData():
|
|||||||
solidFire = "solidfire"
|
solidFire = "solidfire"
|
||||||
storageTag = "SolidFire_SAN_1"
|
storageTag = "SolidFire_SAN_1"
|
||||||
tags = "tags"
|
tags = "tags"
|
||||||
templateName = "templatename"
|
|
||||||
url = "url"
|
url = "url"
|
||||||
user = "user"
|
user = "user"
|
||||||
username = "username"
|
username = "username"
|
||||||
@ -81,7 +79,7 @@ class TestData():
|
|||||||
self.testdata = {
|
self.testdata = {
|
||||||
TestData.solidFire: {
|
TestData.solidFire: {
|
||||||
TestData.mvip: "192.168.139.112",
|
TestData.mvip: "192.168.139.112",
|
||||||
TestData.login: "admin",
|
TestData.username: "admin",
|
||||||
TestData.password: "admin",
|
TestData.password: "admin",
|
||||||
TestData.port: 443,
|
TestData.port: 443,
|
||||||
TestData.url: "https://192.168.139.112:443"
|
TestData.url: "https://192.168.139.112:443"
|
||||||
@ -223,7 +221,6 @@ class TestData():
|
|||||||
TestData.volume_2: {
|
TestData.volume_2: {
|
||||||
TestData.diskName: "test-volume-2",
|
TestData.diskName: "test-volume-2",
|
||||||
},
|
},
|
||||||
TestData.templateName: "CentOS 5.6(64-bit) no GUI (XenServer)",
|
|
||||||
TestData.zoneId: 1,
|
TestData.zoneId: 1,
|
||||||
TestData.clusterId: 1,
|
TestData.clusterId: 1,
|
||||||
TestData.domainId: 1,
|
TestData.domainId: 1,
|
||||||
@ -252,7 +249,9 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
def setUpClass(cls):
|
def setUpClass(cls):
|
||||||
# Set up API client
|
# Set up API client
|
||||||
testclient = super(TestSnapshots, cls).getClsTestClient()
|
testclient = super(TestSnapshots, cls).getClsTestClient()
|
||||||
|
|
||||||
cls.apiClient = testclient.getApiClient()
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
cls.dbConnection = testclient.getDbConnection()
|
cls.dbConnection = testclient.getDbConnection()
|
||||||
|
|
||||||
cls.testdata = TestData().testdata
|
cls.testdata = TestData().testdata
|
||||||
@ -269,12 +268,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
# Set up SolidFire connection
|
# Set up SolidFire connection
|
||||||
cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire])
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
# Get Resources from Cloud Infrastructure
|
# Get Resources from Cloud Infrastructure
|
||||||
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
||||||
cls.cluster = list_clusters(cls.apiClient)[0]
|
cls.cluster = list_clusters(cls.apiClient)[0]
|
||||||
cls.template = get_template(cls.apiClient, cls.zone.id, template_name=cls.testdata[TestData.templateName])
|
cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
# Create test account
|
# Create test account
|
||||||
@ -338,7 +339,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
cls.primary_storage.delete(cls.apiClient)
|
cls.primary_storage.delete(cls.apiClient)
|
||||||
|
|
||||||
sf_util.purge_solidfire_volumes(cls.sf_client)
|
sf_util.purge_solidfire_volumes(cls.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
@ -377,14 +378,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -409,7 +410,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -418,7 +419,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_1)
|
self._delete_and_test_snapshot(vol_snap_1)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -427,7 +428,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_2)
|
self._delete_and_test_snapshot(vol_snap_2)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -454,14 +455,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_1_root_volume_name = vm_1_root_volume.name
|
vm_1_root_volume_name = vm_1_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -502,14 +503,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_2_root_volume_name = vm_2_root_volume.name
|
vm_2_root_volume_name = vm_2_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots_2 = self.sf_client.list_snapshots(volume_id=sf_volume_2['volumeID'])
|
sf_snapshots_2 = self.sfe.list_snapshots(volume_id=sf_volume_2.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -522,7 +523,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -530,7 +531,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
||||||
|
|
||||||
sf_util.check_list(sf_volume_3['volumeAccessGroups'], 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
sf_util.check_list(sf_volume_3.volume_access_groups, 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
||||||
|
|
||||||
volume_created_from_snapshot = virtual_machine.attach_volume(
|
volume_created_from_snapshot = virtual_machine.attach_volume(
|
||||||
self.apiClient,
|
self.apiClient,
|
||||||
@ -542,7 +543,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -562,7 +563,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_1)
|
self._delete_and_test_snapshot(vol_snap_1)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -572,7 +573,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine_2.delete(self.apiClient, True)
|
virtual_machine_2.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -592,7 +593,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -625,21 +626,21 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
primary_storage_db_id = self._get_cs_storage_pool_db_id(self.primary_storage)
|
primary_storage_db_id = self._get_cs_storage_pool_db_id(self.primary_storage)
|
||||||
|
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
sf_volume_size = sf_volume['totalSize']
|
sf_volume_size = sf_volume.total_size
|
||||||
|
|
||||||
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
||||||
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
@ -665,7 +666,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -696,19 +697,19 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_1_root_volume_name = vm_1_root_volume.name
|
vm_1_root_volume_name = vm_1_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
sf_volume_size = sf_volume['totalSize']
|
sf_volume_size = sf_volume.total_size
|
||||||
|
|
||||||
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
||||||
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
@ -750,19 +751,19 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_2_root_volume_name = vm_2_root_volume.name
|
vm_2_root_volume_name = vm_2_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots_2 = self.sf_client.list_snapshots(volume_id=sf_volume_2['volumeID'])
|
sf_snapshots_2 = self.sfe.list_snapshots(volume_id=sf_volume_2.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_id_2 = sf_volume_2['volumeID']
|
sf_volume_id_2 = sf_volume_2.volume_id
|
||||||
sf_volume_size_2 = sf_volume_2['totalSize']
|
sf_volume_size_2 = sf_volume_2.total_size
|
||||||
|
|
||||||
vol_snap_a = self._create_and_test_snapshot_2(vm_2_root_volume.id, sf_volume_id_2, sf_volume_id + 5, primary_storage_db_id, sf_volume_size_2,
|
vol_snap_a = self._create_and_test_snapshot_2(vm_2_root_volume.id, sf_volume_id_2, sf_volume_id + 5, primary_storage_db_id, sf_volume_size_2,
|
||||||
sf_account_id, 6, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
sf_account_id, 6, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
||||||
@ -774,7 +775,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 7, self, TestSnapshots._should_be_seven_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 7, self, TestSnapshots._should_be_seven_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -782,7 +783,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
||||||
|
|
||||||
sf_util.check_list(sf_volume_3['volumeAccessGroups'], 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
sf_util.check_list(sf_volume_3.volume_access_groups, 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
||||||
|
|
||||||
volume_created_from_snapshot = virtual_machine.attach_volume(
|
volume_created_from_snapshot = virtual_machine.attach_volume(
|
||||||
self.apiClient,
|
self.apiClient,
|
||||||
@ -794,7 +795,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -815,7 +816,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine_2.delete(self.apiClient, True)
|
virtual_machine_2.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -833,7 +834,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -860,7 +861,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_1_root_volume_name = vm_1_root_volume.name
|
vm_1_root_volume_name = vm_1_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -870,14 +871,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
sf_volume_size = sf_volume['totalSize']
|
sf_volume_size = sf_volume.total_size
|
||||||
|
|
||||||
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
||||||
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -888,21 +889,21 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
||||||
|
|
||||||
sf_util.check_list(sf_volume_2['volumeAccessGroups'], 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
sf_util.check_list(sf_volume_2.volume_access_groups, 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
||||||
|
|
||||||
volume_created_from_snapshot = virtual_machine.attach_volume(
|
volume_created_from_snapshot = virtual_machine.attach_volume(
|
||||||
self.apiClient,
|
self.apiClient,
|
||||||
volume_created_from_snapshot
|
volume_created_from_snapshot
|
||||||
)
|
)
|
||||||
|
|
||||||
sf_volume_id_2 = sf_volume_2['volumeID']
|
sf_volume_id_2 = sf_volume_2.volume_id
|
||||||
sf_volume_size_2 = sf_volume_2['totalSize']
|
sf_volume_size_2 = sf_volume_2.total_size
|
||||||
|
|
||||||
vol_snap_a = self._create_and_test_snapshot_2(volume_created_from_snapshot.id, sf_volume_id_2, sf_volume_id + 3, primary_storage_db_id, sf_volume_size_2,
|
vol_snap_a = self._create_and_test_snapshot_2(volume_created_from_snapshot.id, sf_volume_id_2, sf_volume_id + 3, primary_storage_db_id, sf_volume_size_2,
|
||||||
sf_account_id, 4, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
sf_account_id, 4, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
||||||
@ -914,7 +915,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -935,7 +936,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -981,21 +982,21 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
primary_storage_db_id = self._get_cs_storage_pool_db_id(self.primary_storage)
|
primary_storage_db_id = self._get_cs_storage_pool_db_id(self.primary_storage)
|
||||||
|
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
sf_volume_size = sf_volume['totalSize']
|
sf_volume_size = sf_volume.total_size
|
||||||
|
|
||||||
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
vol_snap_1 = self._create_and_test_snapshot_2(vm_1_root_volume.id, sf_volume_id, sf_volume_id + 1, primary_storage_db_id, sf_volume_size,
|
||||||
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_account_id, 2, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
@ -1040,14 +1041,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_2_root_volume_name = vm_2_root_volume.name
|
vm_2_root_volume_name = vm_2_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 4, self, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 4, self, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume_2['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume_2.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1056,7 +1057,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_1 = Volume.create_from_snapshot(self.apiClient, vol_snap_2.id, services, account=self.account.name, domainid=self.domain.id)
|
volume_created_from_snapshot_1 = Volume.create_from_snapshot(self.apiClient, vol_snap_2.id, services, account=self.account.name, domainid=self.domain.id)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1096,14 +1097,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_3_root_volume_name = vm_3_root_volume.name
|
vm_3_root_volume_name = vm_3_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 6, self, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 6, self, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, vm_3_root_volume_name)
|
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, vm_3_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume_3['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume_3.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1112,7 +1113,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_a = Volume.create_from_snapshot(self.apiClient, vol_snap_b.id, services, account=self.account.name, domainid=self.domain.id)
|
volume_created_from_snapshot_a = Volume.create_from_snapshot(self.apiClient, vol_snap_b.id, services, account=self.account.name, domainid=self.domain.id)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 7, self, TestSnapshots._should_be_seven_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 7, self, TestSnapshots._should_be_seven_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1124,7 +1125,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
# should still be 7 volumes because the SolidFire volume for the root disk of the VM just destroyed
|
# should still be 7 volumes because the SolidFire volume for the root disk of the VM just destroyed
|
||||||
# is still needed for the SolidFire snapshots
|
# is still needed for the SolidFire snapshots
|
||||||
@ -1133,14 +1134,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine_2.delete(self.apiClient, True)
|
virtual_machine_2.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 6, self, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 6, self, TestSnapshots._should_be_six_volumes_in_list_err_msg)
|
||||||
|
|
||||||
virtual_machine_3.delete(self.apiClient, True)
|
virtual_machine_3.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 5, self, TestSnapshots._should_be_five_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1149,7 +1150,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 4, self, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 4, self, TestSnapshots._should_be_four_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1158,7 +1159,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1167,7 +1168,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_b)
|
self._delete_and_test_snapshot(vol_snap_b)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
# should still be 2 volumes because the SolidFire volume for the root disk of the VM just destroyed
|
# should still be 2 volumes because the SolidFire volume for the root disk of the VM just destroyed
|
||||||
# is still needed for the SolidFire snapshots
|
# is still needed for the SolidFire snapshots
|
||||||
@ -1176,7 +1177,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_a)
|
self._delete_and_test_snapshot(vol_snap_a)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -1211,14 +1212,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self, TestSnapshots._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1247,7 +1248,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -1256,7 +1257,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_archive_snapshot(vol_snap_1_archive)
|
self._delete_and_test_archive_snapshot(vol_snap_1_archive)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -1265,7 +1266,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_2)
|
self._delete_and_test_snapshot(vol_snap_2)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1292,14 +1293,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_1_root_volume_name = vm_1_root_volume.name
|
vm_1_root_volume_name = vm_1_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
sf_volume = self._get_sf_volume_by_name(sf_volumes, vm_1_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume['volumeID'])
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1342,14 +1343,14 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vm_2_root_volume_name = vm_2_root_volume.name
|
vm_2_root_volume_name = vm_2_root_volume.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots_2 = self.sf_client.list_snapshots(volume_id=sf_volume_2['volumeID'])
|
sf_snapshots_2 = self.sfe.list_snapshots(volume_id=sf_volume_2.volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots_2, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1362,7 +1363,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
volume_created_from_snapshot_name = volume_created_from_snapshot.name
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1370,7 +1371,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
sf_volume_2 = self._get_sf_volume_by_name(sf_volumes, vm_2_root_volume_name)
|
||||||
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
sf_volume_3 = self._get_sf_volume_by_name(sf_volumes, volume_created_from_snapshot_name)
|
||||||
|
|
||||||
sf_util.check_list(sf_volume_3['volumeAccessGroups'], 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
sf_util.check_list(sf_volume_3.volume_access_groups, 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
||||||
|
|
||||||
volume_created_from_snapshot = virtual_machine.attach_volume(
|
volume_created_from_snapshot = virtual_machine.attach_volume(
|
||||||
self.apiClient,
|
self.apiClient,
|
||||||
@ -1382,7 +1383,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine.delete(self.apiClient, True)
|
virtual_machine.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 3, self, TestSnapshots._should_be_three_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1404,7 +1405,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._delete_and_test_snapshot(vol_snap_3)
|
self._delete_and_test_snapshot(vol_snap_3)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 2, self, TestSnapshots._should_be_two_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1414,7 +1415,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
virtual_machine_2.delete(self.apiClient, True)
|
virtual_machine_2.delete(self.apiClient, True)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 1, self, TestSnapshots._should_only_be_one_volume_in_list_err_msg)
|
||||||
|
|
||||||
@ -1434,7 +1435,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
data_volume.delete(self.apiClient)
|
data_volume.delete(self.apiClient)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
sf_util.check_list(sf_volumes, 0, self, TestSnapshots._should_be_zero_volumes_in_list_err_msg)
|
||||||
|
|
||||||
@ -1511,10 +1512,10 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_snapshot_to_return = None
|
sf_snapshot_to_return = None
|
||||||
|
|
||||||
for sf_snapshot in sf_snapshots:
|
for sf_snapshot in sf_snapshots:
|
||||||
if (sf_snapshot['snapshotID'] > most_recent_id):
|
if (sf_snapshot.snapshot_id > most_recent_id):
|
||||||
sf_snapshot_to_return = sf_snapshot
|
sf_snapshot_to_return = sf_snapshot
|
||||||
|
|
||||||
most_recent_id = sf_snapshot['snapshotID']
|
most_recent_id = sf_snapshot.snapshot_id
|
||||||
|
|
||||||
if (sf_snapshot_to_return == None):
|
if (sf_snapshot_to_return == None):
|
||||||
raise Exception("Unable to find the most recent SolidFire snapshot in the provided list")
|
raise Exception("Unable to find the most recent SolidFire snapshot in the provided list")
|
||||||
@ -1541,7 +1542,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_volume = None
|
sf_volume = None
|
||||||
|
|
||||||
for volume in sf_volumes:
|
for volume in sf_volumes:
|
||||||
if volume['name'] == sf_volume_name:
|
if volume.name == sf_volume_name:
|
||||||
sf_volume = volume
|
sf_volume = volume
|
||||||
|
|
||||||
break
|
break
|
||||||
@ -1560,7 +1561,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
sf_volume = None
|
sf_volume = None
|
||||||
|
|
||||||
for volume in sf_volumes:
|
for volume in sf_volumes:
|
||||||
if volume['volumeID'] == sf_volume_id:
|
if volume.volume_id == sf_volume_id:
|
||||||
sf_volume = volume
|
sf_volume = volume
|
||||||
|
|
||||||
break
|
break
|
||||||
@ -1582,7 +1583,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
def _check_sf_snapshot_does_not_exist(self, sf_snapshots, sf_snapshot_id):
|
def _check_sf_snapshot_does_not_exist(self, sf_snapshots, sf_snapshot_id):
|
||||||
for sf_snapshot in sf_snapshots:
|
for sf_snapshot in sf_snapshots:
|
||||||
if sf_snapshot["snapshotID"] == sf_snapshot:
|
if sf_snapshot.snapshot_id == sf_snapshot:
|
||||||
raise Exception("The following SolidFire snapshot ID should not exist: " + sf_snapshot_id)
|
raise Exception("The following SolidFire snapshot ID should not exist: " + sf_snapshot_id)
|
||||||
|
|
||||||
def _check_snapshot_details_do_not_exist(self, vol_snap_db_id):
|
def _check_snapshot_details_do_not_exist(self, vol_snap_db_id):
|
||||||
@ -1606,10 +1607,10 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
||||||
|
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume_id)
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, expected_num_snapshots, self, snapshot_err_msg)
|
sf_util.check_list(sf_snapshots, expected_num_snapshots, self, snapshot_err_msg)
|
||||||
|
|
||||||
@ -1621,16 +1622,16 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
vol_snap_db_id = self._get_cs_volume_snapshot_db_id(vol_snap)
|
vol_snap_db_id = self._get_cs_volume_snapshot_db_id(vol_snap)
|
||||||
|
|
||||||
self._check_snapshot_details(sf_snapshot_details, vol_snap_db_id, sf_volume_id, sf_snapshot['snapshotID'], primary_storage_db_id, sf_volume['totalSize'])
|
self._check_snapshot_details(sf_snapshot_details, vol_snap_db_id, sf_volume_id, sf_snapshot.snapshot_id, primary_storage_db_id, sf_volume.total_size)
|
||||||
|
|
||||||
return vol_snap
|
return vol_snap
|
||||||
|
|
||||||
# used when SolidFire snapshots are being used for CloudStack volume snapshots to create a backup on secondary storage
|
# used when SolidFire snapshots are being used for CloudStack volume snapshots to create a backup on secondary storage
|
||||||
def _create_and_test_archive_snapshot(self, volume_id_for_snapshot, sf_volume):
|
def _create_and_test_archive_snapshot(self, volume_id_for_snapshot, sf_volume):
|
||||||
sf_volume_id = sf_volume['volumeID']
|
sf_volume_id = sf_volume.volume_id
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots_orig = self.sf_client.list_snapshots(sf_volume_id)
|
sf_snapshots_orig = self.sfe.list_snapshots(sf_volume_id).snapshots
|
||||||
|
|
||||||
vol_snap = Snapshot.create(
|
vol_snap = Snapshot.create(
|
||||||
self.apiClient,
|
self.apiClient,
|
||||||
@ -1641,7 +1642,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(sf_volume_id)
|
sf_snapshots = self.sfe.list_snapshots(sf_volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, len(sf_snapshots_orig), self, "A new SolidFire snapshot was detected.")
|
sf_util.check_list(sf_snapshots, len(sf_snapshots_orig), self, "A new SolidFire snapshot was detected.")
|
||||||
|
|
||||||
@ -1662,7 +1663,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
self._wait_for_snapshot_state(vol_snap.id, Snapshot.BACKED_UP)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume_id)
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume_id).snapshots
|
||||||
|
|
||||||
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
sf_util.check_list(sf_snapshots, 0, self, TestSnapshots._should_be_zero_snapshots_in_list_err_msg)
|
||||||
|
|
||||||
@ -1675,13 +1676,13 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._check_snapshot_details_2(sf_snapshot_details, vol_snap_db_id, sf_volume_id_for_volume_snapshot, primary_storage_db_id, sf_volume_size)
|
self._check_snapshot_details_2(sf_snapshot_details, vol_snap_db_id, sf_volume_id_for_volume_snapshot, primary_storage_db_id, sf_volume_size)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, expected_num_volumes, self, volume_err_msg)
|
sf_util.check_list(sf_volumes, expected_num_volumes, self, volume_err_msg)
|
||||||
|
|
||||||
sf_volume_for_snapshot = self._get_sf_volume_by_id(sf_volumes, sf_volume_id_for_volume_snapshot)
|
sf_volume_for_snapshot = self._get_sf_volume_by_id(sf_volumes, sf_volume_id_for_volume_snapshot)
|
||||||
|
|
||||||
sf_util.check_list(sf_volume_for_snapshot['volumeAccessGroups'], 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
sf_util.check_list(sf_volume_for_snapshot.volume_access_groups, 0, self, TestSnapshots._should_be_zero_volume_access_groups_in_list_err_msg)
|
||||||
|
|
||||||
return vol_snap
|
return vol_snap
|
||||||
|
|
||||||
@ -1722,7 +1723,7 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
vol_snap.delete(self.apiClient)
|
vol_snap.delete(self.apiClient)
|
||||||
|
|
||||||
# Get snapshot information for volume from SolidFire cluster
|
# Get snapshot information for volume from SolidFire cluster
|
||||||
sf_snapshots = self.sf_client.list_snapshots(volume_id=sf_volume_id)
|
sf_snapshots = self.sfe.list_snapshots(volume_id=sf_volume_id).snapshots
|
||||||
|
|
||||||
self._check_sf_snapshot_does_not_exist(sf_snapshots, sf_snapshot_id)
|
self._check_sf_snapshot_does_not_exist(sf_snapshots, sf_snapshot_id)
|
||||||
|
|
||||||
@ -1741,6 +1742,6 @@ class TestSnapshots(cloudstackTestCase):
|
|||||||
self._check_snapshot_details_do_not_exist(vol_snap_db_id)
|
self._check_snapshot_details_do_not_exist(vol_snap_db_id)
|
||||||
|
|
||||||
# Get volume information from SolidFire cluster
|
# Get volume information from SolidFire cluster
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_util.check_list(sf_volumes, expected_num_volumes, self, volume_err_msg)
|
sf_util.check_list(sf_volumes, expected_num_volumes, self, volume_err_msg)
|
||||||
|
|||||||
@ -20,6 +20,8 @@ import random
|
|||||||
import SignedAPICall
|
import SignedAPICall
|
||||||
import XenAPI
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
from util import sf_util
|
from util import sf_util
|
||||||
|
|
||||||
# All tests inherit from cloudstackTestCase
|
# All tests inherit from cloudstackTestCase
|
||||||
@ -34,8 +36,6 @@ from marvin.lib.common import get_domain, get_template, get_zone, list_clusters,
|
|||||||
# utils - utility classes for common cleanup, external library wrappers, etc.
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
from marvin.lib.utils import cleanup_resources
|
from marvin.lib.utils import cleanup_resources
|
||||||
|
|
||||||
from solidfire import solidfire_element_api as sf_api
|
|
||||||
|
|
||||||
# Prerequisites:
|
# Prerequisites:
|
||||||
# Only one zone
|
# Only one zone
|
||||||
# Only one pod
|
# Only one pod
|
||||||
@ -71,7 +71,6 @@ class TestData():
|
|||||||
storageTag2 = "SolidFire_Volume_1"
|
storageTag2 = "SolidFire_Volume_1"
|
||||||
tags = "tags"
|
tags = "tags"
|
||||||
templateCacheName = "centos56-x86-64-xen"
|
templateCacheName = "centos56-x86-64-xen"
|
||||||
templateName = "templatename"
|
|
||||||
testAccount = "testaccount"
|
testAccount = "testaccount"
|
||||||
url = "url"
|
url = "url"
|
||||||
user = "user"
|
user = "user"
|
||||||
@ -86,7 +85,7 @@ class TestData():
|
|||||||
self.testdata = {
|
self.testdata = {
|
||||||
TestData.solidFire: {
|
TestData.solidFire: {
|
||||||
TestData.mvip: "192.168.139.112",
|
TestData.mvip: "192.168.139.112",
|
||||||
TestData.login: "admin",
|
TestData.username: "admin",
|
||||||
TestData.password: "admin",
|
TestData.password: "admin",
|
||||||
TestData.port: 443,
|
TestData.port: 443,
|
||||||
TestData.url: "https://192.168.139.112:443"
|
TestData.url: "https://192.168.139.112:443"
|
||||||
@ -208,7 +207,6 @@ class TestData():
|
|||||||
TestData.volume_1: {
|
TestData.volume_1: {
|
||||||
TestData.diskName: "test-volume",
|
TestData.diskName: "test-volume",
|
||||||
},
|
},
|
||||||
TestData.templateName: "CentOS 5.6(64-bit) no GUI (XenServer)",
|
|
||||||
TestData.zoneId: 1,
|
TestData.zoneId: 1,
|
||||||
TestData.clusterId1: 1,
|
TestData.clusterId1: 1,
|
||||||
TestData.clusterId2: 2,
|
TestData.clusterId2: 2,
|
||||||
@ -224,7 +222,9 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
def setUpClass(cls):
|
def setUpClass(cls):
|
||||||
# Set up API client
|
# Set up API client
|
||||||
testclient = super(TestVMMigrationWithStorage, cls).getClsTestClient()
|
testclient = super(TestVMMigrationWithStorage, cls).getClsTestClient()
|
||||||
|
|
||||||
cls.apiClient = testclient.getApiClient()
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
cls.dbConnection = testclient.getDbConnection()
|
cls.dbConnection = testclient.getDbConnection()
|
||||||
|
|
||||||
cls.testdata = TestData().testdata
|
cls.testdata = TestData().testdata
|
||||||
@ -250,13 +250,15 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
cls.xen_session_2.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
cls.xen_session_2.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
# Set up SolidFire connection
|
# Set up SolidFire connection
|
||||||
cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire])
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
# Get Resources from Cloud Infrastructure
|
# Get Resources from Cloud Infrastructure
|
||||||
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
||||||
cls.cluster_1 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId1])[0]
|
cls.cluster_1 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId1])[0]
|
||||||
cls.cluster_2 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId2])[0]
|
cls.cluster_2 = list_clusters(cls.apiClient, id=cls.testdata[TestData.clusterId2])[0]
|
||||||
cls.template = get_template(cls.apiClient, cls.zone.id, template_name=cls.testdata[TestData.templateName])
|
cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
# Create test account
|
# Create test account
|
||||||
@ -330,6 +332,8 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
cleanup_resources(cls.apiClient, cls._cleanup)
|
cleanup_resources(cls.apiClient, cls._cleanup)
|
||||||
|
|
||||||
cls.primary_storage.delete(cls.apiClient)
|
cls.primary_storage.delete(cls.apiClient)
|
||||||
|
|
||||||
|
sf_util.purge_solidfire_volumes(cls.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
@ -340,7 +344,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
try:
|
try:
|
||||||
cleanup_resources(self.apiClient, self.cleanup)
|
cleanup_resources(self.apiClient, self.cleanup)
|
||||||
|
|
||||||
sf_util.purge_solidfire_volumes(self.sf_client)
|
sf_util.purge_solidfire_volumes(self.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(self): %s" % e)
|
logging.debug("Exception in tearDownClass(self): %s" % e)
|
||||||
|
|
||||||
@ -366,7 +370,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
||||||
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
||||||
|
|
||||||
@ -386,7 +390,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
cs_data_volume
|
cs_data_volume
|
||||||
)
|
)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
||||||
|
|
||||||
@ -451,7 +455,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
||||||
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
||||||
|
|
||||||
@ -497,7 +501,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
sf_account_id = sf_util.get_sf_account_id(self.cs_api, self.account.id, self.primary_storage.id, self,
|
||||||
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
TestVMMigrationWithStorage._sf_account_id_should_be_non_zero_int_err_msg)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
||||||
|
|
||||||
@ -517,7 +521,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
cs_data_volume
|
cs_data_volume
|
||||||
)
|
)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
||||||
|
|
||||||
@ -549,7 +553,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
cs_root_volume = self._get_updated_cs_volume(cs_root_volume.id)
|
cs_root_volume = self._get_updated_cs_volume(cs_root_volume.id)
|
||||||
cs_data_volume = self._get_updated_cs_volume(cs_data_volume.id)
|
cs_data_volume = self._get_updated_cs_volume(cs_data_volume.id)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
||||||
dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
||||||
@ -580,7 +584,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
|
|
||||||
cs_volume = self._get_updated_cs_volume(cs_volume.id)
|
cs_volume = self._get_updated_cs_volume(cs_volume.id)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
dest_sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_volume.name, self)
|
dest_sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_volume.name, self)
|
||||||
|
|
||||||
@ -624,7 +628,7 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
self._verifyFields(cs_root_volume_refreshed, src_sf_root_volume)
|
self._verifyFields(cs_root_volume_refreshed, src_sf_root_volume)
|
||||||
self._verifyFields(cs_data_volume_refreshed, src_sf_data_volume)
|
self._verifyFields(cs_data_volume_refreshed, src_sf_data_volume)
|
||||||
|
|
||||||
sf_volumes = sf_util.get_not_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_not_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
dest_sf_root_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_root_volume.name, self)
|
||||||
dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
dest_sf_data_volume = sf_util.check_and_get_sf_volume(sf_volumes, cs_data_volume.name, self)
|
||||||
@ -633,11 +637,11 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
self._verify_xenserver_state(dest_xen_session, dest_sf_data_volume, src_xen_session, src_sf_data_volume)
|
self._verify_xenserver_state(dest_xen_session, dest_sf_data_volume, src_xen_session, src_sf_data_volume)
|
||||||
|
|
||||||
def _verify_different_volume_access_groups(self, src_sf_volume, dest_sf_volume):
|
def _verify_different_volume_access_groups(self, src_sf_volume, dest_sf_volume):
|
||||||
src_vags = src_sf_volume['volumeAccessGroups']
|
src_vags = src_sf_volume.volume_access_groups
|
||||||
|
|
||||||
sf_util.check_list(src_vags, 1, self, "'src_vags' should be a list with only one element in it.")
|
sf_util.check_list(src_vags, 1, self, "'src_vags' should be a list with only one element in it.")
|
||||||
|
|
||||||
dest_vags = dest_sf_volume['volumeAccessGroups']
|
dest_vags = dest_sf_volume.volume_access_groups
|
||||||
|
|
||||||
sf_util.check_list(dest_vags, 1, self, "'dest_vags' should be a list with only one element in it.")
|
sf_util.check_list(dest_vags, 1, self, "'dest_vags' should be a list with only one element in it.")
|
||||||
|
|
||||||
@ -647,23 +651,23 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
return list_volumes(self.apiClient, listall=True, id=cs_volume_id)[0]
|
return list_volumes(self.apiClient, listall=True, id=cs_volume_id)[0]
|
||||||
|
|
||||||
def _verify_same_account(self, src_sf_volume, dest_sf_volume):
|
def _verify_same_account(self, src_sf_volume, dest_sf_volume):
|
||||||
self.assertEqual(src_sf_volume['accountID'], dest_sf_volume['accountID'], "The source and destination volumes should be in the same SolidFire account.")
|
self.assertEqual(src_sf_volume.account_id, dest_sf_volume.account_id, "The source and destination volumes should be in the same SolidFire account.")
|
||||||
|
|
||||||
def _verifySfVolumeIds(self, src_sf_volume, dest_sf_volume):
|
def _verifySfVolumeIds(self, src_sf_volume, dest_sf_volume):
|
||||||
self.assert_(src_sf_volume['volumeID'] < dest_sf_volume['volumeID'],
|
self.assert_(src_sf_volume.volume_id < dest_sf_volume.volume_id,
|
||||||
"The destination SolidFire root volume's ID should be greater than the id of the source one.")
|
"The destination SolidFire root volume's ID should be greater than the id of the source one.")
|
||||||
|
|
||||||
# verify the name, folder, and iscsi_name
|
# verify the name, folder, and iscsi_name
|
||||||
def _verifyFields(self, cs_volume, sf_volume):
|
def _verifyFields(self, cs_volume, sf_volume):
|
||||||
self.assert_(cs_volume.name == sf_volume['name'], "The CloudStack volume name does not match the SolidFire volume name.")
|
self.assert_(cs_volume.name == sf_volume.name, "The CloudStack volume name does not match the SolidFire volume name.")
|
||||||
|
|
||||||
cs_volume_folder = self._get_cs_volume_folder(cs_volume.id)
|
cs_volume_folder = self._get_cs_volume_folder(cs_volume.id)
|
||||||
|
|
||||||
self.assert_(int(cs_volume_folder) == sf_volume['volumeID'], "The CloudStack folder name does not match the SolidFire volume ID.")
|
self.assert_(int(cs_volume_folder) == sf_volume.volume_id, "The CloudStack folder name does not match the SolidFire volume ID.")
|
||||||
|
|
||||||
cs_volume_iscsi_name = self._get_cs_volume_iscsi_name(cs_volume.id)
|
cs_volume_iscsi_name = self._get_cs_volume_iscsi_name(cs_volume.id)
|
||||||
|
|
||||||
self.assert_(cs_volume_iscsi_name == sf_util.format_iqn(sf_volume['iqn']), "The CloudStack volume iscsi_name does not match the SolidFire volume IQN.")
|
self.assert_(cs_volume_iscsi_name == sf_util.format_iqn(sf_volume.iqn), "The CloudStack volume iscsi_name does not match the SolidFire volume IQN.")
|
||||||
|
|
||||||
def _get_cs_volume_property(self, cs_volume_id, volume_property):
|
def _get_cs_volume_property(self, cs_volume_id, volume_property):
|
||||||
sql_query = "Select " + volume_property + " From volumes Where uuid = '" + cs_volume_id + "'"
|
sql_query = "Select " + volume_property + " From volumes Where uuid = '" + cs_volume_id + "'"
|
||||||
@ -688,10 +692,10 @@ class TestVMMigrationWithStorage(cloudstackTestCase):
|
|||||||
sf_util.check_list(sql_result, 0, self, "The cloud.volume_details table should not have any name fields that start with 'basic_'.")
|
sf_util.check_list(sql_result, 0, self, "The cloud.volume_details table should not have any name fields that start with 'basic_'.")
|
||||||
|
|
||||||
def _verify_xenserver_state(self, xen_session_1, sf_volume_1, xen_session_2, sf_volume_2):
|
def _verify_xenserver_state(self, xen_session_1, sf_volume_1, xen_session_2, sf_volume_2):
|
||||||
sr_name = sf_util.format_iqn(sf_volume_1["iqn"])
|
sr_name = sf_util.format_iqn(sf_volume_1.iqn)
|
||||||
|
|
||||||
sf_util.check_xen_sr(sr_name, xen_session_1, self, False)
|
sf_util.check_xen_sr(sr_name, xen_session_1, self, False)
|
||||||
|
|
||||||
sr_name = sf_util.format_iqn(sf_volume_2["iqn"])
|
sr_name = sf_util.format_iqn(sf_volume_2.iqn)
|
||||||
|
|
||||||
sf_util.check_xen_sr(sr_name, xen_session_2, self)
|
sf_util.check_xen_sr(sr_name, xen_session_2, self)
|
||||||
|
|||||||
@ -20,6 +20,8 @@ import random
|
|||||||
import SignedAPICall
|
import SignedAPICall
|
||||||
import XenAPI
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
from util import sf_util
|
from util import sf_util
|
||||||
|
|
||||||
# All tests inherit from cloudstackTestCase
|
# All tests inherit from cloudstackTestCase
|
||||||
@ -36,8 +38,6 @@ from marvin.lib.common import get_domain, get_template, get_zone, list_hosts, li
|
|||||||
# utils - utility classes for common cleanup, external library wrappers, etc.
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
from marvin.lib.utils import cleanup_resources
|
from marvin.lib.utils import cleanup_resources
|
||||||
|
|
||||||
from solidfire import solidfire_element_api as sf_api
|
|
||||||
|
|
||||||
# Prerequisites:
|
# Prerequisites:
|
||||||
# Only one zone
|
# Only one zone
|
||||||
# Only one pod
|
# Only one pod
|
||||||
@ -63,7 +63,6 @@ class TestData:
|
|||||||
solidFire = "solidfire"
|
solidFire = "solidfire"
|
||||||
storageTag = "SolidFire_SAN_1"
|
storageTag = "SolidFire_SAN_1"
|
||||||
tags = "tags"
|
tags = "tags"
|
||||||
templateName = "templatename"
|
|
||||||
url = "url"
|
url = "url"
|
||||||
user = "user"
|
user = "user"
|
||||||
username = "username"
|
username = "username"
|
||||||
@ -76,7 +75,7 @@ class TestData:
|
|||||||
self.testdata = {
|
self.testdata = {
|
||||||
TestData.solidFire: {
|
TestData.solidFire: {
|
||||||
TestData.mvip: "192.168.139.112",
|
TestData.mvip: "192.168.139.112",
|
||||||
TestData.login: "admin",
|
TestData.username: "admin",
|
||||||
TestData.password: "admin",
|
TestData.password: "admin",
|
||||||
TestData.port: 443,
|
TestData.port: 443,
|
||||||
TestData.url: "https://192.168.139.112:443"
|
TestData.url: "https://192.168.139.112:443"
|
||||||
@ -211,7 +210,6 @@ class TestData:
|
|||||||
"volume2": {
|
"volume2": {
|
||||||
"diskname": "testvolume2",
|
"diskname": "testvolume2",
|
||||||
},
|
},
|
||||||
TestData.templateName: "CentOS 5.6(64-bit) no GUI (XenServer)",
|
|
||||||
TestData.zoneId: 1,
|
TestData.zoneId: 1,
|
||||||
TestData.clusterId: 1,
|
TestData.clusterId: 1,
|
||||||
TestData.domainId: 1,
|
TestData.domainId: 1,
|
||||||
@ -237,7 +235,9 @@ class TestVMSnapshots(cloudstackTestCase):
|
|||||||
def setUpClass(cls):
|
def setUpClass(cls):
|
||||||
# Set up API client
|
# Set up API client
|
||||||
testclient = super(TestVMSnapshots, cls).getClsTestClient()
|
testclient = super(TestVMSnapshots, cls).getClsTestClient()
|
||||||
|
|
||||||
cls.apiClient = testclient.getApiClient()
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
|
|
||||||
cls.testdata = TestData().testdata
|
cls.testdata = TestData().testdata
|
||||||
|
|
||||||
@ -252,11 +252,13 @@ class TestVMSnapshots(cloudstackTestCase):
|
|||||||
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
# Set up SolidFire connection
|
# Set up SolidFire connection
|
||||||
cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire])
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
# Get Resources from Cloud Infrastructure
|
# Get Resources from Cloud Infrastructure
|
||||||
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
||||||
template = get_template(cls.apiClient, cls.zone.id, template_name=cls.testdata[TestData.templateName])
|
template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
# Create test account
|
# Create test account
|
||||||
@ -332,7 +334,7 @@ class TestVMSnapshots(cloudstackTestCase):
|
|||||||
|
|
||||||
cls.primary_storage.delete(cls.apiClient)
|
cls.primary_storage.delete(cls.apiClient)
|
||||||
|
|
||||||
sf_util.purge_solidfire_volumes(cls.sf_client)
|
sf_util.purge_solidfire_volumes(cls.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
|
|||||||
@ -20,6 +20,8 @@ import random
|
|||||||
import SignedAPICall
|
import SignedAPICall
|
||||||
import XenAPI
|
import XenAPI
|
||||||
|
|
||||||
|
from solidfire.factory import ElementFactory
|
||||||
|
|
||||||
from util import sf_util
|
from util import sf_util
|
||||||
|
|
||||||
# All tests inherit from cloudstackTestCase
|
# All tests inherit from cloudstackTestCase
|
||||||
@ -39,8 +41,6 @@ from marvin.lib.common import get_domain, get_template, get_zone, list_clusters,
|
|||||||
# utils - utility classes for common cleanup, external library wrappers, etc.
|
# utils - utility classes for common cleanup, external library wrappers, etc.
|
||||||
from marvin.lib.utils import cleanup_resources
|
from marvin.lib.utils import cleanup_resources
|
||||||
|
|
||||||
from solidfire import solidfire_element_api as sf_api
|
|
||||||
|
|
||||||
# Prerequisites:
|
# Prerequisites:
|
||||||
# Only one zone
|
# Only one zone
|
||||||
# Only one pod
|
# Only one pod
|
||||||
@ -71,7 +71,6 @@ class TestData():
|
|||||||
storageTag = "SolidFire_SAN_1"
|
storageTag = "SolidFire_SAN_1"
|
||||||
tags = "tags"
|
tags = "tags"
|
||||||
templateCacheName = "centos56-x86-64-xen"
|
templateCacheName = "centos56-x86-64-xen"
|
||||||
templateName = "templatename"
|
|
||||||
testAccount = "testaccount"
|
testAccount = "testaccount"
|
||||||
url = "url"
|
url = "url"
|
||||||
user = "user"
|
user = "user"
|
||||||
@ -87,7 +86,7 @@ class TestData():
|
|||||||
self.testdata = {
|
self.testdata = {
|
||||||
TestData.solidFire: {
|
TestData.solidFire: {
|
||||||
TestData.mvip: "192.168.139.112",
|
TestData.mvip: "192.168.139.112",
|
||||||
TestData.login: "admin",
|
TestData.username: "admin",
|
||||||
TestData.password: "admin",
|
TestData.password: "admin",
|
||||||
TestData.port: 443,
|
TestData.port: 443,
|
||||||
TestData.url: "https://192.168.139.112:443"
|
TestData.url: "https://192.168.139.112:443"
|
||||||
@ -168,7 +167,6 @@ class TestData():
|
|||||||
TestData.volume_2: {
|
TestData.volume_2: {
|
||||||
TestData.diskName: "test-volume-2",
|
TestData.diskName: "test-volume-2",
|
||||||
},
|
},
|
||||||
TestData.templateName: "CentOS 5.6(64-bit) no GUI (XenServer)",
|
|
||||||
TestData.zoneId: 1,
|
TestData.zoneId: 1,
|
||||||
TestData.clusterId: 1,
|
TestData.clusterId: 1,
|
||||||
TestData.domainId: 1,
|
TestData.domainId: 1,
|
||||||
@ -192,6 +190,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
# Set up API client
|
# Set up API client
|
||||||
testclient = super(TestVolumes, cls).getClsTestClient()
|
testclient = super(TestVolumes, cls).getClsTestClient()
|
||||||
cls.apiClient = testclient.getApiClient()
|
cls.apiClient = testclient.getApiClient()
|
||||||
|
cls.configData = testclient.getParsedTestDataConfig()
|
||||||
cls.dbConnection = testclient.getDbConnection()
|
cls.dbConnection = testclient.getDbConnection()
|
||||||
|
|
||||||
cls.testdata = TestData().testdata
|
cls.testdata = TestData().testdata
|
||||||
@ -212,12 +211,14 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password])
|
||||||
|
|
||||||
# Set up SolidFire connection
|
# Set up SolidFire connection
|
||||||
cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire])
|
solidfire = cls.testdata[TestData.solidFire]
|
||||||
|
|
||||||
|
cls.sfe = ElementFactory.create(solidfire[TestData.mvip], solidfire[TestData.username], solidfire[TestData.password])
|
||||||
|
|
||||||
# Get Resources from Cloud Infrastructure
|
# Get Resources from Cloud Infrastructure
|
||||||
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
cls.zone = get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId])
|
||||||
cls.cluster = list_clusters(cls.apiClient)[0]
|
cls.cluster = list_clusters(cls.apiClient)[0]
|
||||||
cls.template = get_template(cls.apiClient, cls.zone.id, template_name=cls.testdata[TestData.templateName])
|
cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"])
|
||||||
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId])
|
||||||
|
|
||||||
# Create test account
|
# Create test account
|
||||||
@ -304,7 +305,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
|
|
||||||
cls.primary_storage.delete(cls.apiClient)
|
cls.primary_storage.delete(cls.apiClient)
|
||||||
|
|
||||||
sf_util.purge_solidfire_volumes(cls.sf_client)
|
sf_util.purge_solidfire_volumes(cls.sfe)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
logging.debug("Exception in tearDownClass(cls): %s" % e)
|
||||||
|
|
||||||
@ -328,16 +329,16 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, TestData.templateCacheName, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, TestData.templateCacheName, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
"The volume should not be in a VAG."
|
"The volume should not be in a VAG."
|
||||||
)
|
)
|
||||||
|
|
||||||
sf_account_id = sf_volume["accountID"]
|
sf_account_id = sf_volume.account_id
|
||||||
|
|
||||||
sf_account = self.sf_client.get_account_by_id(sf_account_id)["account"]
|
sf_account = self.sfe.get_account_by_id(sf_account_id).account
|
||||||
|
|
||||||
sf_account_name = sf_account["username"]
|
sf_account_name = sf_account.username
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
sf_account_name.endswith("_1"),
|
sf_account_name.endswith("_1"),
|
||||||
@ -504,7 +505,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
"The volume should not be in a VAG."
|
"The volume should not be in a VAG."
|
||||||
)
|
)
|
||||||
@ -723,7 +724,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -747,7 +748,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -847,7 +848,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -871,7 +872,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
sf_volume = sf_util.check_and_get_sf_volume(sf_volumes, vol.name, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -1075,7 +1076,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
|
sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -1182,7 +1183,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
|
sf_util.check_size_and_iops(sf_volume, vol, sf_volume_size, self)
|
||||||
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
0,
|
0,
|
||||||
TestVolumes._volume_should_not_be_in_a_vag
|
TestVolumes._volume_should_not_be_in_a_vag
|
||||||
)
|
)
|
||||||
@ -1489,7 +1490,7 @@ class TestVolumes(cloudstackTestCase):
|
|||||||
sf_util.check_xen_sr(xen_sr_name, self.xen_session, self, should_exist)
|
sf_util.check_xen_sr(xen_sr_name, self.xen_session, self, should_exist)
|
||||||
|
|
||||||
def _get_active_sf_volumes(self, sf_account_id=None):
|
def _get_active_sf_volumes(self, sf_account_id=None):
|
||||||
sf_volumes = sf_util.get_active_sf_volumes(self.sf_client, sf_account_id)
|
sf_volumes = sf_util.get_active_sf_volumes(self.sfe, sf_account_id)
|
||||||
|
|
||||||
self.assertNotEqual(
|
self.assertNotEqual(
|
||||||
len(sf_volumes),
|
len(sf_volumes),
|
||||||
|
|||||||
@ -68,20 +68,20 @@ def _set_supports_resign_for_table(supports_resign, db_connection, table):
|
|||||||
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
# make sure you can connect to MySQL: https://teamtreehouse.com/community/cant-connect-remotely-to-mysql-server-with-mysql-workbench
|
||||||
db_connection.execute(sql_query)
|
db_connection.execute(sql_query)
|
||||||
|
|
||||||
def purge_solidfire_volumes(sf_client):
|
def purge_solidfire_volumes(sfe):
|
||||||
deleted_volumes = sf_client.list_deleted_volumes()
|
deleted_volumes = sfe.list_deleted_volumes()
|
||||||
|
|
||||||
for deleted_volume in deleted_volumes:
|
for deleted_volume in deleted_volumes.volumes:
|
||||||
sf_client.purge_deleted_volume(deleted_volume['volumeID'])
|
sfe.purge_deleted_volume(deleted_volume.volume_id)
|
||||||
|
|
||||||
def get_not_active_sf_volumes(sf_client, sf_account_id=None):
|
def get_not_active_sf_volumes(sfe, sf_account_id=None):
|
||||||
if sf_account_id is not None:
|
if sf_account_id is not None:
|
||||||
sf_volumes = sf_client.list_volumes_for_account(sf_account_id)
|
sf_volumes = sfe.list_volumes_for_account(sf_account_id).volumes
|
||||||
|
|
||||||
if sf_volumes is not None and len(sf_volumes) > 0:
|
if sf_volumes is not None and len(sf_volumes) > 0:
|
||||||
sf_volumes = _get_not_active_sf_volumes_only(sf_volumes)
|
sf_volumes = _get_not_active_sf_volumes_only(sf_volumes)
|
||||||
else:
|
else:
|
||||||
sf_volumes = sf_client.list_deleted_volumes()
|
sf_volumes = sfe.list_deleted_volumes().volumes
|
||||||
|
|
||||||
return sf_volumes
|
return sf_volumes
|
||||||
|
|
||||||
@ -89,19 +89,19 @@ def _get_not_active_sf_volumes_only(sf_volumes):
|
|||||||
not_active_sf_volumes_only = []
|
not_active_sf_volumes_only = []
|
||||||
|
|
||||||
for sf_volume in sf_volumes:
|
for sf_volume in sf_volumes:
|
||||||
if sf_volume["status"] != "active":
|
if sf_volume.status != "active":
|
||||||
not_active_sf_volumes_only.append(sf_volume)
|
not_active_sf_volumes_only.append(sf_volume)
|
||||||
|
|
||||||
return not_active_sf_volumes_only
|
return not_active_sf_volumes_only
|
||||||
|
|
||||||
def get_active_sf_volumes(sf_client, sf_account_id=None):
|
def get_active_sf_volumes(sfe, sf_account_id=None):
|
||||||
if sf_account_id is not None:
|
if sf_account_id is not None:
|
||||||
sf_volumes = sf_client.list_volumes_for_account(sf_account_id)
|
sf_volumes = sfe.list_volumes_for_account(sf_account_id).volumes
|
||||||
|
|
||||||
if sf_volumes is not None and len(sf_volumes) > 0:
|
if sf_volumes is not None and len(sf_volumes) > 0:
|
||||||
sf_volumes = _get_active_sf_volumes_only(sf_volumes)
|
sf_volumes = _get_active_sf_volumes_only(sf_volumes)
|
||||||
else:
|
else:
|
||||||
sf_volumes = sf_client.list_active_volumes()
|
sf_volumes = sfe.list_active_volumes().volumes
|
||||||
|
|
||||||
return sf_volumes
|
return sf_volumes
|
||||||
|
|
||||||
@ -109,7 +109,7 @@ def _get_active_sf_volumes_only(sf_volumes):
|
|||||||
active_sf_volumes_only = []
|
active_sf_volumes_only = []
|
||||||
|
|
||||||
for sf_volume in sf_volumes:
|
for sf_volume in sf_volumes:
|
||||||
if sf_volume["status"] == "active":
|
if sf_volume.status == "active":
|
||||||
active_sf_volumes_only.append(sf_volume)
|
active_sf_volumes_only.append(sf_volume)
|
||||||
|
|
||||||
return active_sf_volumes_only
|
return active_sf_volumes_only
|
||||||
@ -118,7 +118,7 @@ def check_and_get_sf_volume(sf_volumes, sf_volume_name, obj_assert, should_exist
|
|||||||
sf_volume = None
|
sf_volume = None
|
||||||
|
|
||||||
for volume in sf_volumes:
|
for volume in sf_volumes:
|
||||||
if volume['name'] == sf_volume_name:
|
if volume.name == sf_volume_name:
|
||||||
sf_volume = volume
|
sf_volume = volume
|
||||||
break
|
break
|
||||||
|
|
||||||
@ -155,13 +155,13 @@ def check_xen_sr(xen_sr_name, xen_session, obj_assert, should_exist=True):
|
|||||||
|
|
||||||
def check_vag(sf_volume, sf_vag_id, obj_assert):
|
def check_vag(sf_volume, sf_vag_id, obj_assert):
|
||||||
obj_assert.assertEqual(
|
obj_assert.assertEqual(
|
||||||
len(sf_volume['volumeAccessGroups']),
|
len(sf_volume.volume_access_groups),
|
||||||
1,
|
1,
|
||||||
"The volume should only be in one VAG."
|
"The volume should only be in one VAG."
|
||||||
)
|
)
|
||||||
|
|
||||||
obj_assert.assertEqual(
|
obj_assert.assertEqual(
|
||||||
sf_volume['volumeAccessGroups'][0],
|
sf_volume.volume_access_groups[0],
|
||||||
sf_vag_id,
|
sf_vag_id,
|
||||||
"The volume is not in the VAG with the following ID: " + str(sf_vag_id) + "."
|
"The volume is not in the VAG with the following ID: " + str(sf_vag_id) + "."
|
||||||
)
|
)
|
||||||
@ -185,21 +185,21 @@ def format_iqn(iqn):
|
|||||||
|
|
||||||
def check_size_and_iops(sf_volume, cs_volume, size, obj_assert):
|
def check_size_and_iops(sf_volume, cs_volume, size, obj_assert):
|
||||||
obj_assert.assertEqual(
|
obj_assert.assertEqual(
|
||||||
sf_volume['qos']['minIOPS'],
|
sf_volume.qos.min_iops,
|
||||||
cs_volume.miniops,
|
cs_volume.miniops,
|
||||||
"Check QoS - Min IOPS: " + str(sf_volume['qos']['minIOPS'])
|
"Check QoS - Min IOPS: " + str(sf_volume.qos.min_iops)
|
||||||
)
|
)
|
||||||
|
|
||||||
obj_assert.assertEqual(
|
obj_assert.assertEqual(
|
||||||
sf_volume['qos']['maxIOPS'],
|
sf_volume.qos.max_iops,
|
||||||
cs_volume.maxiops,
|
cs_volume.maxiops,
|
||||||
"Check QoS - Max IOPS: " + str(sf_volume['qos']['maxIOPS'])
|
"Check QoS - Max IOPS: " + str(sf_volume.qos.max_iops)
|
||||||
)
|
)
|
||||||
|
|
||||||
obj_assert.assertEqual(
|
obj_assert.assertEqual(
|
||||||
sf_volume['totalSize'],
|
sf_volume.total_size,
|
||||||
size,
|
size,
|
||||||
"Check SolidFire volume size: " + str(sf_volume['totalSize'])
|
"Check SolidFire volume size: " + str(sf_volume.total_size)
|
||||||
)
|
)
|
||||||
|
|
||||||
def get_volume_size_with_hsr(cs_api, cs_volume, obj_assert):
|
def get_volume_size_with_hsr(cs_api, cs_volume, obj_assert):
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user