mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 08:42:29 +01:00
This PR implements Storage filesystem as a first class feature. https://cwiki.apache.org/confluence/display/CLOUDSTACK/Storage+Filesystem+as+a+First+Class+Feature Documentation PR: apache/cloudstack-documentation#420 Co-authored-by: Wei Zhou <weizhou@apache.org>
278 lines
11 KiB
Python
278 lines
11 KiB
Python
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
"""
|
|
Tests for Shared FileSystem
|
|
"""
|
|
import time
|
|
|
|
# Import Local Modules
|
|
from nose.plugins.attrib import attr
|
|
from marvin.cloudstackAPI import (createFirewallRule,
|
|
createPortForwardingRule)
|
|
from marvin.cloudstackTestCase import cloudstackTestCase
|
|
from marvin.lib.utils import (cleanup_resources,
|
|
wait_until)
|
|
from marvin.lib.base import (Account,
|
|
VirtualMachine,
|
|
Network,
|
|
SharedFS,
|
|
ServiceOffering,
|
|
NetworkOffering,
|
|
DiskOffering,
|
|
PublicIPAddress,
|
|
)
|
|
from marvin.lib.common import (get_domain,
|
|
get_zone,
|
|
get_template)
|
|
from marvin.codes import FAILED
|
|
|
|
from marvin.lib.decoratorGenerators import skipTestIf
|
|
|
|
|
|
class TestSharedFSLifecycle(cloudstackTestCase):
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
|
|
cls.testClient = super(TestSharedFSLifecycle, cls).getClsTestClient()
|
|
cls.apiclient = cls.testClient.getApiClient()
|
|
cls.services = cls.testClient.getParsedTestDataConfig()
|
|
|
|
cls.domain = get_domain(cls.apiclient)
|
|
cls.zone = get_zone(cls.apiclient, cls.testClient.getZoneForTests())
|
|
cls._cleanup = []
|
|
|
|
cls.hypervisor = cls.testClient.getHypervisorInfo()
|
|
cls.hypervisorNotSupported = False
|
|
if cls.hypervisor.lower() not in ["kvm", "vmware"]:
|
|
cls.hypervisorNotSupported = True
|
|
return
|
|
|
|
cls.services["service_offering"]["name"] = 'FSVM offering';
|
|
cls.services["service_offering"]["offerha"] = True;
|
|
cls.services["service_offering"]["cpunumber"] = 2;
|
|
cls.services["service_offering"]["cpuspeed"] = 500;
|
|
cls.services["service_offering"]["memory"] = 1024;
|
|
cls.service_offering = ServiceOffering.create(
|
|
cls.apiclient,
|
|
cls.services["service_offering"]
|
|
)
|
|
cls._cleanup.append(cls.service_offering)
|
|
|
|
cls.services["disk_offering"]["disksize"] = 1;
|
|
cls.disk_offering = DiskOffering.create(
|
|
cls.apiclient,
|
|
cls.services["disk_offering"],
|
|
custom=True
|
|
)
|
|
cls._cleanup.append(cls.disk_offering)
|
|
|
|
cls.useraccount = Account.create(
|
|
cls.apiclient,
|
|
cls.services["account"],
|
|
domainid=cls.domain.id
|
|
)
|
|
cls._cleanup.append(cls.useraccount)
|
|
|
|
cls.adminaccount = Account.create(
|
|
cls.apiclient,
|
|
cls.services["account"],
|
|
domainid=cls.domain.id,
|
|
admin=True
|
|
)
|
|
cls._cleanup.append(cls.adminaccount)
|
|
|
|
cls.network_offering_isolated = NetworkOffering.create(
|
|
cls.apiclient,
|
|
cls.services["isolated_network_offering"]
|
|
)
|
|
cls.network_offering_isolated.update(cls.apiclient, state='Enabled')
|
|
cls._cleanup.append(cls.network_offering_isolated)
|
|
cls.services["network"]["name"] = "Test Network Isolated"
|
|
cls.user_network = Network.create(
|
|
cls.apiclient,
|
|
cls.services["network"],
|
|
networkofferingid=cls.network_offering_isolated.id,
|
|
domainid=cls.domain.id,
|
|
accountid=cls.adminaccount.name,
|
|
zoneid=cls.zone.id
|
|
)
|
|
cls._cleanup.insert(0, cls.user_network)
|
|
cls.public_ipaddress = None
|
|
cls.sshpublicport = 1000
|
|
|
|
cls.template = get_template(
|
|
cls.apiclient,
|
|
cls.zone.id,
|
|
cls.services["ostype"])
|
|
if cls.template == FAILED:
|
|
assert False, "get_template() failed to return template with description %s" % cls.services["ostype"]
|
|
|
|
cls.services["domainid"] = cls.domain.id
|
|
cls.services["zoneid"] = cls.zone.id
|
|
cls.services["diskofferingid"] = cls.disk_offering.id
|
|
cls.services["serviceofferingid"] = cls.service_offering.id
|
|
cls.services["networkid"] = cls.user_network.id
|
|
cls.services["account"] = cls.adminaccount.name
|
|
|
|
cls.sharedfs = SharedFS.create(
|
|
cls.apiclient,
|
|
cls.services,
|
|
name='Test Shared FileSystem 1',
|
|
size=2,
|
|
filesystem='XFS'
|
|
)
|
|
cls._cleanup.insert(0, cls.sharedfs)
|
|
|
|
cls.virtual_machine1 = VirtualMachine.create(
|
|
cls.apiclient,
|
|
cls.services["virtual_machine"],
|
|
templateid=cls.template.id,
|
|
serviceofferingid=cls.service_offering.id,
|
|
networkids=cls.user_network.id,
|
|
domainid=cls.domain.id,
|
|
accountid=cls.adminaccount.name,
|
|
zoneid=cls.zone.id
|
|
)
|
|
cls._cleanup.insert(0, cls.virtual_machine1)
|
|
|
|
cls.public_ipaddress = cls.setUpSNAT(cls, cls.user_network)
|
|
cls.debug("Public ipaddress: " + cls.public_ipaddress.ipaddress)
|
|
port = cls.setUpPortForwarding(cls, cls.virtual_machine1.id)
|
|
cls.vm1_ssh_client = cls.getSSHClient(cls, cls.virtual_machine1, port)
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
try:
|
|
cleanup_resources(cls.apiclient, cls._cleanup)
|
|
except Exception as e:
|
|
raise Exception("Warning: Exception during cleanup : %s" % e)
|
|
|
|
def setUp(self):
|
|
self.apiclient = self.testClient.getApiClient()
|
|
self.dbclient = self.testClient.getDbConnection()
|
|
self.cleanup = []
|
|
return
|
|
|
|
def tearDown(self):
|
|
try:
|
|
self.debug("Cleaning up the resources")
|
|
cleanup_resources(self.apiclient, self.cleanup)
|
|
self.debug("Cleanup complete!")
|
|
except Exception as e:
|
|
self.debug("Warning! Exception in tearDown: %s" % e)
|
|
|
|
def setUpSNAT(self, network):
|
|
public_ipaddress = PublicIPAddress.list(
|
|
self.apiclient,
|
|
account=self.adminaccount.name,
|
|
domainid=self.domain.id,
|
|
associatednetworkid=network.id
|
|
)
|
|
createFwRule = createFirewallRule.createFirewallRuleCmd()
|
|
createFwRule.cidrlist = "0.0.0.0/0"
|
|
createFwRule.startport = 22
|
|
createFwRule.endport = 22
|
|
createFwRule.ipaddressid = public_ipaddress[0].id
|
|
createFwRule.protocol = "tcp"
|
|
self.apiclient.createFirewallRule(createFwRule)
|
|
return public_ipaddress[0]
|
|
|
|
def setUpPortForwarding(self, virtualmachineid):
|
|
createPfRule = createPortForwardingRule.createPortForwardingRuleCmd()
|
|
self.sshpublicport += 1
|
|
createPfRule.publicport = self.sshpublicport
|
|
createPfRule.privateport = 22
|
|
createPfRule.virtualmachineid = virtualmachineid
|
|
createPfRule.ipaddressid = self.public_ipaddress.id
|
|
createPfRule.protocol = "tcp"
|
|
self.apiclient.createPortForwardingRule(createPfRule)
|
|
self.debug("Successfully programmed PF rule for :%s"%self.public_ipaddress.ipaddress)
|
|
return createPfRule.publicport
|
|
|
|
def getSSHClient(self, virtual_machine, port):
|
|
try:
|
|
ssh_client = virtual_machine.get_ssh_client(ipaddress=self.public_ipaddress.ipaddress, port=port)
|
|
except Exception as e:
|
|
self.fail("SSH failed for virtual machine: %s - %s" % (virtual_machine.ipaddress, e))
|
|
return ssh_client
|
|
|
|
def mountSharedFSOnVM(self, ssh_client, sharedfs):
|
|
sharedfs_ip = sharedfs.nic[0].ipaddress
|
|
ssh_client.execute("mkdir /mnt/fs1")
|
|
cmd = "mount -t nfs -o nolock " + sharedfs_ip + ":/export /mnt/fs1"
|
|
ssh_client.execute(cmd)
|
|
|
|
@attr( tags=[ "advanced", "advancedns", "smokes"], required_hardware="true")
|
|
@skipTestIf("hypervisorNotSupported")
|
|
def test_mount_shared_fs(self):
|
|
"""Mount Shared FileSystem on two VMs and match contents
|
|
"""
|
|
self.mountSharedFSOnVM(self.vm1_ssh_client, self.sharedfs)
|
|
self.vm1_ssh_client.execute("df -Th /mnt/fs1")
|
|
self.vm1_ssh_client.execute("touch /mnt/fs1/test")
|
|
|
|
try:
|
|
self.virtual_machine2 = VirtualMachine.create(
|
|
self.apiclient,
|
|
self.services["virtual_machine"],
|
|
templateid=self.template.id,
|
|
serviceofferingid=self.service_offering.id,
|
|
networkids=self.user_network.id,
|
|
domainid=self.domain.id,
|
|
accountid=self.adminaccount.name,
|
|
zoneid=self.zone.id
|
|
)
|
|
except Exception as e:
|
|
self.vm1_ssh_client.execute("rm /mnt/fs1/test")
|
|
self.fail(e)
|
|
|
|
self.cleanup.append(self.virtual_machine2)
|
|
|
|
port = self.setUpPortForwarding(self.virtual_machine2.id)
|
|
ssh_client = self.getSSHClient(self.virtual_machine2, port)
|
|
self.assertIsNotNone(ssh_client)
|
|
|
|
self.mountSharedFSOnVM(ssh_client, self.sharedfs)
|
|
ssh_client.execute("df -Th /mnt/fs1")
|
|
result = ssh_client.execute("ls /mnt/fs1/test")
|
|
self.assertEqual(result[0], "/mnt/fs1/test")
|
|
|
|
@attr( tags=[ "advanced", "advancedns", "smokes"], required_hardware="true")
|
|
@skipTestIf("hypervisorNotSupported")
|
|
def test_resize_shared_fs(self):
|
|
"""Resize the shared filesystem by changing the disk offering and validate
|
|
"""
|
|
self.mountSharedFSOnVM(self.vm1_ssh_client, self.sharedfs)
|
|
result = self.vm1_ssh_client.execute("df -Th /mnt/fs1 | grep nfs")[0]
|
|
self.debug(result)
|
|
size = result.split()[-5]
|
|
self.debug("Size of the filesystem is " + size)
|
|
self.assertEqual(size, "2.0G", "SharedFS size should be 2.0G")
|
|
|
|
response = SharedFS.stop(self.sharedfs, self.apiclient)
|
|
response = SharedFS.changediskoffering(self.sharedfs, self.apiclient, self.disk_offering.id, 3)
|
|
self.debug(response)
|
|
response = SharedFS.start(self.sharedfs, self.apiclient)
|
|
time.sleep(10)
|
|
|
|
result = self.vm1_ssh_client.execute("df -Th /mnt/fs1 | grep nfs")[0]
|
|
size = result.split()[-5]
|
|
self.debug("Size of the filesystem is " + size)
|
|
self.assertEqual(size, "3.0G", "SharedFS size should be 3.0G")
|