Fix spelling (#6064)

* Fix spelling

- `interupted` to `interrupted`
- `paramter` to `parameter`

* Fix more typos
This commit is contained in:
John Bampton 2022-03-09 02:02:35 +10:00 committed by GitHub
parent 7facd49d33
commit 6401c850b7
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
58 changed files with 96 additions and 96 deletions

View File

@ -119,11 +119,11 @@ domr.scripts.dir=scripts/network/domr/kvm
# agent.hooks.libvirt_vm_xml_transformer.script=libvirt-vm-xml-transformer.groovy
# agent.hooks.libvirt_vm_xml_transformer.method=transform
#
# The hook is called right after libvirt successfuly launched VM
# The hook is called right after libvirt successfully launched VM
# agent.hooks.libvirt_vm_on_start.script=libvirt-vm-state-change.groovy
# agent.hooks.libvirt_vm_on_start.method=onStart
#
# The hook is called right after libvirt successfuly stopped VM
# The hook is called right after libvirt successfully stopped VM
# agent.hooks.libvirt_vm_on_stop.script=libvirt-vm-state-change.groovy
# agent.hooks.libvirt_vm_on_stop.method=onStop
#

View File

@ -513,7 +513,7 @@ public class AgentShell implements IAgentShell, Daemon {
while (!_exit)
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] AgentShell was interupted.");
s_logger.debug("[ignored] AgentShell was interrupted.");
}
} catch (final ConfigurationException e) {

View File

@ -607,7 +607,7 @@ public class DeployVMCmd extends BaseAsyncCreateCustomIdCmd implements SecurityG
public Map<Long, DiskOffering> getDataDiskTemplateToDiskOfferingMap() {
if (diskOfferingId != null && dataDiskTemplateToDiskOfferingList != null) {
throw new InvalidParameterValueException("diskofferingid paramter can't be specified along with datadisktemplatetodiskofferinglist parameter");
throw new InvalidParameterValueException("diskofferingid parameter can't be specified along with datadisktemplatetodiskofferinglist parameter");
}
if (MapUtils.isEmpty(dataDiskTemplateToDiskOfferingList)) {
return new HashMap<Long, DiskOffering>();

View File

@ -102,7 +102,7 @@ public class VmRulesetLogDaoImpl extends GenericDaoBase<VmRulesetLogVO, Long> im
try {
Thread.sleep(delayMs);
} catch (InterruptedException ie) {
s_logger.debug("[ignored] interupted while inserting security group rule log.");
s_logger.debug("[ignored] interrupted while inserting security group rule log.");
}
} else
s_logger.warn("Caught another deadlock exception while retrying inserting security group rule log, giving up");

View File

@ -275,7 +275,7 @@ public class StorageCacheManagerImpl implements StorageCacheManager, Manager {
try {
lock.wait(miliSeconds);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting for cache copy completion.");
s_logger.debug("[ignored] interrupted while waiting for cache copy completion.");
}
s_logger.debug("waken up");

View File

@ -171,7 +171,7 @@ public class ClusterServiceServletContainer {
try {
Thread.sleep(1000);
} catch (InterruptedException e1) {
s_logger.debug("[ignored] interupted while waiting to retry running the servlet container.");
s_logger.debug("[ignored] interrupted while waiting to retry running the servlet container.");
}
}
}

View File

@ -144,7 +144,7 @@ public class Merovingian2 extends StandardMBean implements MerovingianMBean {
}
Thread.sleep(5000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while aquiring " + key);
s_logger.debug("[ignored] interrupted while aquiring " + key);
}
}
String msg = "Timed out on acquiring lock " + key + " . Waited for " + ((InaccurateClock.getTime() - startTime)/1000) + "seconds";

View File

@ -54,7 +54,7 @@ public class GlobalLockTest {
Thread.sleep(jobDuration * 1000);
}
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while testing global lock.");
s_logger.debug("[ignored] interrupted while testing global lock.");
} finally {
if (locked) {
boolean unlocked = WorkLock.unlock();

View File

@ -326,7 +326,7 @@ public class MessageBusBase implements MessageBus {
try {
wait();
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while guarding re-entrance on message bus.");
s_logger.debug("[ignored] interrupted while guarding re-entrance on message bus.");
}
} else {
break;

View File

@ -41,7 +41,7 @@ public class MessageDetector implements MessageSubscriber {
try {
wait(timeoutInMiliseconds);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting on any message.");
s_logger.debug("[ignored] interrupted while waiting on any message.");
}
}
}

View File

@ -84,7 +84,7 @@ public class StaticStrategy implements BalanceStrategy {
try {
Thread.sleep(250);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while fail over in progres.");
s_logger.debug("[ignored] interrupted while fail over in progres.");
}
// start fresh

View File

@ -303,7 +303,7 @@ public class BareMetalResourceBase extends ManagerBase implements ServerResource
try {
TimeUnit.SECONDS.sleep(1);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting to retry running script.");
s_logger.debug("[ignored] interrupted while waiting to retry running script.");
}
continue;
} else if (res == null) {

View File

@ -2371,7 +2371,7 @@ public class HypervDirectConnectResource extends ServerResourceBase implements S
try {
Thread.sleep(5000);
} catch (final InterruptedException ex) {
s_logger.debug("[ignored] interupted while waiting to retry connecting to vm after exception: "+e.getLocalizedMessage());
s_logger.debug("[ignored] interrupted while waiting to retry connecting to vm after exception: "+e.getLocalizedMessage());
}
}
}
@ -2379,7 +2379,7 @@ public class HypervDirectConnectResource extends ServerResourceBase implements S
try {
Thread.sleep(1000);
} catch (final InterruptedException ex) {
s_logger.debug("[ignored] interupted while connecting to vm.");
s_logger.debug("[ignored] interrupted while connecting to vm.");
}
}

View File

@ -104,7 +104,7 @@ public final class LibvirtBackupSnapshotCommandWrapper extends CommandWrapper<Ba
r.confSet("key", primaryPool.getAuthSecret());
r.confSet("client_mount_timeout", "30");
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet("mon_host"));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet("mon_host"));
final IoCTX io = r.ioCtxCreate(primaryPool.getSourceDir());
final Rbd rbd = new Rbd(io);

View File

@ -121,7 +121,7 @@ public final class LibvirtManageSnapshotCommandWrapper extends CommandWrapper<Ma
r.confSet("key", primaryPool.getAuthSecret());
r.confSet("client_mount_timeout", "30");
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet("mon_host"));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet("mon_host"));
final IoCTX io = r.ioCtxCreate(primaryPool.getSourceDir());
final Rbd rbd = new Rbd(io);

View File

@ -159,7 +159,7 @@ public class IscsiStorageCleanupMonitor implements Runnable{
try {
Thread.sleep(CLEANUP_INTERVAL_SEC * 1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted between heartbeats.");
s_logger.debug("[ignored] interrupted between heartbeats.");
}
Thread monitorThread = new Thread(new Monitor());
@ -167,7 +167,7 @@ public class IscsiStorageCleanupMonitor implements Runnable{
try {
monitorThread.join();
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted joining monitor.");
s_logger.debug("[ignored] interrupted joining monitor.");
}
}
}

View File

@ -325,7 +325,7 @@ public class KVMStoragePoolManager {
try {
Thread.sleep(3000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while trying to get storage pool.");
s_logger.debug("[ignored] interrupted while trying to get storage pool.");
}
cnt++;
}

View File

@ -1604,7 +1604,7 @@ public class KVMStorageProcessor implements StorageProcessor {
r.confSet(CEPH_AUTH_KEY, primaryPool.getAuthSecret());
r.confSet(CEPH_CLIENT_MOUNT_TIMEOUT, CEPH_DEFAULT_MOUNT_TIMEOUT);
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet(CEPH_MON_HOST));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet(CEPH_MON_HOST));
return r;
}

View File

@ -509,7 +509,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
pool.setUsed(storage.getInfo().allocation);
pool.setAvailable(storage.getInfo().available);
s_logger.debug("Succesfully refreshed pool " + uuid +
s_logger.debug("Successfully refreshed pool " + uuid +
" Capacity: " + toHumanReadableSize(storage.getInfo().capacity) +
" Used: " + toHumanReadableSize(storage.getInfo().allocation) +
" Available: " + toHumanReadableSize(storage.getInfo().available));
@ -721,7 +721,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
s.free();
}
s_logger.info("Storage pool " + uuid + " was succesfully removed from libvirt.");
s_logger.info("Storage pool " + uuid + " was successfully removed from libvirt.");
return true;
} catch (LibvirtException e) {
@ -908,7 +908,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
r.confSet("key", pool.getAuthSecret());
r.confSet("client_mount_timeout", "30");
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet("mon_host"));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet("mon_host"));
IoCTX io = r.ioCtxCreate(pool.getSourceDir());
Rbd rbd = new Rbd(io);
@ -926,7 +926,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
s_logger.debug("Removing snapshot " + pool.getSourceDir() + "/" + uuid + "@" + snap.name);
image.snapRemove(snap.name);
}
s_logger.info("Succesfully unprotected and removed any remaining snapshots (" + snaps.size() + ") of "
s_logger.info("Successfully unprotected and removed any remaining snapshots (" + snaps.size() + ") of "
+ pool.getSourceDir() + "/" + uuid + " Continuing to remove the RBD image");
} catch (RbdException e) {
s_logger.error("Failed to remove snapshot with exception: " + e.toString() +
@ -1105,7 +1105,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
r.confSet("key", srcPool.getAuthSecret());
r.confSet("client_mount_timeout", "30");
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet("mon_host"));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet("mon_host"));
IoCTX io = r.ioCtxCreate(srcPool.getSourceDir());
Rbd rbd = new Rbd(io);
@ -1154,7 +1154,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
}
rbd.clone(template.getName(), rbdTemplateSnapName, io, disk.getName(), RBD_FEATURES, rbdOrder);
s_logger.debug("Succesfully cloned " + template.getName() + "@" + rbdTemplateSnapName + " to " + disk.getName());
s_logger.debug("Successfully cloned " + template.getName() + "@" + rbdTemplateSnapName + " to " + disk.getName());
/* We also need to resize the image if the VM was deployed with a larger root disk size */
if (disk.getVirtualSize() > template.getVirtualSize()) {
RbdImage diskImage = rbd.open(disk.getName());
@ -1176,14 +1176,14 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
rSrc.confSet("key", srcPool.getAuthSecret());
rSrc.confSet("client_mount_timeout", "30");
rSrc.connect();
s_logger.debug("Succesfully connected to source Ceph cluster at " + rSrc.confGet("mon_host"));
s_logger.debug("Successfully connected to source Ceph cluster at " + rSrc.confGet("mon_host"));
Rados rDest = new Rados(destPool.getAuthUserName());
rDest.confSet("mon_host", destPool.getSourceHost() + ":" + destPool.getSourcePort());
rDest.confSet("key", destPool.getAuthSecret());
rDest.confSet("client_mount_timeout", "30");
rDest.connect();
s_logger.debug("Succesfully connected to source Ceph cluster at " + rDest.confGet("mon_host"));
s_logger.debug("Successfully connected to source Ceph cluster at " + rDest.confGet("mon_host"));
IoCTX sIO = rSrc.ioCtxCreate(srcPool.getSourceDir());
Rbd sRbd = new Rbd(sIO);
@ -1343,7 +1343,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
s_logger.debug("Starting copy from source image " + srcFile.getFileName() + " to RBD image " + rbdDestPath);
qemu.convert(srcFile, destFile);
s_logger.debug("Succesfully converted source image " + srcFile.getFileName() + " to RBD image " + rbdDestPath);
s_logger.debug("Successfully converted source image " + srcFile.getFileName() + " to RBD image " + rbdDestPath);
/* We have to stat the RBD image to see how big it became afterwards */
Rados r = new Rados(destPool.getAuthUserName());
@ -1351,7 +1351,7 @@ public class LibvirtStorageAdaptor implements StorageAdaptor {
r.confSet("key", destPool.getAuthSecret());
r.confSet("client_mount_timeout", "30");
r.connect();
s_logger.debug("Succesfully connected to Ceph cluster at " + r.confGet("mon_host"));
s_logger.debug("Successfully connected to Ceph cluster at " + r.confGet("mon_host"));
IoCTX io = r.ioCtxCreate(destPool.getSourceDir());
Rbd rbd = new Rbd(io);

View File

@ -271,7 +271,7 @@ public class ScaleIOStorageAdaptor implements StorageAdaptor {
LOGGER.debug("Starting copy from source disk image " + srcFile.getFileName() + " to PowerFlex volume: " + destDisk.getPath());
qemu.convert(srcFile, destFile, true);
LOGGER.debug("Succesfully converted source disk image " + srcFile.getFileName() + " to PowerFlex volume: " + destDisk.getPath());
LOGGER.debug("Successfully converted source disk image " + srcFile.getFileName() + " to PowerFlex volume: " + destDisk.getPath());
} catch (QemuImgException | LibvirtException e) {
try {
Map<String, String> srcInfo = qemu.info(srcFile);
@ -373,7 +373,7 @@ public class ScaleIOStorageAdaptor implements StorageAdaptor {
LOGGER.debug("Starting copy from source downloaded template " + srcFile.getFileName() + " to PowerFlex template volume: " + destDisk.getPath());
QemuImg qemu = new QemuImg(timeout);
qemu.convert(srcFile, destFile);
LOGGER.debug("Succesfully converted source downloaded template " + srcFile.getFileName() + " to PowerFlex template volume: " + destDisk.getPath());
LOGGER.debug("Successfully converted source downloaded template " + srcFile.getFileName() + " to PowerFlex template volume: " + destDisk.getPath());
} catch (QemuImgException | LibvirtException e) {
LOGGER.error("Failed to convert from " + srcFile.getFileName() + " to " + destFile.getFileName() + " the error was: " + e.getMessage(), e);
destDisk = null;

View File

@ -1158,7 +1158,7 @@ public class VmwareResource implements StoragePoolResource, ServerResource, Vmwa
Thread.currentThread();
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while trying to get mac.");
s_logger.debug("[ignored] interrupted while trying to get mac.");
}
}
@ -6496,7 +6496,7 @@ public class VmwareResource implements StoragePoolResource, ServerResource, Vmwa
try {
Thread.sleep(5000);
} catch (InterruptedException ex) {
s_logger.debug("[ignored] interupted while waiting to retry connect after failure.", e);
s_logger.debug("[ignored] interrupted while waiting to retry connect after failure.", e);
}
}
}
@ -6504,7 +6504,7 @@ public class VmwareResource implements StoragePoolResource, ServerResource, Vmwa
try {
Thread.sleep(1000);
} catch (InterruptedException ex) {
s_logger.debug("[ignored] interupted while waiting to retry connect.");
s_logger.debug("[ignored] interrupted while waiting to retry connect.");
}
}

View File

@ -853,7 +853,7 @@ public class XenServerStorageProcessor implements StorageProcessor {
VDI.Record vdir;
vdir = vdi.getRecord(conn);
s_logger.debug("Succesfully created VDI: Uuid = " + vdir.uuid);
s_logger.debug("Successfully created VDI: Uuid = " + vdir.uuid);
final VolumeObjectTO newVol = new VolumeObjectTO();
newVol.setName(vdir.nameLabel);

View File

@ -72,7 +72,7 @@ public final class CitrixCreateCommandWrapper extends CommandWrapper<CreateComma
VDI.Record vdir;
vdir = vdi.getRecord(conn);
s_logger.debug("Succesfully created VDI for " + command + ". Uuid = " + vdir.uuid);
s_logger.debug("Successfully created VDI for " + command + ". Uuid = " + vdir.uuid);
final VolumeTO vol =
new VolumeTO(command.getVolumeId(), dskch.getType(), pool.getType(), pool.getUuid(), vdir.nameLabel, pool.getPath(), vdir.uuid, vdir.virtualSize, null);

View File

@ -961,7 +961,7 @@ public class NetscalerResource implements ServerResource {
try {
Thread.sleep(10000);
} catch (final InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting for netscaler to be 'up'.");
s_logger.debug("[ignored] interrupted while waiting for netscaler to be 'up'.");
}
final ns refreshNsObj = new ns();
refreshNsObj.set_id(newVpx.get_id());

View File

@ -964,7 +964,7 @@ public class ElastistorUtil {
public List<Pair<String, String>> getCommandParameters();
/*
* Adds new key-value pair to the query paramters lists.
* Adds new key-value pair to the query parameters lists.
*/
public void putCommandParameter(String key, String value);

View File

@ -206,15 +206,15 @@ done
#install_cloud_agent $dflag
#install_cloud_consoleP $dflag
paramters=
parameters=
if [ -n "$pubNic" ]
then
paramters=" --pubNic=$pubNic"
parameters=" --pubNic=$pubNic"
fi
if [ -n "$prvNic" ]
then
paramters=" --prvNic=$prvNic $paramters"
parameters=" --prvNic=$prvNic $parameters"
fi
selenabled=`cat /selinux/enforce`
@ -224,5 +224,5 @@ then
setenforce 0
fi
cloudstack-setup-agent --host=$host --zone=$zone --pod=$pod --cluster=$cluster --guid=$guid $paramters -a > /dev/null
cloudstack-setup-agent --host=$host --zone=$zone --pod=$pod --cluster=$cluster --guid=$guid $parameters -a > /dev/null
#cloud_consoleP_setup $host $zone $pod

View File

@ -1250,7 +1250,7 @@ public class ApiDBUtils {
// Currently, KVM only supports RBD and PowerFlex images of type RAW.
// This results in a weird collision with OVM volumes which
// can only be raw, thus making KVM RBD volumes show up as OVM
// rather than RBD. This block of code can (hopefuly) by checking to
// rather than RBD. This block of code can (hopefully) by checking to
// see if the pool is using either RBD or NFS. However, it isn't
// quite clear what to do if both storage types are used. If the image
// format is RAW, it narrows the hypervisor choice down to OVM and KVM / RBD or KVM / CLVM

View File

@ -239,7 +239,7 @@ public class NetworkMigrationManagerImpl implements NetworkMigrationManager {
assignRouterNicsToNewNetwork(network.getId(), networkCopyId);
if (s_logger.isDebugEnabled()) {
s_logger.debug("Succesfully created a copy of network " + originalNetwork.getName() + "(" + originalNetwork.getUuid() + ") id is " + originalNetwork.getId() + " for migration. The network copy has uuid " + network.getUuid() + " and id " + network.getId());
s_logger.debug("Successfully created a copy of network " + originalNetwork.getName() + "(" + originalNetwork.getUuid() + ") id is " + originalNetwork.getId() + " for migration. The network copy has uuid " + network.getUuid() + " and id " + network.getId());
}
return networkCopyId;
}
@ -311,7 +311,7 @@ public class NetworkMigrationManagerImpl implements NetworkMigrationManager {
reassignGatewayToNewVpc(vpcId, copyOfVpcId);
copyVpcResourceTagsToNewVpc(vpcId, copyOfVpcId);
if (s_logger.isDebugEnabled()) {
s_logger.debug("Succesfully created a copy of network " + vpc.getName() + "(" + vpc.getUuid() + ") id is " + vpc.getId() + " for migration. The network copy has uuid " + copyVpcVO.getUuid() + " and id " + copyOfVpc.getId());
s_logger.debug("Successfully created a copy of network " + vpc.getName() + "(" + vpc.getUuid() + ") id is " + vpc.getId() + " for migration. The network copy has uuid " + copyVpcVO.getUuid() + " and id " + copyOfVpc.getId());
}
} catch (ResourceAllocationException e) {
throw new CloudRuntimeException(e.getMessage());

View File

@ -175,7 +175,7 @@ public class VpcVirtualNetworkApplianceManagerImpl extends VirtualNetworkApplian
s_logger.warn("Failed to remove the router " + router + " from network " + network + " as a part of cleanup");
}
} else {
s_logger.debug("Succesfully added router " + router + " to guest network " + network);
s_logger.debug("Successfully added router " + router + " to guest network " + network);
}
}

View File

@ -1528,7 +1528,7 @@ public class RulesManagerImpl extends ManagerBase implements RulesManager, Rules
_ipAddrMgr.handleSystemIpRelease(ip);
throw new CloudRuntimeException("Failed to enable static nat on system ip for the vm " + vm);
} else {
s_logger.warn("Succesfully enabled static nat on system ip " + ip + " for the vm " + vm);
s_logger.warn("Successfully enabled static nat on system ip " + ip + " for the vm " + vm);
}
}
}

View File

@ -1141,7 +1141,7 @@ public class VpcManagerImpl extends ManagerBase implements VpcManager, VpcProvis
// update the instance with removed flag only when the cleanup is
// executed successfully
if (_vpcDao.remove(vpc.getId())) {
s_logger.debug("Vpc " + vpc + " is destroyed succesfully");
s_logger.debug("Vpc " + vpc + " is destroyed successfully");
return true;
} else {
s_logger.warn("Vpc " + vpc + " failed to destroy");
@ -1433,7 +1433,7 @@ public class VpcManagerImpl extends ManagerBase implements VpcManager, VpcProvis
for (final VpcProvider element : getVpcElements()) {
if (providersToImplement.contains(element.getProvider())) {
if (element.shutdownVpc(vpc, context)) {
s_logger.debug("Vpc " + vpc + " has been shutdown succesfully");
s_logger.debug("Vpc " + vpc + " has been shutdown successfully");
} else {
s_logger.warn("Vpc " + vpc + " failed to shutdown");
success = false;
@ -1961,7 +1961,7 @@ public class VpcManagerImpl extends ManagerBase implements VpcManager, VpcProvis
}
}
if (success) {
s_logger.debug("Private gateway " + gateway + " was applied succesfully on the backend");
s_logger.debug("Private gateway " + gateway + " was applied successfully on the backend");
if (vo.getState() != VpcGateway.State.Ready) {
vo.setState(VpcGateway.State.Ready);
_vpcGatewayDao.update(vo.getId(), vo);
@ -2029,7 +2029,7 @@ public class VpcManagerImpl extends ManagerBase implements VpcManager, VpcProvis
for (final VpcProvider provider : getVpcElements()) {
if (providersToImplement.contains(provider.getProvider())) {
if (provider.deletePrivateGateway(gateway)) {
s_logger.debug("Private gateway " + gateway + " was applied succesfully on the backend");
s_logger.debug("Private gateway " + gateway + " was applied successfully on the backend");
} else {
s_logger.warn("Private gateway " + gateway + " failed to apply on the backend");
gatewayVO.setState(VpcGateway.State.Ready);

View File

@ -256,7 +256,7 @@ public class SnapshotManagerImpl extends MutualExclusiveIdsManagerBase implement
try {
Thread.sleep(_pauseInterval * 1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while retry cmd.");
s_logger.debug("[ignored] interrupted while retry cmd.");
}
s_logger.debug("Retrying...");

View File

@ -61,7 +61,7 @@ public class SystemVmLoadScanner<T> {
try {
_capacityScanScheduler.awaitTermination(1000, TimeUnit.MILLISECONDS);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while stopping systemvm load scanner.");
s_logger.debug("[ignored] interrupted while stopping systemvm load scanner.");
}
_capacityScanLock.releaseRef();

View File

@ -1127,7 +1127,7 @@ public class UnmanagedVMsManagerImpl implements UnmanagedVMsManager {
if (template == null) {
template = createDefaultDummyVmImportTemplate();
if (template == null) {
throw new InvalidParameterValueException(String.format("Default VM import template with unique name: %s for hypervisor: %s cannot be created. Please use templateid paramter for import", VM_IMPORT_DEFAULT_TEMPLATE_NAME, cluster.getHypervisorType().toString()));
throw new InvalidParameterValueException(String.format("Default VM import template with unique name: %s for hypervisor: %s cannot be created. Please use templateid parameter for import", VM_IMPORT_DEFAULT_TEMPLATE_NAME, cluster.getHypervisorType().toString()));
}
}
} else {

View File

@ -264,7 +264,7 @@ public class SyncLink implements Link {
try {
Thread.sleep(100);
} catch (InterruptedException e) {
s_logger.info("[ignored] interupted during pull", e);
s_logger.info("[ignored] interrupted during pull", e);
}
}

View File

@ -69,7 +69,7 @@ public class FakeSource extends BaseElement {
try {
Thread.sleep(delay);
} catch (InterruptedException e) {
s_logger.info("[ignored] interupted while creating latency", e);
s_logger.info("[ignored] interrupted while creating latency", e);
}
}

View File

@ -230,7 +230,7 @@ public abstract class ConsoleProxyClientBase implements ConsoleProxyClient, Cons
try {
Thread.sleep(100);
} catch (InterruptedException e) {
s_logger.debug("[ignored] Console proxy was interupted while waiting for viewer to become ready.");
s_logger.debug("[ignored] Console proxy was interrupted while waiting for viewer to become ready.");
}
}
return false;
@ -343,7 +343,7 @@ public abstract class ConsoleProxyClientBase implements ConsoleProxyClient, Cons
try {
tileDirtyEvent.wait(3000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] Console proxy ajax update was interupted while waiting for viewer to become ready.");
s_logger.debug("[ignored] Console proxy ajax update was interrupted while waiting for viewer to become ready.");
}
}
}

View File

@ -111,7 +111,7 @@ public class ConsoleProxyGCThread extends Thread {
try {
Thread.sleep(5000);
} catch (InterruptedException ex) {
s_logger.debug("[ignored] Console proxy was interupted during GC.");
s_logger.debug("[ignored] Console proxy was interrupted during GC.");
}
}
}

View File

@ -494,7 +494,7 @@ class TestSystemVmLocalStorage(cloudstackTestCase):
name = 'secstorage.service.offering'
else:
raise Exception(
"type paramter is not correct it should be system vm "
"type parameter is not correct it should be system vm "
"type{console proxy,secsroragevm}")
# 3-update global settings with system offering uuid

View File

@ -562,7 +562,7 @@ class TestSnapshotList(cloudstackTestCase):
True,
"Account access check failed!!")
## Domain Admin - Test cases without passing listall paramter
## Domain Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listSnapshot_as_domainadmin(self):
@ -1308,7 +1308,7 @@ class TestSnapshotList(cloudstackTestCase):
True,
"Account access check failed!!")
## ROOT Admin - Test cases without passing listall paramter
## ROOT Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listSnapshot_as_rootadmin(self):
@ -2022,7 +2022,7 @@ class TestSnapshotList(cloudstackTestCase):
True,
"Account access check failed!!")
## Regular User - Test cases without passing listall paramter
## Regular User - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listSnapshot_as_regularuser(self):

View File

@ -535,7 +535,7 @@ class TestVMList(cloudstackTestCase):
True,
"Account access check failed!!")
## Domain Admin - Test cases without passing listall paramter
## Domain Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVM_as_domainadmin(self):
@ -1281,7 +1281,7 @@ class TestVMList(cloudstackTestCase):
True,
"Account access check failed!!")
## ROOT Admin - Test cases without passing listall paramter
## ROOT Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVM_as_rootadmin(self):
@ -1996,7 +1996,7 @@ class TestVMList(cloudstackTestCase):
True,
"Account access check failed!!")
## Regular User - Test cases without passing listall paramter
## Regular User - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVM_as_regularuser(self):

View File

@ -548,7 +548,7 @@ class TestVolumeList(cloudstackTestCase):
True,
"Account access check failed!!")
## Domain Admin - Test cases without passing listall paramter
## Domain Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVolume_as_domainadmin(self):
@ -1290,7 +1290,7 @@ class TestVolumeList(cloudstackTestCase):
True,
"Account access check failed!!")
## ROOT Admin - Test cases without passing listall paramter
## ROOT Admin - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVolume_as_rootadmin(self):
@ -2004,7 +2004,7 @@ class TestVolumeList(cloudstackTestCase):
True,
"Account access check failed!!")
## Regular User - Test cases without passing listall paramter
## Regular User - Test cases without passing listall parameter
@attr("simulator_only", tags=["advanced"], required_hardware="false")
def test_listVolume_as_regularuser(self):

View File

@ -940,7 +940,7 @@ class TestTemplates(cloudstackTestCase):
# 2. Perform snapshot on the root disk of this VM.
# 3. Create a template from snapshot.
# 4. Delete the template and create a new template with same name
# 5. Template should be created succesfully
# 5. Template should be created successfully
# Create a snapshot from the ROOTDISK
if self.hypervisor.lower() in ['hyperv']:

View File

@ -858,7 +858,7 @@ class TestHAProxyStickyness(cloudstackTestCase):
# * by passing the Invlaid parameter
# * Invalid method name
# * required parameter not present
# * passing invalid values to valid paramters.
# * passing invalid values to valid parameters.
self.debug("Creating LB rule for account: %s" % self.account.name)
lb_rule = self.create_LB_Rule(

View File

@ -851,7 +851,7 @@ class TestResizeVolume(cloudstackTestCase):
except Exception as e:
raise Exception("Warning: Exception during"
" VM deployment with new"
" rootdisk paramter : %s" % e)
" rootdisk parameter : %s" % e)
@attr(tags=["advanced"], required_hardware="true")

View File

@ -604,7 +604,7 @@ class TestVPCNetworkPFRules(cloudstackTestCase):
# 7. Use the Create PF rule ssh and http for vm1 and vm2 in network1.
# 8. Use the Create PF rule ssh and http for vm3 and vm4 in network2.
# 9. Successfully ssh into the Guest vm1, vm2, vm3 and vm4 using the PF rule.
# 10. Succesfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 10. Successfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 11. Stop VPC Virtual Router.
# 12. Delete all PF rultes for vm1, vm2, vm3 and vm4.
# 12. Start VPC Virtual Router.
@ -684,7 +684,7 @@ class TestVPCNetworkPFRules(cloudstackTestCase):
# 7. Use the Create PF rule ssh and http for vm1 and vm2 in network1.
# 8. Use the Create PF rule ssh and http for vm3 and vm4 in network2.
# 9. Successfully ssh into the Guest vm1, vm2, vm3 and vm4 using the PF rule.
# 10. Succesfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 10. Successfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 12. Delete all PF rultes for vm1, vm2, vm3 and vm4.
# 13. Fail to ssh and http to vm1, vm2, vm3 and vm4.
"""

View File

@ -549,7 +549,7 @@ class TestVPCNetworkPFRules(cloudstackTestCase):
# 7. Use the Create PF rule ssh and http for vm1 and vm2 in network1.
# 8. Use the Create PF rule ssh and http for vm3 and vm4 in network2.
# 9. Successfully ssh into the Guest vm1, vm2, vm3 and vm4 using the PF rule.
# 10. Succesfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 10. Successfully wget a file from http server present on vm1, vm2, vm3 and vm4.
# 12. Delete all PF rultes for vm1, vm2, vm3 and vm4.
# 13. Fail to ssh and http to vm1, vm2, vm3 and vm4.
"""

View File

@ -1542,7 +1542,7 @@ class TestUserLogin(cloudstackTestCase):
parentdomainid=self.domain.id
)
self.cleanup.append(domain)
self.debug("Domain: %s is created succesfully." % domain.name)
self.debug("Domain: %s is created successfully." % domain.name)
self.debug(
"Checking if the created domain is listed in list domains API")
domains = Domain.list(self.apiclient, id=domain.id, listall=True)
@ -1818,7 +1818,7 @@ class TestDomainForceRemove(cloudstackTestCase):
parentdomainid=self.domain.id
)
self.cleanup.append(self.child_domain)
self.debug("Domain is created succesfully.")
self.debug("Domain is created successfully.")
self.debug(
"Checking if the created domain is listed in list domains API")
domains = Domain.list(self.apiclient, id=self.child_domain.id, listall=True)

View File

@ -448,7 +448,7 @@ class TestSnapshotsHardning(cloudstackTestCase):
"""snapshot hardning
1. Take VM snapshot then migrate the VM to another host
and again take volume snapshot and check its integrity
2. Verify that snapshot gets created successfuly while VM
2. Verify that snapshot gets created successfully while VM
is getting Migrated to another host
3. Verify that snapshot should succeed after vm's are HA-ed
to different host and also check its integrity
@ -860,7 +860,7 @@ class TestSnapshotsHardning(cloudstackTestCase):
1. Take VM snapshot then migrate the VM to another
host and again take
volume snapshot and check its intigrity
2. Verify that snapshot gets created successfuly
2. Verify that snapshot gets created successfully
while VM is getting
Migrated to another host
3. Verify that snapshot should succeed after vm's are

View File

@ -471,7 +471,7 @@ class TestPathVolume(cloudstackTestCase):
self.cluster_volume.id), "volume does not exist %s" %
self.cluster_volume.id)
self.debug(
"volume id %s got created successfuly" %
"volume id %s got created successfully" %
list_cluster_volume[0].id)
self.virtual_machine_zone_4.attach_volume(self.userapiclient,
self.cluster_volume
@ -634,7 +634,7 @@ class TestPathVolume(cloudstackTestCase):
self.template_from_vm1_root_disk.delete(self.userapiclient
)
self.debug(
"Template id: %s got deleted successfuly" %
"Template id: %s got deleted successfully" %
self.template_from_vm1_root_disk.id)
except Exception as e:
raise Exception("Template deletion failed with error %s" % e)
@ -690,7 +690,7 @@ class TestPathVolume(cloudstackTestCase):
list_volume[0].id, str(
self.migrate_volume.id), "volume does not exist %s" %
self.migrate_volume.id)
self.debug("volume id %s got created successfuly" % list_volume[0].id)
self.debug("volume id %s got created successfully" % list_volume[0].id)
self.virtual_machine_1.attach_volume(self.userapiclient,
self.migrate_volume
@ -720,7 +720,7 @@ class TestPathVolume(cloudstackTestCase):
list_volume[0].id, str(
self.migrate_volume.id), "volume does not exist %s" %
self.migrate_volume.id)
self.debug("volume id %s got created successfuly" % list_volume[0].id)
self.debug("volume id %s got created successfully" % list_volume[0].id)
list_pool = StoragePool.list(self.apiclient,
id=list_volume[0].storageid
)

View File

@ -802,7 +802,7 @@ public class ApiCommand {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while during async job result query.");
s_logger.debug("[ignored] interrupted while during async job result query.");
}
} else {
break;

View File

@ -1337,7 +1337,7 @@ public class StressTestDirectAttach {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while during async job result query.");
s_logger.debug("[ignored] interrupted while during async job result query.");
}
} else {
break;

View File

@ -2273,7 +2273,7 @@ public class TestClientWithAPI {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while during async job result query.");
s_logger.debug("[ignored] interrupted while during async job result query.");
}
} else {
break;

View File

@ -67,7 +67,7 @@ public class ConsoleProxy implements Runnable {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted.");
s_logger.debug("[ignored] interrupted.");
}
}

View File

@ -3339,7 +3339,7 @@ function Browser(window, document, $log, $sniffer) {
* Cancels a defered task identified with `deferId`.
*
* @param {*} deferId Token returned by the `$browser.defer` function.
* @returns {boolean} Returns `true` if the task hasn't executed yet and was successfuly canceled.
* @returns {boolean} Returns `true` if the task hasn't executed yet and was successfully canceled.
*/
self.defer.cancel = function(deferId) {
if (pendingDeferIds[deferId]) {

View File

@ -1759,7 +1759,7 @@ public class HypervisorHostHelper {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting to config vm.");
s_logger.debug("[ignored] interrupted while waiting to config vm.");
}
}
}

View File

@ -275,7 +275,7 @@ public class VirtualMachineMO extends BaseMO {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while dealing with vm questions.");
s_logger.debug("[ignored] interrupted while dealing with vm questions.");
}
}
s_logger.info("VM Question monitor stopped");
@ -544,7 +544,7 @@ public class VirtualMachineMO extends BaseMO {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while waiting for snapshot to be done.");
s_logger.debug("[ignored] interrupted while waiting for snapshot to be done.");
}
}
@ -1651,7 +1651,7 @@ public class VirtualMachineMO extends BaseMO {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while handling vm question about iso detach.");
s_logger.debug("[ignored] interrupted while handling vm question about iso detach.");
}
}
s_logger.info("VM Question monitor stopped");
@ -3367,7 +3367,7 @@ public class VirtualMachineMO extends BaseMO {
try {
Thread.sleep(1000);
} catch (InterruptedException e) {
s_logger.debug("[ignored] interupted while handling vm question about umount tools install.");
s_logger.debug("[ignored] interrupted while handling vm question about umount tools install.");
}
}

View File

@ -681,7 +681,7 @@ public class VmwareContext {
try {
Thread.sleep(CONNECT_RETRY_INTERVAL);
} catch (InterruptedException ex) {
s_logger.debug("[ignored] interupted while connecting.");
s_logger.debug("[ignored] interrupted while connecting.");
}
}
}