mirror of
https://github.com/apache/cloudstack.git
synced 2025-10-26 01:32:18 +02:00
Remove unneeded duplicate words (#7850)
This commit is contained in:
parent
45616aaf61
commit
4eb110af73
@ -106,7 +106,7 @@ public interface ManagementService {
|
||||
Pair<List<? extends Configuration>, Integer> searchForConfigurations(ListCfgsByCmd c);
|
||||
|
||||
/**
|
||||
* returns the the configuration groups
|
||||
* returns the configuration groups
|
||||
*
|
||||
* @return list of configuration groups
|
||||
*/
|
||||
|
||||
@ -51,7 +51,7 @@ public class ListNicsCmd extends BaseListCmd {
|
||||
//////////////// API parameters /////////////////////
|
||||
/////////////////////////////////////////////////////
|
||||
|
||||
@Parameter(name = ApiConstants.NIC_ID, type = CommandType.UUID, entityType = NicResponse.class, required = false, description = "the ID of the nic to to list IPs")
|
||||
@Parameter(name = ApiConstants.NIC_ID, type = CommandType.UUID, entityType = NicResponse.class, required = false, description = "the ID of the nic to list IPs")
|
||||
private Long nicId;
|
||||
|
||||
@Parameter(name = ApiConstants.VIRTUAL_MACHINE_ID, type = CommandType.UUID, entityType = UserVmResponse.class, required = true, description = "the ID of the vm")
|
||||
|
||||
@ -488,7 +488,7 @@ public class OVFHelperTest {
|
||||
" </Property>\n" +
|
||||
" <Property ovf:key=\"guestinfo.cis.db.servername\" ovf:type=\"string\" ovf:userConfigurable=\"false\" ovf:value=\"\">\n" +
|
||||
" <Label>Database Server</Label>\n" +
|
||||
" <Description>String naming the the hostname of the server on which the external database is running (ignored when db.type is 'embedded').</Description>\n" +
|
||||
" <Description>String naming the hostname of the server on which the external database is running (ignored when db.type is 'embedded').</Description>\n" +
|
||||
" </Property>\n" +
|
||||
" <Property ovf:key=\"guestinfo.cis.db.serverport\" ovf:type=\"string\" ovf:userConfigurable=\"false\" ovf:value=\"\">\n" +
|
||||
" <Label>Database Port</Label>\n" +
|
||||
|
||||
@ -240,7 +240,7 @@ public class UsageEventUtils {
|
||||
try {
|
||||
s_eventBus.publish(event);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to publish usage event on the the event bus.");
|
||||
s_logger.warn("Failed to publish usage event on the event bus.");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -53,7 +53,7 @@ public interface HighAvailabilityManager extends Manager {
|
||||
ConfigKey<Long> TimeBetweenCleanup = new ConfigKey<>("Advanced", Long.class,
|
||||
"time.between.cleanup", "86400", "The time in seconds to wait before the"
|
||||
+ " cleanup thread runs for the different HA-Worker-Threads. The cleanup thread finds all the work items "
|
||||
+ "that were successful and is now ready to be purged from the the database (table: op_ha_work).",
|
||||
+ "that were successful and is now ready to be purged from the database (table: op_ha_work).",
|
||||
true, Cluster);
|
||||
|
||||
ConfigKey<Integer> MaxRetries = new ConfigKey<>("Advanced", Integer.class, "max.retries",
|
||||
|
||||
@ -94,7 +94,7 @@ public class NetworkStateListener implements StateListener<State, Event, Network
|
||||
try {
|
||||
s_eventBus.publish(eventMsg);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to publish state change event on the the event bus.");
|
||||
s_logger.warn("Failed to publish state change event on the event bus.");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -73,7 +73,7 @@ public class StorageUtil {
|
||||
* With managed storage on XenServer and vSphere, CloudStack needs to use an iSCSI SR (XenServer) or datastore (vSphere) per CloudStack
|
||||
* volume. Since XenServer and vSphere are limited to the hundreds with regards to how many SRs or datastores can be leveraged per
|
||||
* compute cluster, this method is used to check a Global Setting (that specifies the maximum number of SRs or datastores per compute cluster)
|
||||
* against what is being requested. KVM does not apply here here because it does not suffer from the same scalability limits as XenServer and
|
||||
* against what is being requested. KVM does not apply here because it does not suffer from the same scalability limits as XenServer and
|
||||
* vSphere do. With XenServer and vSphere, each host is configured to see all the SRs/datastores of the cluster. With KVM, each host typically
|
||||
* is only configured to see the managed volumes of the VMs that are currently running on that host.
|
||||
*
|
||||
|
||||
@ -2492,7 +2492,7 @@ public class NetworkOrchestrator extends ManagerBase implements NetworkOrchestra
|
||||
IPv6Address.class.getName(), null);
|
||||
}
|
||||
|
||||
//remove the secondary ip addresses corresponding to to this nic
|
||||
//remove the secondary ip addresses corresponding to this nic
|
||||
if (!removeVmSecondaryIpsOfNic(nic.getId())) {
|
||||
s_logger.debug("Removing nic " + nic.getId() + " secondary ip addresses failed");
|
||||
}
|
||||
|
||||
@ -1204,7 +1204,7 @@ public class Upgrade218to22 implements DbUpgrade {
|
||||
selectnonRemovedVms.setLong(2, dataCenterId);
|
||||
try (ResultSet nonRemovedVms = selectnonRemovedVms.executeQuery();) {
|
||||
if (nonRemovedVms.next()) {
|
||||
s_logger.warn("Failed to find domR for for account id=" + accountId + " in zone id=" + dataCenterId +
|
||||
s_logger.warn("Failed to find domR for account id=" + accountId + " in zone id=" + dataCenterId +
|
||||
"; will try to locate domR based on user_vm info");
|
||||
//try to get domR information from the user_vm belonging to the account
|
||||
try (PreparedStatement selectNetworkType =
|
||||
|
||||
@ -231,7 +231,7 @@ CREATE TABLE `cloud`.`port_forwarding_rules` (
|
||||
`instance_id` bigint unsigned NOT NULL COMMENT 'vm instance id',
|
||||
`dest_ip_address` char(40) NOT NULL COMMENT 'id_address',
|
||||
`dest_port_start` int(10) NOT NULL COMMENT 'starting port of the port range to map to',
|
||||
`dest_port_end` int(10) NOT NULL COMMENT 'end port of the the port range to map to',
|
||||
`dest_port_end` int(10) NOT NULL COMMENT 'end port of the port range to map to',
|
||||
PRIMARY KEY (`id`),
|
||||
CONSTRAINT `fk_port_forwarding_rules__id` FOREIGN KEY(`id`) REFERENCES `firewall_rules`(`id`) ON DELETE CASCADE
|
||||
) ENGINE=InnoDB DEFAULT CHARSET=utf8;
|
||||
|
||||
@ -445,7 +445,7 @@ CREATE TABLE `cloud`.`external_load_balancer_devices` (
|
||||
`allocation_state` varchar(32) NOT NULL DEFAULT 'Free' COMMENT 'Allocation state (Free/Shared/Dedicated/Provider) of the device',
|
||||
`is_dedicated` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if device/appliance is provisioned for dedicated use only',
|
||||
`is_inline` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer will be used in in-line configuration with firewall',
|
||||
`is_managed` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer appliance is provisioned and its life cycle is managed by by cloudstack',
|
||||
`is_managed` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer appliance is provisioned and its life cycle is managed by cloudstack',
|
||||
`host_id` bigint unsigned NOT NULL COMMENT 'host id corresponding to the external load balancer device',
|
||||
`parent_host_id` bigint unsigned COMMENT 'if the load balancer appliance is cloudstack managed, then host id on which this appliance is provisioned',
|
||||
PRIMARY KEY (`id`),
|
||||
|
||||
@ -1986,7 +1986,7 @@ CREATE TABLE `cloud`.`account_vnet_map` (
|
||||
`uuid` varchar(255) UNIQUE,
|
||||
`vnet_range` varchar(255) NOT NULL COMMENT 'dedicated guest vlan range',
|
||||
`account_id` bigint unsigned NOT NULL COMMENT 'account id. foreign key to account table',
|
||||
`physical_network_id` bigint unsigned NOT NULL COMMENT 'physical network id. foreign key to the the physical network table',
|
||||
`physical_network_id` bigint unsigned NOT NULL COMMENT 'physical network id. foreign key to the physical network table',
|
||||
PRIMARY KEY (`id`),
|
||||
CONSTRAINT `fk_account_vnet_map__physical_network_id` FOREIGN KEY (`physical_network_id`) REFERENCES `physical_network` (`id`) ON DELETE CASCADE,
|
||||
INDEX `i_account_vnet_map__physical_network_id`(`physical_network_id`),
|
||||
|
||||
@ -529,7 +529,7 @@ namespace HypervResource
|
||||
{
|
||||
// Get the virtual switch
|
||||
VirtualEthernetSwitch vSwitch = GetExternalVirtSwitch(vSwitchName);
|
||||
//check the the recevied vSwitch is the same as vSwitchName.
|
||||
//check the recevied vSwitch is the same as vSwitchName.
|
||||
if (!vSwitchName.Equals("") && !vSwitch.ElementName.Equals(vSwitchName))
|
||||
{
|
||||
var errMsg = string.Format("Internal error, coudl not find Virtual Switch with the name : " +vSwitchName);
|
||||
|
||||
@ -801,7 +801,7 @@ public abstract class CitrixResourceBase extends ServerResourceBase implements S
|
||||
}
|
||||
}
|
||||
} catch (final Throwable e) {
|
||||
final String msg = "Unable to get vms through host " + _host.getUuid() + " due to to " + e.toString();
|
||||
final String msg = "Unable to get vms through host " + _host.getUuid() + " due to " + e.toString();
|
||||
s_logger.warn(msg, e);
|
||||
throw new CloudRuntimeException(msg);
|
||||
}
|
||||
|
||||
@ -411,7 +411,7 @@ public class KubernetesClusterActionWorker {
|
||||
return new Pair<>(address.getAddress().addr(), port);
|
||||
}
|
||||
}
|
||||
LOGGER.warn(String.format("No public IP found for the the VPC tier: %s, Kubernetes cluster : %s", network, kubernetesCluster.getName()));
|
||||
LOGGER.warn(String.format("No public IP found for the VPC tier: %s, Kubernetes cluster : %s", network, kubernetesCluster.getName()));
|
||||
return new Pair<>(null, port);
|
||||
}
|
||||
|
||||
@ -584,7 +584,7 @@ public class KubernetesClusterActionWorker {
|
||||
writer.write(data);
|
||||
writer.close();
|
||||
} catch (IOException e) {
|
||||
logAndThrow(Level.ERROR, String.format("Kubernetes Cluster %s : Failed to to fetch script %s",
|
||||
logAndThrow(Level.ERROR, String.format("Kubernetes Cluster %s : Failed to fetch script %s",
|
||||
kubernetesCluster.getName(), filename), e);
|
||||
}
|
||||
return file;
|
||||
|
||||
@ -248,7 +248,7 @@ public class NetscalerResource implements ServerResource {
|
||||
enableLoadBalancingFeature();
|
||||
SSL.enableSslFeature(_netscalerService, _isSdx);
|
||||
|
||||
//if the the device is cloud stack provisioned then make it part of the public network
|
||||
//if the device is cloud stack provisioned then make it part of the public network
|
||||
if (_cloudManaged) {
|
||||
_publicIP = (String)params.get("publicip");
|
||||
_publicIPNetmask = (String)params.get("publicipnetmask");
|
||||
@ -1856,7 +1856,7 @@ public class NetscalerResource implements ServerResource {
|
||||
public static void linkCerts(final nitro_service ns, final String userCertName, final String caCertName) throws ExecutionException {
|
||||
try {
|
||||
|
||||
// the assumption is that that both userCertName and caCertName are present on NS
|
||||
// the assumption is that both userCertName and caCertName are present on NS
|
||||
|
||||
final sslcertkey caCert = sslcertkey.get(ns, caCertName);
|
||||
final sslcertkey userCert = sslcertkey.get(ns, userCertName);
|
||||
|
||||
@ -140,7 +140,7 @@ public final class NiciraNvpConfigureSharedNetworkUuidCommandWrapper extends Com
|
||||
cleanupLSwitchPort(logicalSwitchUuid, lSwitchPort, niciraNvpApi);
|
||||
return handleException(e, command, niciraNvpResource);
|
||||
}
|
||||
s_logger.debug("Logical Router Port " + lRouterPort.getUuid() + " (" + lRouterPort.getDisplayName() + ") successfully attached to to Logical Switch Port " + lSwitchPort.getUuid() + " (" + lSwitchPort.getDisplayName() + ") with a PatchAttachment");
|
||||
s_logger.debug("Logical Router Port " + lRouterPort.getUuid() + " (" + lRouterPort.getDisplayName() + ") successfully attached to Logical Switch Port " + lSwitchPort.getUuid() + " (" + lSwitchPort.getDisplayName() + ") with a PatchAttachment");
|
||||
|
||||
|
||||
//Step 5: Attach lSwitchPort to lRouterPort with a PatchAttachment
|
||||
@ -154,7 +154,7 @@ public final class NiciraNvpConfigureSharedNetworkUuidCommandWrapper extends Com
|
||||
cleanupLSwitchPort(logicalSwitchUuid, lSwitchPort, niciraNvpApi);
|
||||
return handleException(e, command, niciraNvpResource);
|
||||
}
|
||||
s_logger.debug("Logical Switch Port " + lSwitchPort.getUuid() + " (" + lSwitchPort.getDisplayName() + ") successfully attached to to Logical Router Port " + lRouterPort.getUuid() + " (" + lRouterPort.getDisplayName() + ") with a PatchAttachment");
|
||||
s_logger.debug("Logical Switch Port " + lSwitchPort.getUuid() + " (" + lSwitchPort.getDisplayName() + ") successfully attached to Logical Router Port " + lRouterPort.getUuid() + " (" + lRouterPort.getDisplayName() + ") with a PatchAttachment");
|
||||
|
||||
s_logger.info("Successfully attached Logical Switch " + logicalSwitchUuid + " on Logical Router " + logicalRouterUuid + " for Shared Network " + networkId);
|
||||
return new ConfigureSharedNetworkUuidAnswer(command, true, "OK");
|
||||
|
||||
@ -376,7 +376,7 @@ public class LdapListUsersCmd extends BaseListCmd {
|
||||
}
|
||||
|
||||
/**
|
||||
* @return true unless the the user is imported in the specified cloudstack domain from LDAP
|
||||
* @return true unless the user is imported in the specified cloudstack domain from LDAP
|
||||
*/
|
||||
private boolean isNotAlreadyImportedInTheCurrentDomain(LdapUserResponse user) {
|
||||
UserResponse cloudstackUser = getCloudstackUser(user);
|
||||
|
||||
@ -114,7 +114,7 @@ copyvhd()
|
||||
dd if=$srcvhd of=$desvhd bs=512 seek=$(($(($vsize/512))-1)) count=1
|
||||
$VHDUTIL modify -s $vsize -n $desvhd
|
||||
if [ $? -ne 0 ]; then
|
||||
echo "32#failed to set new vhd physical size for vdi vdi $uuid"
|
||||
echo "32#failed to set new vhd physical size for vdi $uuid"
|
||||
cleanup
|
||||
exit 0
|
||||
fi
|
||||
|
||||
@ -373,7 +373,7 @@ public class ApiServer extends ManagerBase implements HttpRequestHandler, ApiSer
|
||||
try {
|
||||
eventBus.publish(event);
|
||||
} catch (EventBusException evx) {
|
||||
String errMsg = "Failed to publish async job event on the the event bus.";
|
||||
String errMsg = "Failed to publish async job event on the event bus.";
|
||||
s_logger.warn(errMsg, evx);
|
||||
}
|
||||
}
|
||||
|
||||
@ -236,7 +236,7 @@ public class ActionEventUtils {
|
||||
try {
|
||||
s_eventBus.publish(event);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to publish action event on the the event bus.");
|
||||
s_logger.warn("Failed to publish action event on the event bus.");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -109,7 +109,7 @@ public class AlertGenerator {
|
||||
try {
|
||||
s_eventBus.publish(event);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to publish alert on the the event bus.");
|
||||
s_logger.warn("Failed to publish alert on the event bus.");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@ -135,7 +135,7 @@ public class Ipv6AddressManagerImpl extends ManagerBase implements Ipv6AddressMa
|
||||
/**
|
||||
* Allocates a public IPv6 address for the guest NIC. It will throw exceptions in the following cases:
|
||||
* <ul>
|
||||
* <li>the the requested IPv6 address is already in use in the network;</li>
|
||||
* <li>the requested IPv6 address is already in use in the network;</li>
|
||||
* <li>IPv6 address is equals to the Gateway;</li>
|
||||
* <li>the network offering is empty;</li>
|
||||
* <li>the IPv6 address is not in the network.</li>
|
||||
@ -151,7 +151,7 @@ public class Ipv6AddressManagerImpl extends ManagerBase implements Ipv6AddressMa
|
||||
/**
|
||||
* Performs some checks on the given IPv6 address. It will throw exceptions in the following cases:
|
||||
* <ul>
|
||||
* <li>the the requested IPv6 address is already in use in the network;</li>
|
||||
* <li>the requested IPv6 address is already in use in the network;</li>
|
||||
* <li>IPv6 address is equals to the Gateway;</li>
|
||||
* <li>the network offering is empty;</li>
|
||||
* <li>the IPv6 address is not in the network.</li>
|
||||
|
||||
@ -834,7 +834,7 @@ public class NetworkServiceImpl extends ManagerBase implements NetworkService, C
|
||||
}
|
||||
|
||||
/**
|
||||
* It allocates a secondary IP alias on the NIC. It can be either an Ipv4 or an Ipv6 or even both, according to the the given IpAddresses object.
|
||||
* It allocates a secondary IP alias on the NIC. It can be either an Ipv4 or an Ipv6 or even both, according to the given IpAddresses object.
|
||||
*/
|
||||
@Override
|
||||
@ActionEvent(eventType = EventTypes.EVENT_NIC_SECONDARY_IP_ASSIGN, eventDescription = "assigning secondary ip to nic", create = true)
|
||||
|
||||
@ -222,7 +222,7 @@ public class DirectPodBasedNetworkGuru extends DirectNetworkGuru {
|
||||
* Linux, FreeBSD and Windows all calculate the same IPv6 address when configured properly.
|
||||
*
|
||||
* Using Router Advertisements the routers in the network should announce the IPv6 CIDR which is configured
|
||||
* in in the vlan table in the database.
|
||||
* in the vlan table in the database.
|
||||
*
|
||||
* This way the NIC will be populated with a IPv6 address on which the Instance is reachable.
|
||||
*/
|
||||
|
||||
@ -1134,7 +1134,7 @@ public class RulesManagerImpl extends ManagerBase implements RulesManager, Rules
|
||||
// revoke all port forwarding rules
|
||||
success = success && applyPortForwardingRules(ipId, _ipAddrMgr.RulesContinueOnError.value(), caller);
|
||||
|
||||
// revoke all all static nat rules
|
||||
// revoke all static nat rules
|
||||
success = success && applyStaticNatRulesForIp(ipId, _ipAddrMgr.RulesContinueOnError.value(), caller, true);
|
||||
|
||||
// revoke static nat for the ip address
|
||||
@ -1190,7 +1190,7 @@ public class RulesManagerImpl extends ManagerBase implements RulesManager, Rules
|
||||
success = success && applyPortForwardingRulesForNetwork(networkId, true, caller);
|
||||
success = success && applyPortForwardingRulesForNetwork(networkId, _ipAddrMgr.RulesContinueOnError.value(), caller);
|
||||
|
||||
// revoke all all static nat rules for the network
|
||||
// revoke all static nat rules for the network
|
||||
success = success && applyStaticNatRulesForNetwork(networkId, true, caller);
|
||||
success = success && applyStaticNatRulesForNetwork(networkId, _ipAddrMgr.RulesContinueOnError.value(), caller);
|
||||
|
||||
|
||||
@ -476,7 +476,7 @@ public class SecurityGroupManagerImpl extends ManagerBase implements SecurityGro
|
||||
protected List<Long> getAffectedVmsForVmStop(VMInstanceVO vm) {
|
||||
List<Long> affectedVms = new ArrayList<Long>();
|
||||
List<SecurityGroupVMMapVO> groupsForVm = _securityGroupVMMapDao.listByInstanceId(vm.getId());
|
||||
// For each group, find the security rules rules that allow the group
|
||||
// For each group, find the security rules that allow the group
|
||||
for (SecurityGroupVMMapVO mapVO : groupsForVm) {// FIXME: use custom sql in the dao
|
||||
//Add usage events for security group remove
|
||||
UsageEventUtils.publishUsageEvent(EventTypes.EVENT_SECURITY_GROUP_REMOVE, vm.getAccountId(), vm.getDataCenterId(), vm.getId(), mapVO.getSecurityGroupId(), vm
|
||||
|
||||
@ -51,7 +51,7 @@ public class ImageStoreDetailsUtil {
|
||||
* Obtain NFS protocol version (if provided) for a store id, if not use default config value<br/>
|
||||
* @param storeId image store id
|
||||
* @return {@code null} if {@code secstorage.nfs.version} is not found for storeId <br/>
|
||||
* {@code X} if {@code secstorage.nfs.version} is found found for storeId
|
||||
* {@code X} if {@code secstorage.nfs.version} is found for storeId
|
||||
*/
|
||||
public String getNfsVersion(long storeId) throws NumberFormatException {
|
||||
|
||||
@ -68,7 +68,7 @@ public class ImageStoreDetailsUtil {
|
||||
* Obtain NFS protocol version (if provided) for a store uuid.<br/>
|
||||
* @param storeUuid image store id
|
||||
* @return {@code null} if {@code secstorage.nfs.version} is not found for storeUuid <br/>
|
||||
* {@code X} if {@code secstorage.nfs.version} is found found for storeUuid
|
||||
* {@code X} if {@code secstorage.nfs.version} is found for storeUuid
|
||||
*/
|
||||
public String getNfsVersionByUuid(String storeUuid){
|
||||
ImageStoreVO imageStore = imageStoreDao.findByUuid(storeUuid);
|
||||
|
||||
@ -548,7 +548,7 @@ public class VolumeApiServiceImpl extends ManagerBase implements VolumeApiServic
|
||||
|
||||
sanitizeFormat(format);
|
||||
|
||||
// Check that the the disk offering specified is valid
|
||||
// Check that the disk offering specified is valid
|
||||
if (diskOfferingId != null) {
|
||||
DiskOfferingVO diskOffering = _diskOfferingDao.findById(diskOfferingId);
|
||||
if ((diskOffering == null) || diskOffering.getRemoved() != null || diskOffering.isComputeOnly()) {
|
||||
@ -738,7 +738,7 @@ public class VolumeApiServiceImpl extends ManagerBase implements VolumeApiServic
|
||||
}
|
||||
}
|
||||
|
||||
// Check that the the disk offering is specified
|
||||
// Check that the disk offering is specified
|
||||
diskOffering = _diskOfferingDao.findById(diskOfferingId);
|
||||
if ((diskOffering == null) || diskOffering.getRemoved() != null || diskOffering.isComputeOnly()) {
|
||||
throw new InvalidParameterValueException("Please specify a valid disk offering.");
|
||||
|
||||
@ -106,7 +106,7 @@ public class SnapshotStateListener implements StateListener<State, Event, Snapsh
|
||||
try {
|
||||
s_eventBus.publish(eventMsg);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to publish state change event on the the event bus.");
|
||||
s_logger.warn("Failed to publish state change event on the event bus.");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -122,7 +122,7 @@ public class VolumeStateListener implements StateListener<State, Event, Volume>
|
||||
try {
|
||||
s_eventBus.publish(eventMsg);
|
||||
} catch (EventBusException e) {
|
||||
s_logger.warn("Failed to state change event on the the event bus.");
|
||||
s_logger.warn("Failed to state change event on the event bus.");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -6134,7 +6134,7 @@ public class UserVmManagerImpl extends ManagerBase implements UserVmManager, Vir
|
||||
*/
|
||||
protected void validateKvmExtraConfig(String decodedUrl) {
|
||||
String[] allowedConfigOptionList = KvmAdditionalConfigAllowList.value().split(",");
|
||||
// Skip allowed keys validation validation for DPDK
|
||||
// Skip allowed keys validation for DPDK
|
||||
if (!decodedUrl.contains(":")) {
|
||||
try {
|
||||
DocumentBuilder builder = ParserUtils.getSaferDocumentBuilderFactory().newDocumentBuilder();
|
||||
|
||||
@ -152,7 +152,7 @@ public class UserVmStateListener implements StateListener<State, VirtualMachine.
|
||||
try {
|
||||
s_eventBus.publish(eventMsg);
|
||||
} catch (org.apache.cloudstack.framework.events.EventBusException e) {
|
||||
s_logger.warn("Failed to publish state change event on the the event bus.");
|
||||
s_logger.warn("Failed to publish state change event on the event bus.");
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
@ -324,7 +324,7 @@ public class RdpClient extends PipelineImpl {
|
||||
|
||||
);
|
||||
|
||||
// Last element of handshake sequence will wake up queue and and socket
|
||||
// Last element of handshake sequence will wake up queue and socket
|
||||
// output pull loop, which will switch links, between socket output and
|
||||
// queue, from push mode to pull mode.
|
||||
link(HANDSHAKE_END + " >queue", "queue", "OUT");
|
||||
|
||||
@ -543,7 +543,7 @@ for example:
|
||||
|
||||
This command sets up the CloudStack Management Server and CloudStack Usage Server database configuration (connection credentials and host information) based on the first argument.
|
||||
|
||||
If the the --deploy-as option is present, this command will also connect to the database using the administrative credentials specified as the value for the --deploy-as argument, construct the database environment needed to run the CloudStack Management Server, and alter the password specified for the user in the first argument. In this case, the user name specified in --deploy-as= cannot be the same as the user name specified for the connection credentials that the CloudStack Management Server will be set up with.
|
||||
If the --deploy-as option is present, this command will also connect to the database using the administrative credentials specified as the value for the --deploy-as argument, construct the database environment needed to run the CloudStack Management Server, and alter the password specified for the user in the first argument. In this case, the user name specified in --deploy-as= cannot be the same as the user name specified for the connection credentials that the CloudStack Management Server will be set up with.
|
||||
|
||||
If a server-setup.xml cloud setup information file is specified with the --auto option, this command will also construct a customized database environment according to the cloud setup information in the file.
|
||||
|
||||
|
||||
@ -801,7 +801,7 @@ CREATE TABLE `cloud`.`port_forwarding_rules` (
|
||||
`instance_id` bigint unsigned NOT NULL COMMENT 'vm instance id',
|
||||
`dest_ip_address` char(40) NOT NULL COMMENT 'id_address',
|
||||
`dest_port_start` int(10) NOT NULL COMMENT 'starting port of the port range to map to',
|
||||
`dest_port_end` int(10) NOT NULL COMMENT 'end port of the the port range to map to',
|
||||
`dest_port_end` int(10) NOT NULL COMMENT 'end port of the port range to map to',
|
||||
PRIMARY KEY (`id`),
|
||||
CONSTRAINT `fk_port_forwarding_rules__id` FOREIGN KEY(`id`) REFERENCES `firewall_rules`(`id`) ON DELETE CASCADE,
|
||||
CONSTRAINT `fk_port_forwarding_rules__instance_id` FOREIGN KEY `fk_port_forwarding_rules__instance_id` (`instance_id`) REFERENCES `vm_instance` (`id`) ON DELETE CASCADE
|
||||
@ -2137,7 +2137,7 @@ CREATE TABLE `cloud`.`external_load_balancer_devices` (
|
||||
`allocation_state` varchar(32) NOT NULL DEFAULT 'Free' COMMENT 'Allocation state (Free/Shared/Dedicated/Provider) of the device',
|
||||
`is_dedicated` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if device/appliance is provisioned for dedicated use only',
|
||||
`is_inline` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer will be used in in-line configuration with firewall',
|
||||
`is_managed` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer appliance is provisioned and its life cycle is managed by by cloudstack',
|
||||
`is_managed` int(1) unsigned NOT NULL DEFAULT 0 COMMENT '1 if load balancer appliance is provisioned and its life cycle is managed by cloudstack',
|
||||
`host_id` bigint unsigned NOT NULL COMMENT 'host id corresponding to the external load balancer device',
|
||||
`parent_host_id` bigint unsigned COMMENT 'if the load balancer appliance is cloudstack managed, then host id on which this appliance is provisioned',
|
||||
PRIMARY KEY (`id`),
|
||||
|
||||
@ -704,7 +704,7 @@ const UI = {
|
||||
|
||||
updateControlbarHandle() {
|
||||
// Since the control bar is fixed on the viewport and not the page,
|
||||
// the move function expects coordinates relative the the viewport.
|
||||
// the move function expects coordinates relative the viewport.
|
||||
const handle = document.getElementById("noVNC_control_bar_handle");
|
||||
const handleBounds = handle.getBoundingClientRect();
|
||||
UI.moveControlbarHandle(handleBounds.top);
|
||||
|
||||
@ -347,7 +347,7 @@ export default class Display {
|
||||
// 1. https://bugzilla.mozilla.org/show_bug.cgi?id=1194719
|
||||
//
|
||||
// We need to set these every time since all properties are reset
|
||||
// when the the size is changed
|
||||
// when the size is changed
|
||||
this._drawCtx.mozImageSmoothingEnabled = false;
|
||||
this._drawCtx.webkitImageSmoothingEnabled = false;
|
||||
this._drawCtx.msImageSmoothingEnabled = false;
|
||||
|
||||
@ -883,7 +883,7 @@ export default class RFB extends EventTargetMixin {
|
||||
size.w + 'x' + size.h);
|
||||
}
|
||||
|
||||
// Gets the the size of the available screen
|
||||
// Gets the size of the available screen
|
||||
_screenSize() {
|
||||
let r = this._screen.getBoundingClientRect();
|
||||
return { w: r.width, h: r.height };
|
||||
|
||||
@ -483,7 +483,7 @@ class TestHostHighAvailability(cloudstackTestCase):
|
||||
self.assertEqual(
|
||||
vm_response.id,
|
||||
vm.id,
|
||||
"The virtual machine id and the the virtual machine from listVirtualMachines is not matching."
|
||||
"The virtual machine id and the virtual machine from listVirtualMachines is not matching."
|
||||
)
|
||||
|
||||
self.assertEqual(
|
||||
|
||||
@ -427,7 +427,7 @@ class TestIsolatedNetwork(cloudstackTestCase):
|
||||
self.cleanup.append(network)
|
||||
self.assertEqual(network.state.lower() == ALLOCATED.lower(),
|
||||
True,
|
||||
"Admin User is not able to create a network for for other users in other domain")
|
||||
"Admin User is not able to create a network for other users in other domain")
|
||||
|
||||
|
||||
## Test cases relating to createNetwork as domain admin user
|
||||
@ -575,7 +575,7 @@ class TestIsolatedNetwork(cloudstackTestCase):
|
||||
except Exception as e:
|
||||
self.debug("When user tries to create network for users in their domain %s" % e)
|
||||
if not CloudstackAclException.verifyMsginException(e, CloudstackAclException.UNABLE_TO_LIST_NETWORK_ACCOUNT):
|
||||
self.fail("Error message validation failed when when User tries to create network for other users in their domain ")
|
||||
self.fail("Error message validation failed when User tries to create network for other users in their domain ")
|
||||
|
||||
@attr("simulator_only", tags=["advanced"], required_hardware="false")
|
||||
def test_10_createNetwork_user_foruserinotherdomain(self):
|
||||
@ -674,7 +674,7 @@ class TestIsolatedNetwork(cloudstackTestCase):
|
||||
self.cleanup.append(vm)
|
||||
self.assertEqual(vm.state.lower() == RUNNING.lower() and vm.account == self.account_d2a.name and vm.domainid == self.account_d2a.domainid,
|
||||
True,
|
||||
"Admin User is not able to deploy VM for users users in other domain")
|
||||
"Admin User is not able to deploy VM for users in other domain")
|
||||
|
||||
@attr("simulator_only", tags=["advanced"], required_hardware="false")
|
||||
def test_13_1_deployvm_admin_foruserinotherdomain_crossnetwork(self):
|
||||
|
||||
@ -1714,7 +1714,7 @@ class TestSharedNetworkImpersonation(cloudstackTestCase):
|
||||
accountid=self.account_d2a.name,
|
||||
domainid=self.account_d2a.domainid
|
||||
)
|
||||
self.fail("Domain admin is able able to deploy a VM for an regular user from a differnt domain in a shared network with scope=account")
|
||||
self.fail("Domain admin is able to deploy a VM for an regular user from a differnt domain in a shared network with scope=account")
|
||||
except Exception as e:
|
||||
self.debug("When a user from different domain deploys a VM in a shared network with scope=account %s" % e)
|
||||
if not CloudstackAclException.verifyMsginException(e, CloudstackAclException.NO_PERMISSION_TO_OPERATE_DOMAIN):
|
||||
|
||||
@ -331,7 +331,7 @@ class TestBaseImageUpdate(cloudstackTestCase):
|
||||
except Exception as e:
|
||||
self.fail("Failed to reboot the virtual machines, %s" % e)
|
||||
|
||||
# Check if the the root disk was destroyed and recreated for isVolatile=True
|
||||
# Check if the root disk was destroyed and recreated for isVolatile=True
|
||||
self.debug("Checking root disk of VM with isVolatile=True")
|
||||
vms = VirtualMachine.list(
|
||||
self.apiclient,
|
||||
@ -358,7 +358,7 @@ class TestBaseImageUpdate(cloudstackTestCase):
|
||||
%(vm_with_reset.nic[0].ipaddress, self.vm_with_reset.nic[0].ipaddress)
|
||||
)
|
||||
|
||||
# Check if the the root disk was not destroyed for isVolatile=False
|
||||
# Check if the root disk was not destroyed for isVolatile=False
|
||||
self.debug("Checking root disk of VM with isVolatile=False")
|
||||
vms = VirtualMachine.list(
|
||||
self.apiclient,
|
||||
@ -473,7 +473,7 @@ class TestBaseImageUpdate(cloudstackTestCase):
|
||||
%(vm_with_reset.nic[0].ipaddress, self.vm_with_reset.nic[0].ipaddress)
|
||||
)
|
||||
|
||||
# Check if the the root disk was not destroyed for isVolatile=False
|
||||
# Check if the root disk was not destroyed for isVolatile=False
|
||||
self.debug("Checking template id of VM with isVolatile=False")
|
||||
vms = VirtualMachine.list(
|
||||
self.apiclient,
|
||||
@ -606,7 +606,7 @@ class TestBaseImageUpdate(cloudstackTestCase):
|
||||
except Exception as e:
|
||||
self.fail("Failed to reboot the virtual machine. Error: %s" % e)
|
||||
|
||||
# Check if the the root disk was destroyed and recreated for isVolatile=True
|
||||
# Check if the root disk was destroyed and recreated for isVolatile=True
|
||||
self.debug("Checking whether root disk of VM with isVolatile=True was destroyed")
|
||||
vms = VirtualMachine.list(
|
||||
self.apiclient,
|
||||
|
||||
@ -366,7 +366,7 @@ class ConfigDriveUtils:
|
||||
return None
|
||||
|
||||
def _get_config_drive_data(self, ssh, file, name, fail_on_missing=True):
|
||||
"""Fetches the content of a file file on the config drive
|
||||
"""Fetches the content of a file on the config drive
|
||||
|
||||
:param ssh: SSH connection to the VM
|
||||
:param file: path to the file to fetch
|
||||
@ -1613,7 +1613,7 @@ class TestConfigDrive(cloudstackTestCase, ConfigDriveUtils):
|
||||
"List network should return a valid list"
|
||||
)
|
||||
self.assertEqual(network.name, networks[0].name,
|
||||
"Name of the network should match with with the "
|
||||
"Name of the network should match with the "
|
||||
"returned list data"
|
||||
)
|
||||
if state:
|
||||
@ -1777,11 +1777,11 @@ class TestConfigDrive(cloudstackTestCase, ConfigDriveUtils):
|
||||
# And the VM is successfully deployed and is in the "Running" state
|
||||
# And there is no VR is deployed.
|
||||
# 4. And the user data in the ConfigDrive device is as expected
|
||||
# 5. And the the vm password in the ConfigDrive device is as expected
|
||||
# 5. And the vm password in the ConfigDrive device is as expected
|
||||
|
||||
# 6. When I stop, reset the password, and start the VM
|
||||
# 7. Then I can login into the VM using the new password.
|
||||
# 8. And the the vm password in the ConfigDrive device is the new one
|
||||
# 8. And the vm password in the ConfigDrive device is the new one
|
||||
|
||||
# 9. Verify various scenarios and check the data in configdriveIso
|
||||
# 10. Delete all the created objects (cleanup).
|
||||
@ -2000,11 +2000,11 @@ class TestConfigDrive(cloudstackTestCase, ConfigDriveUtils):
|
||||
# And the VM is successfully deployed and is in the "Running" state
|
||||
|
||||
# 4. And the user data in the ConfigDrive device is as expected
|
||||
# 5. And the the vm password in the ConfigDrive device is as expected
|
||||
# 5. And the vm password in the ConfigDrive device is as expected
|
||||
|
||||
# 6. When I stop, reset the password, and start the VM
|
||||
# 7. Then I can login into the VM using the new password.
|
||||
# 8. And the the vm password in the ConfigDrive device is the new one
|
||||
# 8. And the vm password in the ConfigDrive device is the new one
|
||||
|
||||
# 9. Verify various scenarios and check the data in configdriveIso
|
||||
# 10. Delete all the created objects (cleanup).
|
||||
|
||||
@ -105,7 +105,7 @@ class TestLdap(cloudstackTestCase):
|
||||
self.apiClient.deleteAccount(deleteAcct)
|
||||
|
||||
self.debug(
|
||||
"Deleted the the following account name %s:" %
|
||||
"Deleted the following account name %s:" %
|
||||
acct_name)
|
||||
|
||||
if self.ldapconfRes == 1:
|
||||
|
||||
@ -379,7 +379,7 @@ class TestLdap(cloudstackTestCase):
|
||||
self.apiClient.deleteAccount(deleteAcct2)
|
||||
|
||||
self.debug(
|
||||
"Deleted the the following account name %s:" %
|
||||
"Deleted the following account name %s:" %
|
||||
acct_name)
|
||||
|
||||
except Exception as e:
|
||||
|
||||
@ -297,7 +297,7 @@ class TestBasicOperations(cloudstackTestCase):
|
||||
# Validations:
|
||||
# 1. Step 3 should succeed
|
||||
# 2. Step 4 should fail
|
||||
# 3. Step 5 should should fail
|
||||
# 3. Step 5 should fail
|
||||
# 4. Step 6 should fail
|
||||
|
||||
self.account = Account.create(
|
||||
|
||||
@ -850,7 +850,7 @@ class TestPrivateGwACL(cloudstackTestCase):
|
||||
self.assertTrue(check_state == 0, "Routers private gateway interface should not be on the same state!")
|
||||
self.assertTrue(check_mac == 0, "Routers private gateway interface should not have the same mac address!")
|
||||
else:
|
||||
self.assertTrue(check_state == 1, "Routers private gateway interface should should have been removed!")
|
||||
self.assertTrue(check_state == 1, "Routers private gateway interface should have been removed!")
|
||||
|
||||
def check_routers_state(self, status_to_check="PRIMARY", expected_count=1):
|
||||
routers = self.query_routers()
|
||||
|
||||
@ -157,7 +157,7 @@ class TestMultipleVolumeSnapshots(cloudstackTestCase):
|
||||
storage pools available in the setup")
|
||||
if len(list(storagePool for storagePool in self.pools
|
||||
if storagePool.scope == "CLUSTER")) < 2:
|
||||
self.skipTest("There must be at at least two cluster wide\
|
||||
self.skipTest("There must be at least two cluster wide\
|
||||
storage pools available in the setup")
|
||||
except Exception as e:
|
||||
self.skipTest(e)
|
||||
|
||||
@ -141,7 +141,7 @@ class TestRestoreVM(cloudstackTestCase):
|
||||
"Check: Failed to list cluster wide storage pools")
|
||||
|
||||
if len(self.pools) < 2:
|
||||
self.skipTest("There must be at at least two cluster wide\
|
||||
self.skipTest("There must be at least two cluster wide\
|
||||
storage pools available in the setup")
|
||||
|
||||
except Exception as e:
|
||||
|
||||
@ -432,7 +432,7 @@ class TestUsage(cloudstackTestCase):
|
||||
def test_01_positive_tests_usage(self):
|
||||
""" Positive test for usage test path
|
||||
|
||||
# 1. Register a template and verify that usage usage is generated
|
||||
# 1. Register a template and verify that usage is generated
|
||||
for correct size of template
|
||||
# 2. Register an ISO, verify usage is generate for the correct size
|
||||
of ISO
|
||||
|
||||
@ -732,7 +732,7 @@ class TestPathVolume(cloudstackTestCase):
|
||||
self.assertGreater(
|
||||
len(list_pool),
|
||||
0,
|
||||
"Check the list list storagepoolresponse for vm id: %s" %
|
||||
"Check the list storagepoolresponse for vm id: %s" %
|
||||
list_volume[0].storageid)
|
||||
list_pools = StoragePool.list(self.apiclient,
|
||||
scope=list_pool[0].scope
|
||||
|
||||
@ -49,7 +49,7 @@ deploy_server() {
|
||||
echo "Deploying management server software on remote machine $1"
|
||||
|
||||
rsync -avzh -e ssh $path/cloudstack-oss root@$1:/root/
|
||||
if [ $? -gt 0 ]; then echo "failed to rsync software between $path/cloudstack-oss and and remote machine $1"; return 2; fi
|
||||
if [ $? -gt 0 ]; then echo "failed to rsync software between $path/cloudstack-oss and remote machine $1"; return 2; fi
|
||||
|
||||
ssh root@$1 "cd /root/cloudstack-oss && ant clean build-all deploy-server"
|
||||
if [ $? -gt 0 ]; then echo "failed to deploy cluster on $1"; return 2; fi
|
||||
|
||||
@ -280,7 +280,7 @@ class UpdateConfigTestCase(SystemVMTestCase):
|
||||
unique["eth%s" % ips["nic_dev_id"]] = 1
|
||||
|
||||
# If this is the first run, the drops will not be there yet
|
||||
# this is so I can get get a true count of what is explicitly added
|
||||
# this is so I can get a true count of what is explicitly added
|
||||
drops = len(unique)
|
||||
for dev in unique:
|
||||
drops -= ip.count_fw_rules('ACL_INBOUND_%s -j DROP' % dev)
|
||||
|
||||
@ -90,7 +90,7 @@ h. puppet - set puppetmaster to listen on 8140
|
||||
|
||||
2. NFS storage - the nfs server is a single server serving as both primary and secondary storage. This is likely a limitation when compared to true production deployments but serves in good stead for a test setup. Where it becomes a limitation is in testing different storage backends. Object stores, local storage, clustered local storage etc are not addressed by this setup.
|
||||
|
||||
3. Hypervisor hosts - There currently are 4 hosts in this environment. These are arranged at the moment in three pods so as to be capable of being deployed in a two zone environment. One zone with two pods and and a second zone with a single pod. This covers tests that depend on
|
||||
3. Hypervisor hosts - There currently are 4 hosts in this environment. These are arranged at the moment in three pods so as to be capable of being deployed in a two zone environment. One zone with two pods and a second zone with a single pod. This covers tests that depend on
|
||||
a. single zone/pod/cluster
|
||||
b. multiple cluster
|
||||
c. inter-zone tests
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user