This is a very dangerous file. Although we do not package it, it would be very dangerous to do so.
If this file would be present on a HyperVisor it would cause all instances to be stopped on a libvirt restart.
There is no need for stopping all instances when libvirt is being stopped (or restarted).
Using this via Vagrant assumes that Vagrant is patched with the changes
made by Edison. The pull request to upstream these changes into the
Vagrant project are here: https://github.com/mitchellh/vagrant/pull/1043 )
This patch adds RBD (RADOS Block Device) support for primary storage in combination with KVM.
To get this patch working you need:
- libvirt-java 0.4.8
- libvirt with RBD storage pool support (>0.9.13)
- Qemu with RBD support (>0.14)
The primary storage does not support all the functions of CloudStack yet, for example snapshotting is disabled
due to the fact that backupping up a RBD snapshot is not possible in the way CloudStack wants to do it.
Creating templates from RBD volumes goes well, creating a VM from a template however is still a hit-and-miss.
NFS primary storage is also still required, you are not able to run your System VM's from RBD, they will need
to run on NFS.
Other then these points you can run instances with RBD backed disks.
un-allocated space is insufficient on primary storage
check the availability of un-allocated primary storage space during
planning stage, for multiple-volume VM creation scenario
modification in StorageManagerImpl.java and StorageManager.java:
add a new method storagePoolHasEnoughSpace(List<Volumes>, StoragePool),
check if storagePool has enough space for all requested volumes
modification in FirstfitPlanner.findPotentialDeploymentResources:
handle multiple volume case, keep track of allocated volumes for pools
and call storagePoolHasEnoughSpace to check space availability
modification in AbstractStoragePoolAllocator.java:
extract capacity computation logic and make a new method in
StorageManagerImpl
RB: https://reviews.apache.org/r/6028/
Send-by: mice_xia@tcloudcomputing.com
This has been broken for a long time right now, this is a quick fix to get the generation working again.
This enables us to build the Debian packages again with:
$ dpkg-buildpackage
WAF still needs to go though :)
host if there are multiple storage pools in a cluster.
The issue is as follows:
1. When CloudStack detects that a host is not responding to ping
requests it'll send a fence command for this host to another host in the
cluster.
2. The agent takes a long time to respond to this check if the storage
is fenced. This is because the agent checks if the first host is writing
to its heartbeat file on all pools in the cluster. It is doing this in a
sequential manner on all storage pool.
Making a fix to get rid of sleep, wait during HA. The behavior is now
similar to Xenserver.
RB: https://reviews.apache.org/r/6133/
Send-by:devdeep.singh@citrix.com
We also exit earlier, we don't display that we are even trying to start.
When we detect the agent is already running we exit right away with a message.
Although this master branch doesn't contain 3.0.2 since we moved passed that
it didn't feel apprioriate to bump the version to 4.0.0 (yet).
This file should be updated to 4.0.0 at the point where we release 4.0
No need for these packages to depend on any cloud-agent package.
If they need files from this packages, then we need to move those files.
The client/management server should never depend on agent packages.
With LSB there is no need for having different init script for different distributions.
This init script should be fully LSB compliant and should work on all Linux platforms we support.
RHEL, (Open)SUSE, Debian and Ubuntu all support at least LSB 3.1