Displaying 10 results from an estimated 10 matches for "storagedevic".
Did you mean:
storagedevs
2009 May 28
0
[PATCH server] Use qpid for migration and add more debugging to taskomatic.
This patch uses the qpid migration call which requires the latest
libvirt-qpid and libvirt. I also add a bunch of debug logging here
which is switchable but I've left it on for now so all users will
have this in their logs.
Signed-off-by: Ian Main <imain at redhat.com>
---
src/task-omatic/taskomatic.rb | 36 ++++++++++++++++++++++++++++--------
1 files changed, 28 insertions(+), 8
2010 Aug 25
2
[PATCH] Virtio support
From: Simon Courtois <scourtois at linagora.com>
Signed-off-by: Simon Courtois <scourtois at linagora.com>
---
src/app/controllers/vm_controller.rb | 19 ++++++++++++++++-
src/app/views/vm/_form.rhtml | 25 ++++++++++++++++++++---
src/db/migrate/044_add_virtio_to_vm_and_nic.rb | 11 ++++++++++
src/public/stylesheets/components.css | 5
2010 Sep 02
1
[PATCH 1/1] Introduce an option to always pxe-boot a vm.
Previously, a pxe-booted vm would always boot to HD at next startup.
Signed-off-by: Nicolas Ochem <nicolas.ochem at alcatel-lucent.com>
---
src/app/controllers/vm_controller.rb | 1 +
src/app/models/vm.rb | 15 +++++++++++----
src/app/views/vm/_form.rhtml | 2 +-
src/task-omatic/taskomatic.rb | 13 +++++++++++--
4 files changed, 24 insertions(+), 7
2009 Sep 24
0
[PATCH server] Make volume finding more robust.
From: Ian Main <imain at management.mains.priv>
This patch makes finding of volumes within attached pools more rubust
by looking for the filename if it fails to find it by key. This
fixes the volume finding for cobbler iso images mounted on nfs.
Signed-off-by: Ian Main <imain at redhat.com>
---
src/task-omatic/taskomatic.rb | 8 +++++++-
1 files changed, 7 insertions(+), 1
2009 Jul 24
1
permit many-to-many vms / networks relationship redux
redux patchset permitting a vm to be associated with
multiple networks and vice-versa. updated patchset so
as to be applicable against current oVirt server HEAD
these patches may be applied in any order, they all
need to be pushed together
2009 Jul 09
2
permit many-to-many vms / networks relationship
This patchset contains changes to the ovirt server
frontend, backend, and tests components, permitting vms
to be associated with multiple networks and vice versa.
Also included are two patches which are required for the frontend
bits; a patch adding collapsable sections to the vm form, which
in itself depends on the second patch that provides default values
for the cpu and memory vm table fields
2007 Nov 13
4
sd_max_throttle
Hello,
we are using hardware array and its vendor recommends the following setting in
/etc/system:
set sd:sd_max_throttle = <value>
or
set ssd:ssd_max_throttle = <value>
Is it possible to monitor *somehow* whether the variable becomes
sort of bottleneck ? Or how its value influences io traffic ?
Regards
przemol
2009 Jul 13
1
[PATCH] Use volume key instead of path to identify volume.
This patch teaches taskomatic to use the volume 'key' instead of the
path from libvirt to key the volume off of in the database. This fixes
the duplicate iscsi volume bug we were seeing. The issue was that
libvirt changed the way they name storage volumes and included a local
ID that changed each time it was attached.
Note that the first run with this new patch will cause duplicate
2009 Jun 30
0
[PATCH server] permit many-to-many vms / networks relationship
- db, model, service, controller, and view changes
- many tests additions
- various fixes / cleanup to get things working
depends on my last two (still unacked) patches:
- add collapsable sections to vm form
- provide default vm allocated cpu and memory values
ovirt-agent might need to be updated to work with the changes
---
src/app/controllers/pool_controller.rb | 2 +-
2008 Apr 01
29
OpenSolaris ZFS NAS Setup
If it''s of interest, I''ve written up some articles on my experiences of building a ZFS NAS box which you can read here:
http://breden.org.uk/2008/03/02/a-home-fileserver-using-zfs/
I used CIFS to share the filesystems, but it will be a simple matter to use NFS instead: issue the command ''zfs set sharenfs=on pool/filesystem'' instead of ''zfs set