search for: georgoulia

Displaying 10 results from an estimated 10 matches for "georgoulia".

Did you mean: georgoulias
2008 Jul 28
4
RHEL 5.2 i386 image - Xen kernel doesn''t recognize the total amount of system memory
Hello, We have installed RHEL 5.2 Server ( i386 image ) on a machine with 64GB of RAM. When booting the default PAE kernel ( 2.6.18-92.el5PAE ) that is shipped with RHEL 5.2 installation the OS detects 64GB of RAM. But when booting the Xen kernel ( 2.6.18-92.el5xen ) the OS detects only 14GB of the total amount of available RAM. Did anyone encounter this problem before ? Is it a normal
2011 Feb 02
1
iSCSI storage pool questions
Hi All, I've been trying to figure out the best way of using an iSCSI SAN with KVM and thanks to a helpful post by Tom Georgoulias that I found on this list (https://www.redhat.com/archives/libvirt-users/2010-May/msg00008.html), it appears I have a solution. What I'm wondering is the following: 1) If I use an iSCSI LUN as the storage pool (instead of creating an LVM VG from this iSCSI LUN manually first and using LVM in...
2010 Mar 10
2
unable to get domain status from libvirt & KVM
I have a python script that monitors the VMs on physical host servers running Xen, but the script doesn't work properly on a server I just built with KVM. The script runs as a non-root user and simply gathers some details on the status and names of the domains running on the host. Both Xen and KVM servers are running the same version of libvirt (libvirt-0.6.3-20.1.el5_4) and have the
2003 May 16
1
data=journal or data=writeback for NFS server
I'm trying to increase the performance of an NFS server running Red Hat Linux 7.3 and using ext3. This server sees hefty read & write activity and I've been getting a lot of complaints that writing to the disks is taking a long time. The hardware is ample (Dell PE 2650 with 2x2.4GHz Xeons, 6GB RAM, PERC3-DC RAID controller & attached PowerVault 220S) and network connectivity
2010 Jan 21
1
/proc/mounts always shows "nobarrier" option for xfs, even when mounted with "barrier"
Ran into a confusing situation today. When I mount an xfs filesystem on a server running centos 5.4 x86_64 with kernel 2.6.18-164.9.1.el5, the barrier/nobarrier mount option as displayed in /proc/mounts is always set to "nobarrier" Here's an example: [root at host ~]# mount -o nobarrier /dev/vg1/homexfs /mnt [root at host ~]# grep xfs /proc/mounts /dev/vg1/homexfs /mnt xfs
2010 Jun 16
1
how to match the ID of a LUN in a storage pool with the GUID on the target server
I've configured a libvirt storage pool using an iscsi target from a Sun 7310 storage appliance and am using the LUNs in this target as volumes for my KVM guests. The setup is very similar to what Daniel covered in a recent blog posting: http://berrange.com/posts/2010/05/05/provisioning-kvm-virtual-machines-on-iscsi-the-hard-way-part-2-of-2/ It works great, but I can't figure out how
2010 May 28
1
bogus bond0 device showing up in /proc/net/dev
I'm running into a situation where a bogus bonded interface named "bond0" is being created, in addition to the desired "bond2" interface. Can anyone confirm this? Anyone know why it's happening or what I do to get rid of it? I wanted to start my numbering scheme at 2 instead of 0, which I didn't think would be a problem. As you can see, I have no reference
2010 Mar 11
2
unable to get domain status from libvirt & KVM
I have a python script that monitors the VMs on physical host servers running Xen, but the script doesn't work properly on a server I just built with KVM. The script runs as a non-root user (same on all servers) and simply gathers some details on the status and names of the domains running on the host. Both Xen and KVM servers are running the same version of libvirt
2010 Mar 12
1
unable to get domain list of KVM guests from libvirt as non-root user
I have a python script that monitors the guest VMs on a servers running Xen, but the script doesn't work properly on a server I just built with KVM. The script runs as a non-root user (same on all servers) and simply gathers some details on the status and names of the domains running on the host. The Xen and KVM servers are running CentOS 5.4 x86_64 and are using the same version of
2011 Jun 09
0
old, deleted LUN still showing up after a virsh pool-refresh
Today I upgraded one of our KVM physical host servers to Centos 5.6. The other server remains on Centos 5.5. I've configured a storage pool in libvirt to use LUNs from an iscsi storage system as the volumes for each VM. After the upgrade, I decommisioned one of the VMs and removed the LUN from the storage system. Normally after this kind of operation, I simply need to run this command