Displaying 20 results from an estimated 20000 matches similar to: "cluster management"
2008 Nov 10
14
GPL PV Intermitent Network Problems
Well, I seem to have developed a strange issue with the GPL PV drivers
in Windows. I''m running Xen 3.2 on SLES 10 SP2 and running some Windows
XP HVM domUs. Most (12-15) of my Windows machines work just fine. Two
of them, though, have started exhibiting a strange behavior with
intermittent network connectivity. Basically, the network starts up and
connects initially. After some
2008 Aug 18
5
HVM windows - PCI IRQ firing on both CPU''s
I''m just doing some testing on the gplpv drivers with different ways of
handling interrupts, and I''m trying a scheme where each xen device (eg
vbd/vif) driver attaches to the same IRQ as the pci driver, and each
handles it in sequence.
In testing though, I noticed the following when logging what each ISR is
doing:
60.32381439 - evtchn event on port 5
60.32384109 - port 5
2009 May 03
5
W2k3 HVM domain QEMU drivers
So I successfully installed the GPLPV drivers in a Windows installation I
have running under Xen. Turns out SP2 is needed - I wish this was slightly
better documented somewhere, as the machine ended up bluescreening and I had
to use /NOGPLPV to get it to boot again so I could upgrade it to SP2.
My question is this: I ended up using ''disable'' in device manager for the
rtl8139 and
2008 Dec 26
17
Multiple IRQ''s in HVM for Windows
I really need to have the ability to tie event channel port''s to
interrupts for my gplpv drivers under Windows. Is anyone working on
anything like this? Does MSI allow more than one interrupt per PCI
device?
Thanks
James
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
http://lists.xensource.com/xen-devel
2010 Nov 02
4
Windows 2008 SBS suddenly disappearing on Xen 4.0.0
Hi all,
when running Windows 2008 SBS on Xen 4.0.0, it disappears after some time (longest time I had it running was 20 minutes). In /var/log/xen/qemu-dm-win1.log I find:
xc_map_foreign_batch: mmap failed: Cannot allocate memory
xc_map_foreign_bulk error 12
Anyone any idea how to troubleshoot this? The Hypervisor has 16 GB of RAM, currently 2 GB are allocated to Dom0. This specific DomU has 8
2009 Nov 02
4
vps file lost after server crash
Hello, All,
Today one of our servers was crashed, after I rebooted the server, one
vps can not boot up, and it needs fsck, after I run fsck command, a
lot of files lost, and also some file content was changed. Not sure
how could it happen, do you have any advice that can avoid this or
recover the data?
Thanks.
_______________________________________________
Xen-users mailing list
2012 Feb 07
7
GPLPV, RDP and network latency
Hello!
Has anybody experienced network latency problems with combination of Windows 7, GPLPV drivers and RDP connection?
Any Windows pop-up message(such as "command not found" error message in "Run command:" dialog, or dividing by zero in windows calc) causes a short freeze of RDP session and looks like that from dom0:
PING 192.168.44.65 (192.168.44.65) 56(84) bytes of data.
2007 Nov 25
4
behaviour of ''xm block-attach'' changed with 3.1.2?
''xm block-attach'' doesn''t let me assign the same backend device anymore,
even if I use ''r!'' or ''w!''.
I think this has changed since 3.1.1 (or maybe since 3.1.0).
Is this the way it is supposed to work?
Thanks
James
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
2009 Feb 10
7
hang on restore in 3.3.1
I am having problems with save/restore under 3.3.1 in the GPLPV drivers.
I call hvm_shutdown(xpdd, SHUTDOWN_suspend), but as soon as I lower IRQL
(enabling interrupts), qemu goes to 100% CPU and the DomU load goes
right up too.
Xentrace is showing a whole lot of this going on:
CPU0 200130258143212 (+ 770) hypercall [ rip =
0x000000008020632a, eax = 0xffffffff ]
CPU0 200130258151107 (+
2010 Jan 30
20
"Iomem mapping not permitted" during windows crash dump under GPLPV
I''ve recently noticed that my windows crash dumps fail at around 40-50%
under GPLPV. ''xm dmesg'' shows the following:
(XEN) grant_table.c:350:d0 Iomem mapping not permitted ffffffffffffffff
(domain 865)
At first I thought that the cause was just a bug in my grant ref code
but it just occurred to me that this could be happening when Windows
tries to write out the
2012 Jun 28
8
GPLPV, clock drift and PVUSB in Windows XP HVM
1. Shouldn''t the GPLPV drivers take care of the (bad) clock drift I''m experiencing in my Windows XP HVM? Or is there some other way around this problem that I haven''t been able to find on Google? How can I tell if the GPLPV drivers are active? I''ve added the /gplpv switch to the boot.ini file and the virtual NIC is definitely using the GPLPV version but other
2008 Aug 20
44
GPL PV drivers for Windows 0.9.11-pre12
I''ve just uploaded 0.9.11-pre12 of the GPL PV drivers for Windows.
Since -pre10 (and -pre11) I''ve fixed a heap of crashes that were
plaguing xennet under load, and also rewritten the interrupt/event
distribution logic to improve performance.
Under windows 2003 I can now get network speeds of 1-2Gbit/second TX and
600Gbit/second RX, which is considerably better than I was
2008 Aug 20
44
GPL PV drivers for Windows 0.9.11-pre12
I''ve just uploaded 0.9.11-pre12 of the GPL PV drivers for Windows.
Since -pre10 (and -pre11) I''ve fixed a heap of crashes that were
plaguing xennet under load, and also rewritten the interrupt/event
distribution logic to improve performance.
Under windows 2003 I can now get network speeds of 1-2Gbit/second TX and
600Gbit/second RX, which is considerably better than I was
2010 Jun 08
32
Problems with GPLPV network latency
Hi,
DomU is a Win2008 R2 64
When i install the GPLPV drivers the network latency goes from 15ms to
random numbers up to 1200ms and eventually dies. If you run a ping from
the DomU to another host the network stays alive but the high latency is
still there.
Further more if i try and uninstall the network driver i am unable to
use the old one (realtek) as it cannot detect the device.
2008 Jun 08
8
Windows GPLPV under OpenSolaris not working
On Sun, Jun 8, 2008 at 3:28 PM, James Harper <james.harper@bendigoit.com.au>
wrote:
> I''ve just uploaded 0.9.8 to http://www.meadowcourt.org/downloads
>
> As a reminder, the wiki page is
> http://wiki.xensource.com/xenwiki/XenWindowsGplPv
>
> This release fixes a bug which could be hit under high disk load (high
> number of outstanding requests with fragmented
2009 Feb 26
8
GPL PV TAP cow incompatible?
Hi
Is it true that the GPLPV drivers is incompatible with QCOW images?
The Windows DomU works fine with raw standard "file:/" images, and
paravirtualized Linux DomU works fine with QCOW and raw images.
If i use QCOW images with my Windows DomU''s i get a stop error:
0x00000007b (inaccessible boot device).
It looks like the blkback.3.hda process, that normally starts with
2008 Nov 16
3
detecting a ''block-configure'' via xenbus
How should I be detecting a block-configure operation done in Dom0 in a
HVM DomU?
Under Debian Lenny''s Xen 3.2.1, when I execute such a command, Dom0
writes to the backend values ''params'' and ''type'', twice, but doesn''t
update the ''sectors'' value to match the new device, or initiate a state
change. Given that, the best I can
2009 May 16
13
GPLPV uninstaller
I''ve just uploaded a couple of .bat files to
http://www.meadowcourt.org/downloads/gplpv_uninstall_bat.zip that will
manually (and forcibly) uninstall GPLPV from your computer, which seems
to be necessary before doing upgrades from some versions. Run the
uninstaller, reboot, then install. There is a 0.9.x and a 0.10.x
uninstaller, which differ slightly, but the basic steps performed are:
2009 May 16
13
GPLPV uninstaller
I''ve just uploaded a couple of .bat files to
http://www.meadowcourt.org/downloads/gplpv_uninstall_bat.zip that will
manually (and forcibly) uninstall GPLPV from your computer, which seems
to be necessary before doing upgrades from some versions. Run the
uninstaller, reboot, then install. There is a 0.9.x and a 0.10.x
uninstaller, which differ slightly, but the basic steps performed are:
2008 Apr 04
7
Linux TCP Checksum offload limitations
Some version of Windows appear to give the network adapter driver a
packet broken up into fairly small pieces, eg
Page 0: 14 bytes of Ethernet Header
Page 1: 20 bytes of IP Header
Page 2: 20 bytes of TCP Header
Page 3: 1460 bytes of TCP Data
When this happens, Linux appears to not pass the packets beyond the
vifX.Y interface - a tcpdump on (say) vif455.0 shows packets but a
tcpdump on eth0 does