similar to: VF passthrough problems

Displaying 20 results from an estimated 100 matches similar to: "VF passthrough problems"

2007 Jun 24
4
It seems the "machine check exception handling" breaks HVM guest
Hi Jan Beulich, > changeset 15414: 3cf5052ba5e5 x86: machine check exception handling With the c/s, when creating HVM guest, I can only see a white Qemu window. :( Can you have a look? Thanks. PS: the serial log follows: (XEN) HVM1: pci dev 02:0 bar 14 size 00001000: f2000000 (XEN) HVM1: pci dev 03:0 bar 10 size 00000100: 0000c101 (XEN) HVM1: pci dev 03:0 bar 14 size 01000000:
2007 Dec 27
2
VT-d and the GPU
Like others on this list I am trying to employ the VTD-NEO patches in Xen 3.2 unstable to assign the internal graphics device to Dom1/Vista. I have removed the Cirrus Logic emulated device from qemu and replaced the Cirrus Logic vgabios with the the actual vgabios from my GPU. However I am hitting an xen assert and was hoping someone might be able to point me in the right direction. Below
2014 Feb 03
2
4.3.0-3 fresh build - hvmloader Triple fault on VCPU0
When using the 4.3.0-3/4.3.0-3+b1 xen packages from the Debian archive, things work fine. However, when we compile the xen=4.3.0-3 source package ourselves, we run into an issue launching an HVM. As soon as the HVM bios is loaded, VCPU0 gets a triple fault and the HVM is destroyed (this is with "loglvl=all guest_loglvl=all hvm_debug=3" boot options). (XEN) HVM1: Multiprocessor
2013 Mar 12
14
vpmu=1 and running 'perf top' within a PVHVM guest eventually hangs dom0 and hypervisor has stuck vCPUS. Romley-EP (model=45, stepping=2)
This issue I am encountering seems to only happen on multi-socket machines. It also does not help that the only multi-socket box I have is an Romley-EP (so two socket SandyBridge CPUs). The other SandyBridge boxes I''ve (one socket) are not showing this. Granted they are also a different model (42). The problem is that when I run ''perf top'' within an SMP PVHVM guest,
2008 Jan 23
13
Xen 3.2 and Big Real Mode support?
Hello, I read in the announce of Xen 3.2.0 released that it has "preliminary support for a wider range of bootloaders in fully virtualised (HVM) guests, using full emulation of x86 ''real mode''" . I''d like to know what is the level of the emulation of x86 ''real mode''? I ask that because I tried to install OpenSuse 10.3 (I used the iso file
2011 Feb 03
1
pci-passthrough nic but no link
I hope someone can enlighten me. I have a quad-port Intel 82580 nic (igb driver) on my system and I''d like to dedicate each nic to a HVM via VT-d PCI passthrough. The IOMMU on my system seems to work, I can assign the PCI devices to my HVMs. The HVMs see the pci device and load their respective igb drivers, and ethtool -i eth0 works on each HVM, shows the drivers are loaded and the
2013 Jul 23
73
Bug: Limitation of <=2GB RAM in domU persists with 4.3.0
I just built 4.3.0 in order to get > 2GB of RAM in domU with GPU passthrough without crashes. Unfortunately, the same crashes still happen. Massive frame buffer corruption on domU before it locks up solid. It seems the PCI memory stomp is still happening. I am using qemu-dm, as I did on Xen 4.2.x. So whatever fix for this went into 4.3.0 didn''t fix it for me. Passing less than 2GB
2010 Aug 15
24
Xen patches merged to upstream Linux 2.6.36, plans for 2.6.37?
Hello, It looks like upstream linux-2.6.git contains at least the following xen related new features for Linux 2.6.36: - Xen-SWIOTLB support (required for Xen PCI passthru and dom0) - Xen PV-on-HVM drivers - Xen VBD online dynamic resize of guest disks (xvd*) Congratulations! What are the plans for 2.6.37 merge window? I believe at least: - Xen PCI frontend Others? I''m going to
2014 Feb 06
0
4.3.0-3 fresh build - hvmloader Triple fault on VCPU0
I've tracked down the issue to the seabios package being updated to 1.7.4-1 from 1.7.3-3. When Xen is built with the older version, everything works fine. In fact, I bisected the issue down to the following SeaBIOS commit: http://review.coreboot.org/gitweb?p=seabios.git;a=commit;h=e144bb7af49ca8756b7222a75811f3b85b0bc1f5 I don't know if this group is interested in tracking down the
2007 Jul 30
50
boot a existing windows in hvm domain
Hi all don''t get such envionment to try now. so ask for help here. imagine that two os installed. Xen3.1 + Linux is installed in /dev/hda and windows xp is installed in /dev/hdb I don''t want to reboot to use windows. Is it possiable to boot the hdb windows in xen? i saw there are some threads about P2V in this mail list. will the P2V tools help? thanks
2007 Jul 30
50
boot a existing windows in hvm domain
Hi all don''t get such envionment to try now. so ask for help here. imagine that two os installed. Xen3.1 + Linux is installed in /dev/hda and windows xp is installed in /dev/hdb I don''t want to reboot to use windows. Is it possiable to boot the hdb windows in xen? i saw there are some threads about P2V in this mail list. will the P2V tools help? thanks
2010 Nov 22
0
HVM error on booting
Hello everybody, I am trying to install an Ubuntu inside a lvm partition as a HVM. Here''s my config: % cat /etc/xen/hvm1 kernel = "/usr/lib/xen/boot/hvmloader" builder=''hvm'' memory = 1024 name = "hvm1" vif = [ ''bridge=xenbr0'' ] disk =
2013 Feb 22
5
Bug#701213: Hvm domU doesn't boot after dom0 reboot under Debian Wheezy
Package: xen-hypervisor-4.1-amd64 Version: 4.1.4-2 Hi, Hvm domU does not starts up afer dom0 reboot. Bridge br0 is set up manually. Dom0: Debian Wheezy 3.2.0-4-amd64 #1 SMP Debian 3.2.35-2 x86_64 GNU/Linux DomU: Debian Squeeze 2.6.32-5-amd64 #1 SMP Sun Sep 23 10:07:46 UTC 2012 x86_64 GNU/Linux config: ===<======== kernel = '/usr/lib/xen-4.1/boot/hvmloader' builder = 'hvm'
2009 Aug 28
64
[PATCH 2/2] graphics passthrough with VT-d
This patch supports basic gfx passthrough on QEMU: - disable emulated VGA adpater if there is passthroughed gfx - register/unregister legacy VGA I/O ports and MMIOs for passthroughed gfx Signed-off-by: Ben Lin <ben.y.lin@intel.com> Signed-off-by: Weidong Han <weidong.han@intel.com> _______________________________________________ Xen-devel mailing list
2006 Aug 13
4
using physical disks with HVM
Hi, As a previous thread in http://lists.xensource.com/archives/html/xen-users/2006-06/msg01111.html tells, it should be possible tun run hvm domains with a physical disk. There where other threads about this, but no solution for my current problem. When I try to use a disk defined as disk = [ ''phy:data/faitest-client-generic-hvm-disk,ioemu:hda,w''] (The device is a lvm
2013 Jan 08
0
Re: txftrans.c:4498 -- TXF_HARD_ERROR
Apologies for missing Subject. On Wed, Jan 9, 2013 at 3:21 AM, Rushikesh Jadhav <2rushikeshj@gmail.com>wrote: > Hello List, > > Please help me in diagnosing below error in /var/log/messages > > HVM1[16586]: txftrans.c:4498 -- TXF_HARD_ERROR (0xC01A002B) for RM: > fffff8800d099000 > > Im unable to find any google reference related to above C file or its > source
2014 Feb 06
5
Bug#737905: Xen: hvmloader causes immediate VCPU triple fault with SeaBIOS 1.7.4
Package: xen-utils-4.3 Version: 4.3.0-3 Severity: important Tags: upstream patch When using Xen compiled with seabios-1.7.4-1 (now in unstable/testing), HVMs fail immediately upon starting with a VCPU triple fault: (XEN) HVM1: Loading SeaBIOS ... (XEN) hvm.c:1253:d1 Triple fault on VCPU0 - invoking HVM shutdown action 1. This is because SeaBIOS can now be greater than 128k, which breaks an
2020 Jul 28
1
BUG: unable to handle page fault for address nouveau_fence_new
I've had several recent crashes of the nouveau kernel driver over the past month or so. My suspicion is that Firefox is causing it. The screen goes black and then the computer reboots. Nothing much in the syslogs, however I've managed to get netconsole output. It happens very infrequently and I'm afraid I don't know how to reproduce it, however I'll be more than happy to
2006 Jul 18
33
Paravirtualised drivers for fully virtualised domains
(The list appears to have eaten my previous attempt to send this. Apologies if you receive multiple copies.) The attached patches allow you to use paravirtualised network and block interfaces from fully virtualised domains, based on Intel''s patches from a few months ago. These are significantly faster than the equivalent ioemu devices, sometimes by more than an order of magnitude.
2004 Feb 06
1
Rsync 2.5.7 local buffer overflow
Hi folks, I've discovered a local stack overflow bug in rsync 2.5.7. Problem is in open_socket_out function (socket.c). ... char portbuf[10]; char *h; int proxied = 0; char buffer[1024]; char *cp; ... h = getenv("RSYNC_PROXY"); proxied = (h != NULL) && (*h != '\0'); ... if (proxied) {