similar to: [xen-unstable bisection] complete test-amd64-i386-rhel6hvm-amd

Displaying 20 results from an estimated 300 matches similar to: "[xen-unstable bisection] complete test-amd64-i386-rhel6hvm-amd"

2012 Feb 24
0
[xen-unstable test] 12043: regressions - FAIL
flight 12043 xen-unstable real [real] http://www.chiark.greenend.org.uk/~xensrcts/logs/12043/ Regressions :-( Tests which did not succeed and are blocking, including tests which could not be run: test-amd64-i386-rhel6hvm-amd 7 redhat-install fail REGR. vs. 12031 test-amd64-i386-rhel6hvm-intel 7 redhat-install fail REGR. vs. 12031 test-amd64-i386-xl-winxpsp3-vcpus1 7
2012 Nov 12
10
Hypervisor to dom0 communication
Hello, I''m interested in establishing a communication channel between the Xen hypervisor and a dom0 userspace application. Ideally this would be a 2-way channel, with the hypervisor asynchronously notifying this application, and then (for certain types of messages) wait for a reply from the userspace consumer. To this end, I''ve been reading xen-devel posts [1], read the
2012 Sep 19
27
[PATCH] Persistent grant maps for xen blk drivers
This patch implements persistent grants for the xen-blk{front,back} mechanism. The effect of this change is to reduce the number of unmap operations performed, since they cause a (costly) TLB shootdown. This allows the I/O performance to scale better when a large number of VMs are performing I/O. Previously, the blkfront driver was supplied a bvec[] from the request queue. This was granted to
2013 Feb 28
1
[PATCH RFC 09/12] xen-blkback: move pending handles list from blkbk to pending_req
Moving grant ref handles from blkbk to pending_req will allow us to get rid of the shared blkbk structure. Signed-off-by: Roger Pau Monné <roger.pau@citrix.com> Cc: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> Cc: xen-devel@lists.xen.org --- drivers/block/xen-blkback/blkback.c | 16 ++++------------ 1 files changed, 4 insertions(+), 12 deletions(-) diff --git
2010 Aug 12
59
[PATCH 00/15] RFC xen device model support
Hi all, this is the long awaited patch series to add xen device model support in qemu; the main author is Anthony Perard. Developing this series we tried to come up with the cleanest possible solution from the qemu point of view, limiting the amount of changes to common code as much as possible. The end result still requires a couple of hooks in piix_pci but overall the impact should be very
2013 Dec 04
5
qemu-xen-dir + PCI passthrough = BOOM
Hey, I just started noticing it today - with qemu-xen (tip is commit b97307ecaad98360f41ea36cd9674ef810c4f8cf xen_disk: mark ioreq as mapped before unmapping in error case) when I try to pass in a PCI device at bootup it blows up with: char device redirected to /dev/pts/2 (label serial0) qemu: hardware error: xen: failed to populate ram at 40050000 CPU #0: EAX=00000000 EBX=00000000
2011 Sep 01
9
[PATCH V4 0/3] xen-blkfront/blkback discard support
Dear list, This is the V4 of the trim support for xen-blkfront/blkback, Now we move BLKIF_OP_TRIM to BLKIF_OP_DISCARD, and dropped all "trim" stuffs in the patches, and use "discard" instead. Also we updated the helpers of blkif_x86_{32|64}_request or we will meet problems using a non-native protocol. And this patch has been tested with both SSD and raw file, with SSD we will
2006 Aug 24
1
block ring interface: nr_segments = 0 results in BLKIF_RSP_ERROR
I am currently developing a blkfront.c for a custom OS over Xen 3.0.2-2. Typical I/O is working, however, I ran into an error while testing a corner case. On standard I/O, where { 1 <= nr_segments < BLKIF_MAX_SEGMENTS_PER_REQUEST } blkif_int()''s bret->status returns BLKIF_RSP_OKAY. Yet when { nr_segments == 0 } blkif_int''s bret->status is non-zero. (Yes
2006 Jun 21
15
[PATCH] patch to buffer write ioreq
This patch implements a new mechanism to buffer the MMIO_WRITE and PIO_WRITE packet in a queue, and then submit them to qemu-dm in batch. This could save a lot of domain switch time. We implement this as a generic one to do ioreq in batch. However, we only use it to batch the MMIO_WIRTE request for stdvga(0xA0000-0xC0000) in order to solve the win2k boot issue for hvm domain. Without this patch,
2013 May 08
11
[PATCH 1/2] xen, libxc: init msix addr/data with value from qemu via hypercall
Accelerated msix entry is initialized to zero when msixtbl_pt_register is called. This doesn''t match the value from qemu side, although pirq may already be mapped and binded in qemu side. Kernel will get wrong value when reading msix info. Signed-off-by: Zhenzhong Duan <zhenzhong.duan@oracle.com> Tested-by: Yuval Shaia <yuval.shaia@oracle.com> --- tools/libxc/xc_domain.c
2012 Jan 12
4
[PATCH] qemu-dm: add command to flush buffer cache
Add support for a xenstore dm command to flush qemu''s buffer cache. qemu will just keep mapping pages and not release them, which causes problems for the memory pager (since the page is mapped, it won''t get paged out). When the pager has trouble finding a page to page out, it asks qemu to flush its buffer, which releases all the page mappings. This makes it possible to find
2006 Aug 08
11
architecture-specific stuff in xend
Hi Ewan, I''m almost ready to integrate some PPC-specific stuff into xend, and I was wondering if you had a plan for how that should work. First example: the device tree data structure we talked about a few weeks ago. We will need to pass the config data to PPC code, probably in XendDomainInfo.initDomain(), and then pass the resulting data structure into libxc''s xc_linux_load()
2007 Mar 24
5
memsize for HVM save/restore
As you know, HVM save/restore broke recently because restored config miss guest memsize that used by xc_hvm_restore to locate some pfn. After discussion, we decided to remove the pfn deduction logic from restore side by adding a general memory layout. I have a patch for it. But then qemu broke, because it also require the memsize to locate the share page. We can''t use the previous
2006 Oct 23
1
[PATCH] Move ACPI timer to HV
Keir, The patch moves ACPI timer to HV, which is the last platform timer in Qemu. We found Vista polls ACPI timer very frequently (about 15 times averagely) when handling timer (RTC @ 64HZ) interrupt routine. Though the exact reason is known, it should be related to system time adjustment. When it''s in Qemu, the overhead is big. After moving, Vista''s idle overhead decreases
2006 May 17
20
X86_emulate to be moved into qemu...
> Hmmm... wouldn''t it be nice if we didn''t have a bespoke, > buggy & incomplete emulator for hvm mmio. ;-) In the process of solving this one, I''ve got to (another) of the blocking points... At the moment, we translate the virtual address from the guest in the page-fault-handler to guest physical address via the gva_to_gpa() function in svm.c/vmx.c:
2007 Apr 13
18
A different probklem with save/restore on C/S 14823.
I''m not seeing the problem that Fan Zhao is reporting, instead I get this one. Not sure if ti''s the same one or a different problem... This happens with my simple-guest [i.e. not using hvmloader, as I described before]. This worked fine yesterday. (XEN) event_channel.c:178:d0 EVTCHNOP failure: domain 0, error -22, line 178 (XEN) bad shared page: 0 (XEN) domain_crash_sync called
2012 Nov 02
2
[PATCH] xen-blk: persistent-grants fixes
This patch contains fixes for persistent grants implementation v2: * handle == 0 is a valid handle, so initialize grants in blkback setting the handle to BLKBACK_INVALID_HANDLE instead of 0. Reported by Konrad Rzeszutek Wilk. * new_map is a boolean, use "true" or "false" instead of 1 and 0. Reported by Konrad Rzeszutek Wilk. * blkfront announces the
2012 Oct 24
5
[PATCH v3] IOMMU: keep disabled until iommu_setup() is called
The iommu is enabled by default when xen is booting and later disabled in iommu_setup() when no iommu is present. But under some circumstances iommu code can be called before iommu_setup() is processed. If there is no iommu available xen crashes. This can happen for example when panic(...) is called as introduced with the patch "x86-64: detect processors subject to AMD erratum #121 and
2008 Mar 10
12
[RFC][PATCH] Use ioemu block drivers through blktap
When I submitted the qcow2 patch for blktap, suggestions came up that the qemu block drivers should be used also for blktap to eliminate the current code duplication in ioemu and blktap. The attached patch adds support for a tap:ioemu pseudo driver. Devices using this driver won''t use tapdisk (containing the code duplication) any more, but will connect to the qemu-dm of the domain. In
2012 May 14
1
[PATCH, v2] qemu/xendisk: properly update stats in ioreq_release()
While for the "normal" case (called from blk_send_response_all()) decrementing requests_finished is correct, doing so in the parse error case is wrong; requests_inflight needs to be decremented instead. Change in v2: Adjust coding style. Signed-off-by: Jan Beulich <jbeulich@suse.com> --- a/hw/xen_disk.c +++ b/hw/xen_disk.c @@ -154,7 +154,7 @@ static void ioreq_finish(struct