Displaying 20 results from an estimated 1000 matches similar to: "[Xen-ia64-devel] RE: Re: [PATCH] Disable auto-balloon on ia64"
2012 Nov 15
1
[RFC/PATCH v4] XENMEM_claim_pages (subop of existing) hypercall
This is a fourth cut of the hypervisor patch of the proposed
XENMEM_claim_pages hypercall/subop, taking into account
feedback from Jan and Keir and IanC, plus some fixes found
via runtime debugging (using privcmd only) and some
added comments/cleanup.
[Logistical note: I will be out tomorrow (Friday) plus US
holidays next week so my responsiveness to comments may be
slower for awhile. --djm]
2013 Nov 18
6
[PATCH RFC v2] pvh: clearly specify used parameters in vcpu_guest_context
The aim of this patch is to define a stable way in which PVH is
going to do AP bringup.
Since we are running inside of a HVM container, PVH should only need
to set flags, cr3 and user_regs in order to bring up a vCPU, the rest
can be set once the vCPU is started using the bare metal methods.
Additionally, the guest can also set cr0 and cr4, and those values
will be appended to the default values
2013 Nov 14
4
[PATCH] xen/arm: Allow balooning working with 1:1 memory mapping
With the lake of iommu, dom0 must have a 1:1 memory mapping for all
these guest physical address. When the ballon decides to give back a
page to the kernel, this page must have the same address as previously.
Otherwise, we will loose the 1:1 mapping and will break DMA-capable
device.
Signed-off-by: Julien Grall <julien.grall@linaro.org>
CC: Keir Fraser <keir@xen.org>
CC: Jan Beulich
2012 Oct 17
1
[PATCH 1/6] xen: balloon: allow PVMMU interfaces to be compiled out
The ARM platform has no concept of PVMMU and therefor no
HYPERVISOR_update_va_mapping et al. Allow this code to be compiled out
when not required.
In some similar situations (e.g. P2M) we have defined dummy functions
to avoid this, however I think we can/should draw the line at dummying
out actual hypercalls.
Signed-off-by: Ian Campbell <ian.campbell@citrix.com>
---
arch/x86/xen/Kconfig
2012 Mar 09
4
Max gfn?
I have a 12GB 64-bit Linux HVM guest (CentOS 5.5). When I look at the
e820 map in the guest, I see the following:
BIOS-provided physical RAM map:
BIOS-e820: 0000000000010000 - 000000000009e000 (usable)
BIOS-e820: 000000000009e000 - 00000000000a0000 (reserved)
BIOS-e820: 00000000000e0000 - 0000000000100000 (reserved)
BIOS-e820: 0000000000100000 - 00000000f0000000 (usable)
BIOS-e820:
2012 Mar 15
3
[PATCH] arm: allocate top level p2m page for all non-idle VCPUs
Not just dom0.
Signed-off-by: Ian Campbell <ian.campbell@citrix.com>
---
xen/arch/arm/domain.c | 3 +++
xen/arch/arm/domain_build.c | 3 ---
xen/arch/arm/p2m.c | 2 +-
3 files changed, 4 insertions(+), 4 deletions(-)
diff --git a/xen/arch/arm/domain.c b/xen/arch/arm/domain.c
index 5702399..4b38790 100644
--- a/xen/arch/arm/domain.c
+++ b/xen/arch/arm/domain.c
@@
2007 Aug 28
6
[PATCH] Make XEN_DOMCTL_destroydomain hypercall continuable.
# HG changeset patch
# User yamahata@valinux.co.jp
# Date 1188274001 -32400
# Node ID 2c9db26f1d0e0fdd4757d76a67f4b37ba0e40351
# Parent 58d131f1fb35977ff2d8682f553391c8a866d52c
Make XEN_DOMCTL_destroydomain hypercall continuable.
XEN_DOMCTL_destroydomain hypercall frees domain resources, especially
it frees all pages of the domain.
When domain memory is very large, it takes too long resulting in
2012 Jan 05
13
[PATCH] xenpaging:add a new array to speed up page-in in xenpaging
# HG changeset patch
# User hongkaixing<hongkaixing@huawei.com>
# Date 1325149704 -28800
# Node ID 052727b8165ce6e05002184ae894096214c8b537
# Parent 54a5e994a241a506900ee0e197bb42e5f1d8e759
xenpaging:add a new array to speed up page-in in xenpaging
This patch adds a new array named page_out_index to reserve the victim''s index.
When page in a page,it has to go through a for loop
2013 Jun 11
21
[PATCH] xen: fix initialization of wallclock time for PVHVM on migration
The initial values of the wallclock time in the shared info page are
set for PVHVM guests when the hypercall page is initialized, since the
hypercall page is not reinitialized on resume, the hypervisor
wallclock time is not properly set on resume.
Fix it by forcing an update of the wallclock values when the shared
info page is mapped.
Signed-off-by: Roger Pau Monné <roger.pau@citrix.com>
2007 Jul 20
0
[Xen-ia64-devel] [IA64] Weekly benchmark results [ww29]
Hi,
I report a benchmark result of this week on IPF using
ia64/xen-unstable and ia64/linux-2.6.18-xen.
Building the xen-vnif PV-on-HVM driver failed. This issue has already
fixed on staging trees by the following patches:
http://xenbits.xensource.com/staging/xen-unstable.hg?rev/48c8244c47c7
http://xenbits.xensource.com/staging/linux-2.6.18-xen.hg?rev/670f8d5305d2
In LTP, test cases related
2005 Dec 16
3
[PATCH] 0/7 xen: Add basic NUMA support
The patchset will add basic NUMA support to Xen (hypervisor only). We
borrowed from Linux support for NUMA SRAT table parsing, discontiguous
memory tracking (mem chunks), and cpu support (node_to_cpumask etc).
The hypervisor parses the SRAT tables and constructs mappings for each
node such as node to cpu mappings and memory range to node mappings.
Using this information, we also modified the
2012 Nov 29
14
Mem_event API and MEM_EVENT_REASON_SINGLESTEP
Hello,
I''m interested in using the mem_event API (or it''s LibVMI wrapper), but
I''m interested in capturing write events anywhere in a domU guest''s
memory, _and_ not just once for each page write, but every time a page
is being written to.
To this end, I''ve looked at the xen-access.c example, where at first all
the pages are being monitored:
2010 Aug 06
5
[PATCH] GSoC 2010 - Memory hotplug support for Xen guests - second fully working version - once again
Hi,
I am sending this e-mail once again because it probably
has been lost in abyss of Xen-devel/LKLM list.
Here is the second version of memory hotplug support
for Xen guests patch. This one cleanly applies to
git://git.kernel.org/pub/scm/linux/kernel/git/jeremy/xen.git
repository, xen/memory-hotplug head.
Changes:
- /sys/devices/system/memory/probe interface has been removed;
2012 Jan 05
3
[PATCH 0 of 2] xenpaging:speed up page-in
The following two patches are about how to speed up in xenpaging when page in pages.
On suse11-64 with 4G memory,if we page out 2G pages,it will cost about 15.5 seconds,
but take 2088 seconds to finish paging in.If page-in costs too much time,it will cause
unmesurable problems when vm or dom0 access the paged_out page,such as BSOD,crash.
What鈥檚 more,the dom0 is always in high I/O pressure.
2008 Dec 12
0
[Xen-ia64-devel] [IA64] Weekly benchmark results [ww49]
Hi,
I report a benchmark result of this week on IPF using
ia64/xen-unstable and ia64/linux-2.6.18-xen.
All test cases passed.
TEST ENVIRONMENT
Machine : Tiger4
Kernel : 2.6.18.8-xen
Changeset : 18872:3acca92b9597 (ia64/xen-unstable)
753:17adc5c344fe (ia64/linux-2.6.18-xen)
131:9210bd27a551 (efi-vfirmware)
b4d410a
2008 Feb 29
0
[Xen-ia64-devel] [IA64] Weekly benchmark results [ww08]
Hi,
I report a benchmark result of this week on IPF using
ia64/xen-unstable and ia64/linux-2.6.18-xen.
All test cases passed.
TEST ENVIRONMENT
Machine : Tiger4
Kernel : 2.6.18.8-xen
Changeset : 17133:daf39fc8038a (ia64/xen-unstable)
442:0c1e6a1b8e90 (ia64/linux-2.6.18-xen)
70:b6ea2d6130fb (efi-vfirmware)
Dom0 OS : RHEL4
2007 Apr 13
0
[Xen-ia64-devel] [IA64] Weekly benchmark results [ww15]
Hi,
I report a benchmark result of this week on IPF using
xen-ia64-unstable.
I did not test a VT-i domain because compiling PV-on-HVM drivers failed.
In LTP, test cases related with direct IO still failed.
# dio02 dio03 dio06 dio08 dio09 dio12 dio14 dio15 dio17 dio18 dio19
# dio21 dio22 dio23 dio25 dio26 dio27 dio28
And syslog10 failed but passed manually.
TEST ENVIRONMENT
Machine
2008 Oct 03
0
[IA64] Weekly benchmark results [ww39]
Hi,
I report a benchmark result of this week on IPF using
ia64/xen-unstable and ia64/linux-2.6.18-xen.
All test cases passed.
TEST ENVIRONMENT
Machine : Tiger4
Kernel : 2.6.18.8-xen
Changeset : 18503:325904748847 (ia64/xen-unstable)
673:3161879fdf22 (ia64/linux-2.6.18-xen)
126:d255b2cfdd5a (efi-vfirmware)
0f3d638
2007 Dec 07
0
[IA64] Weekly benchmark results [ww49]
Hi,
I report a benchmark result of this week on IPF using
ia64/xen-unstable and ia64/linux-2.6.18-xen.
All test cases passed.
TEST ENVIRONMENT
Machine : Tiger4
Kernel : 2.6.18.8-xen
Changeset : 16540:8ba08f2244b2 (ia64/xen-unstable)
337:4108b5c64f86 (ia64/linux-2.6.18-xen)
35:ebf7052731ec (efi-vfirmware)
Dom0 OS : RHEL4
2007 Mar 02
0
[Xen-ia64-devel] [IA64] Weekly benchmark results [ww09]
Hi,
I report a benchmark result of this week on IPF using
xen-ia64-unstable.
I did not test a VT-i domain because linking PV-on-HVM drivers still
failed.
In LTP, test cases related with direct IO still failed.
# dio02 dio03 dio06 dio08 dio09 dio12 dio14 dio15 dio17 dio18 dio19
# dio21 dio22 dio23 dio25 dio26 dio27 dio28
TEST ENVIRONMENT
Machine : Tiger4
Kernel :