Displaying 20 results from an estimated 100 matches similar to: "[PATCH 3/3] TLB flushing and IO memory mapping"
2012 May 25
0
[PATCH 3/3] gnttab: cleanup
- introduce local variables (shortcuts for frequently used <dom>->grant_table)
- adjust first parameter of mapcount()
- drop lock acquisition from gnttab_get_version()
- remove hard tabs and adjust formatting
Signed-off-by: Jan Beulich <jbeulich@suse.com>
Tested-by: Andrew Thomas <andrew.thomas@oracle.com>
--- a/xen/common/grant_table.c
+++ b/xen/common/grant_table.c
@@
2008 Nov 04
7
[PATCH 1/1] Xen PV support for hugepages
This is the latest version of a patch that adds hugepage support to the Xen
hypervisor in a PV environment. It is against the latest xen-unstable tree
on xenbits.xensource.com. I believe this version addresses the comments
made about the previous version of the patch.
Hugepage support must be enabled via the hypervisor command line option
"allowhugepage".
It assumes the guest
2008 Sep 11
5
[PATCH] Fix arguments passed to SHADOW_PRINTK
Hi.
When I compiled xen-3.3-testing with DEBUG_TRACE_DUMP in xen/include/xen/lib.h
turned on, I got some errors at SHADOW_PRINTK.
The following patch will fix the arguments passed to SHADOW_PRINTK in
xen/arch/x86/mm/shadow/common.c and xen/arch/x86/mm/shadow/multi.c.
Although I haven''t tested, it is the same for xen-unstable.hg
Thanks,
Yoshi
Signed-off-by: Yoshi Tamura
2008 Jul 24
2
[RFC] i386 highmem assist hypercalls
While looking at the origin of very frequently executed hypercalls (as
usual, kernel builds are what''s being measured), I realized that the high
page accessor functions in Linux would be good candidates to handle
in the hypervisor - clearing or copying to/from a high page is a pretty
frequent operation (provided there''s enough memory). However, the
measured results
2006 Sep 29
0
[PATCH 2/6] xen: add per-node bucks to page allocator
This patch adds a per-node bucket to the heap structure in Xen. During
heap initialization the patch determines which bucket to place the
memory. We reserve guard pages between node boundaries in the case that
said boundary isn''t already guarded by the MAX_ORDER boundary to prevent
the buddy allocator from merging pages between nodes.
--
Ryan Harper
Software Engineer; Linux Technology
2013 Nov 14
4
[PATCH] xen/arm: Allow balooning working with 1:1 memory mapping
With the lake of iommu, dom0 must have a 1:1 memory mapping for all
these guest physical address. When the ballon decides to give back a
page to the kernel, this page must have the same address as previously.
Otherwise, we will loose the 1:1 mapping and will break DMA-capable
device.
Signed-off-by: Julien Grall <julien.grall@linaro.org>
CC: Keir Fraser <keir@xen.org>
CC: Jan Beulich
2007 Nov 27
4
spurious warnings from get_page() via gnttab_copy() during frontend shutdown
In testing our implementation of the hypervisor copy based backend-
>frontend networking changes, we see what I believe are spurious
warning messages during the shutdown of the frontend domain:
(XEN) /export/build/schuster/xvm-gate/xen.hg/xen/include/asm/mm.h:
189:d0 Error pfn 30e290: rd=ffff830000fcf100, od=ffff830000fcf100,
caf=00000000, taf=0000000000000000
(XEN) Xen call trace:
2008 Oct 17
6
[PATCH, RFC] i386: highmem access assistance hypercalls
While looking at the origin of very frequently executed hypercalls I
realized that the high page accessor functions in Linux would be good
candidates to handle in the hypervisor - clearing or copying to/from
a high page is a pretty frequent operation (provided there''s enough
memory in the domain). While prior to the first submission I only
measured kernel builds (where the results are not
2006 Dec 01
1
[PATCH 2/10] Add support for netfront/netback acceleration drivers
This set of patches adds the support for acceleration plugins to the
netfront/netback drivers. These plugins are intended to support
virtualisable network hardware that can be directly accessed from the
guest, bypassing dom0.
This is in response to the RFC we posted to xen-devel with an outline
of our approach at the end of September.
To follow will be another set of patches to provide our
2007 Dec 10
0
[PATCH] avoid duplication of domain ID in messages
text doesn''t need to repeat that.
Signed-off-by: Jan Beulich <jbeulich@novell.com>
Index: 2007-12-10/xen/arch/x86/hvm/hvm.c
===================================================================
--- 2007-12-10.orig/xen/arch/x86/hvm/hvm.c 2007-12-10 09:19:12.000000000 +0100
+++ 2007-12-10/xen/arch/x86/hvm/hvm.c 2007-12-10 09:22:53.000000000 +0100
@@ -533,8 +533,7 @@ static void
2011 Sep 23
2
Some problems about xenpaging
Hi, Olaf
we have tested the xenpaging feature and found some problems.
(1) the test case like this : when we start a VM with POD enable, the xenpaging is started at the same time.
this case will cause many problems ,finally, we fixed the BUG, the patch is attached below.
(2) there is a very serious problem. we have observed many VM crash examples, the error code is not always the same.
2007 Jan 11
6
[PATCH 4/8] HVM save restore: vcpu context support
[PATCH 4/8] HVM save restore: vcpu context support
Signed-off-by: Zhai Edwin <edwin.zhai@intel.com>
save/restore HVM vcpu context such as vmcs
diff -r ee20d1905bde xen/arch/x86/domain.c
--- a/xen/arch/x86/domain.c Thu Jan 11 16:40:55 2007 +0800
+++ b/xen/arch/x86/domain.c Thu Jan 11 16:46:59 2007 +0800
@@ -573,6 +573,7 @@ int arch_set_info_guest(
else
{
2010 Jan 30
20
"Iomem mapping not permitted" during windows crash dump under GPLPV
I''ve recently noticed that my windows crash dumps fail at around 40-50%
under GPLPV. ''xm dmesg'' shows the following:
(XEN) grant_table.c:350:d0 Iomem mapping not permitted ffffffffffffffff
(domain 865)
At first I thought that the cause was just a bug in my grant ref code
but it just occurred to me that this could be happening when Windows
tries to write out the
2013 Dec 06
36
[V6 PATCH 0/7]: PVH dom0....
Hi,
V6: The only change from V5 is in patch #6:
- changed comment to reflect autoxlate
- removed a redundant ASSERT
- reworked logic a bit so that get_page_from_gfn() is called with NULL
for p2m type as before. arm has ASSERT wanting it to be NULL.
Tim: patch 4 needs your approval.
Daniel: patch 5 needs your approval.
These patches implement PVH dom0.
Patches 1 and 2
2011 Jan 17
8
[PATCH 0 of 3] Miscellaneous populate-on-demand bugs
This patch series includes a series of bugs related to p2m, ept, and
PoD code which were found as part of our XenServer product testing.
Each of these fixes actual bugs, and the 3.4-based version of the patch
has been tested thoroughly. (There may be bugs in porting the patches,
but most of them are simple enough as to make it unlikely.)
Each patch is conceptually independent, so they can each
2019 Oct 30
0
[PATCH v2 09/15] xen/gntdev: use mmu_range_notifier_insert
On 10/28/19 4:10 PM, Jason Gunthorpe wrote:
> From: Jason Gunthorpe <jgg at mellanox.com>
>
> gntdev simply wants to monitor a specific VMA for any notifier events,
> this can be done straightforwardly using mmu_range_notifier_insert() over
> the VMA's VA range.
>
> The notifier should be attached until the original VMA is destroyed.
>
> It is unclear if any of
2011 Nov 29
10
[PATCH 0 of 2] Fix correctness race in xc_mem_paging_prep
ging_prep ensures that an mfn is backing the paged-out gfn, and
transitions to the next state in the paging state machine for this page.
Foreign mappings of the gfn will now succeed. This is the key idea, as it
allows the pager to now map the gfn and fill in its contents.
Unfortunately, it also allows any other foreign mapper to map the gfn and read
its contents. This is particularly dangerous
2007 Mar 20
62
RFC: [0/2] Remove netloop by lazy copying in netback
Hi Keir:
These two patches remove the need for netloop by performing the
copying in netback and only if it is necessary. The rationale
is that most packets will be processed without delay allowing
them to be freed without copying at all. So instead of copying
every packet destined to dom0 we''ll only copy those that linger
longer than a specified amount of time (currently 0.5s).
As it
2011 Mar 19
1
[patch 2/2] xen-gntdev: unlock on error path in gntdev_mmap()
We should unlock here and also decrement the number of &map->users.
Signed-off-by: Dan Carpenter <error27 at gmail.com>
---
I'm not very familiar with this code. Please review carefully.
diff --git a/drivers/xen/gntdev.c b/drivers/xen/gntdev.c
index 017ce60..b0f9e8f 100644
--- a/drivers/xen/gntdev.c
+++ b/drivers/xen/gntdev.c
@@ -662,7 +662,7 @@ static int gntdev_mmap(struct
2011 Mar 19
1
[patch 2/2] xen-gntdev: unlock on error path in gntdev_mmap()
We should unlock here and also decrement the number of &map->users.
Signed-off-by: Dan Carpenter <error27 at gmail.com>
---
I'm not very familiar with this code. Please review carefully.
diff --git a/drivers/xen/gntdev.c b/drivers/xen/gntdev.c
index 017ce60..b0f9e8f 100644
--- a/drivers/xen/gntdev.c
+++ b/drivers/xen/gntdev.c
@@ -662,7 +662,7 @@ static int gntdev_mmap(struct