similar to: [PATCH 3/3] TLB flushing and IO memory mapping

Displaying 20 results from an estimated 100 matches similar to: "[PATCH 3/3] TLB flushing and IO memory mapping"

2012 May 25
0
[PATCH 3/3] gnttab: cleanup
- introduce local variables (shortcuts for frequently used <dom>->grant_table) - adjust first parameter of mapcount() - drop lock acquisition from gnttab_get_version() - remove hard tabs and adjust formatting Signed-off-by: Jan Beulich <jbeulich@suse.com> Tested-by: Andrew Thomas <andrew.thomas@oracle.com> --- a/xen/common/grant_table.c +++ b/xen/common/grant_table.c @@
2008 Nov 04
7
[PATCH 1/1] Xen PV support for hugepages
This is the latest version of a patch that adds hugepage support to the Xen hypervisor in a PV environment. It is against the latest xen-unstable tree on xenbits.xensource.com. I believe this version addresses the comments made about the previous version of the patch. Hugepage support must be enabled via the hypervisor command line option "allowhugepage". It assumes the guest
2008 Sep 11
5
[PATCH] Fix arguments passed to SHADOW_PRINTK
Hi. When I compiled xen-3.3-testing with DEBUG_TRACE_DUMP in xen/include/xen/lib.h turned on, I got some errors at SHADOW_PRINTK. The following patch will fix the arguments passed to SHADOW_PRINTK in xen/arch/x86/mm/shadow/common.c and xen/arch/x86/mm/shadow/multi.c. Although I haven''t tested, it is the same for xen-unstable.hg Thanks, Yoshi Signed-off-by: Yoshi Tamura
2008 Jul 24
2
[RFC] i386 highmem assist hypercalls
While looking at the origin of very frequently executed hypercalls (as usual, kernel builds are what''s being measured), I realized that the high page accessor functions in Linux would be good candidates to handle in the hypervisor - clearing or copying to/from a high page is a pretty frequent operation (provided there''s enough memory). However, the measured results
2006 Sep 29
0
[PATCH 2/6] xen: add per-node bucks to page allocator
This patch adds a per-node bucket to the heap structure in Xen. During heap initialization the patch determines which bucket to place the memory. We reserve guard pages between node boundaries in the case that said boundary isn''t already guarded by the MAX_ORDER boundary to prevent the buddy allocator from merging pages between nodes. -- Ryan Harper Software Engineer; Linux Technology
2013 Nov 14
4
[PATCH] xen/arm: Allow balooning working with 1:1 memory mapping
With the lake of iommu, dom0 must have a 1:1 memory mapping for all these guest physical address. When the ballon decides to give back a page to the kernel, this page must have the same address as previously. Otherwise, we will loose the 1:1 mapping and will break DMA-capable device. Signed-off-by: Julien Grall <julien.grall@linaro.org> CC: Keir Fraser <keir@xen.org> CC: Jan Beulich
2007 Nov 27
4
spurious warnings from get_page() via gnttab_copy() during frontend shutdown
In testing our implementation of the hypervisor copy based backend- >frontend networking changes, we see what I believe are spurious warning messages during the shutdown of the frontend domain: (XEN) /export/build/schuster/xvm-gate/xen.hg/xen/include/asm/mm.h: 189:d0 Error pfn 30e290: rd=ffff830000fcf100, od=ffff830000fcf100, caf=00000000, taf=0000000000000000 (XEN) Xen call trace:
2008 Oct 17
6
[PATCH, RFC] i386: highmem access assistance hypercalls
While looking at the origin of very frequently executed hypercalls I realized that the high page accessor functions in Linux would be good candidates to handle in the hypervisor - clearing or copying to/from a high page is a pretty frequent operation (provided there''s enough memory in the domain). While prior to the first submission I only measured kernel builds (where the results are not
2006 Dec 01
1
[PATCH 2/10] Add support for netfront/netback acceleration drivers
This set of patches adds the support for acceleration plugins to the netfront/netback drivers. These plugins are intended to support virtualisable network hardware that can be directly accessed from the guest, bypassing dom0. This is in response to the RFC we posted to xen-devel with an outline of our approach at the end of September. To follow will be another set of patches to provide our
2007 Dec 10
0
[PATCH] avoid duplication of domain ID in messages
text doesn''t need to repeat that. Signed-off-by: Jan Beulich <jbeulich@novell.com> Index: 2007-12-10/xen/arch/x86/hvm/hvm.c =================================================================== --- 2007-12-10.orig/xen/arch/x86/hvm/hvm.c 2007-12-10 09:19:12.000000000 +0100 +++ 2007-12-10/xen/arch/x86/hvm/hvm.c 2007-12-10 09:22:53.000000000 +0100 @@ -533,8 +533,7 @@ static void
2011 Sep 23
2
Some problems about xenpaging
Hi, Olaf we have tested the xenpaging feature and found some problems. (1) the test case like this : when we start a VM with POD enable, the xenpaging is started at the same time. this case will cause many problems ,finally, we fixed the BUG, the patch is attached below. (2) there is a very serious problem. we have observed many VM crash examples, the error code is not always the same.
2007 Jan 11
6
[PATCH 4/8] HVM save restore: vcpu context support
[PATCH 4/8] HVM save restore: vcpu context support Signed-off-by: Zhai Edwin <edwin.zhai@intel.com> save/restore HVM vcpu context such as vmcs diff -r ee20d1905bde xen/arch/x86/domain.c --- a/xen/arch/x86/domain.c Thu Jan 11 16:40:55 2007 +0800 +++ b/xen/arch/x86/domain.c Thu Jan 11 16:46:59 2007 +0800 @@ -573,6 +573,7 @@ int arch_set_info_guest( else {
2010 Jan 30
20
"Iomem mapping not permitted" during windows crash dump under GPLPV
I''ve recently noticed that my windows crash dumps fail at around 40-50% under GPLPV. ''xm dmesg'' shows the following: (XEN) grant_table.c:350:d0 Iomem mapping not permitted ffffffffffffffff (domain 865) At first I thought that the cause was just a bug in my grant ref code but it just occurred to me that this could be happening when Windows tries to write out the
2013 Dec 06
36
[V6 PATCH 0/7]: PVH dom0....
Hi, V6: The only change from V5 is in patch #6: - changed comment to reflect autoxlate - removed a redundant ASSERT - reworked logic a bit so that get_page_from_gfn() is called with NULL for p2m type as before. arm has ASSERT wanting it to be NULL. Tim: patch 4 needs your approval. Daniel: patch 5 needs your approval. These patches implement PVH dom0. Patches 1 and 2
2011 Jan 17
8
[PATCH 0 of 3] Miscellaneous populate-on-demand bugs
This patch series includes a series of bugs related to p2m, ept, and PoD code which were found as part of our XenServer product testing. Each of these fixes actual bugs, and the 3.4-based version of the patch has been tested thoroughly. (There may be bugs in porting the patches, but most of them are simple enough as to make it unlikely.) Each patch is conceptually independent, so they can each
2019 Oct 30
0
[PATCH v2 09/15] xen/gntdev: use mmu_range_notifier_insert
On 10/28/19 4:10 PM, Jason Gunthorpe wrote: > From: Jason Gunthorpe <jgg at mellanox.com> > > gntdev simply wants to monitor a specific VMA for any notifier events, > this can be done straightforwardly using mmu_range_notifier_insert() over > the VMA's VA range. > > The notifier should be attached until the original VMA is destroyed. > > It is unclear if any of
2011 Nov 29
10
[PATCH 0 of 2] Fix correctness race in xc_mem_paging_prep
ging_prep ensures that an mfn is backing the paged-out gfn, and transitions to the next state in the paging state machine for this page. Foreign mappings of the gfn will now succeed. This is the key idea, as it allows the pager to now map the gfn and fill in its contents. Unfortunately, it also allows any other foreign mapper to map the gfn and read its contents. This is particularly dangerous
2007 Mar 20
62
RFC: [0/2] Remove netloop by lazy copying in netback
Hi Keir: These two patches remove the need for netloop by performing the copying in netback and only if it is necessary. The rationale is that most packets will be processed without delay allowing them to be freed without copying at all. So instead of copying every packet destined to dom0 we''ll only copy those that linger longer than a specified amount of time (currently 0.5s). As it
2011 Mar 19
1
[patch 2/2] xen-gntdev: unlock on error path in gntdev_mmap()
We should unlock here and also decrement the number of &map->users. Signed-off-by: Dan Carpenter <error27 at gmail.com> --- I'm not very familiar with this code. Please review carefully. diff --git a/drivers/xen/gntdev.c b/drivers/xen/gntdev.c index 017ce60..b0f9e8f 100644 --- a/drivers/xen/gntdev.c +++ b/drivers/xen/gntdev.c @@ -662,7 +662,7 @@ static int gntdev_mmap(struct
2011 Mar 19
1
[patch 2/2] xen-gntdev: unlock on error path in gntdev_mmap()
We should unlock here and also decrement the number of &map->users. Signed-off-by: Dan Carpenter <error27 at gmail.com> --- I'm not very familiar with this code. Please review carefully. diff --git a/drivers/xen/gntdev.c b/drivers/xen/gntdev.c index 017ce60..b0f9e8f 100644 --- a/drivers/xen/gntdev.c +++ b/drivers/xen/gntdev.c @@ -662,7 +662,7 @@ static int gntdev_mmap(struct