search for: invalid_mfn

Displaying 20 results from an estimated 43 matches for "invalid_mfn".

2012 Mar 01
14
[PATCH 0 of 3] RFC Paging support for AMD NPT V2
There has been some progress, but still no joy. Definitely not intended for inclusion at this point. Tim, Wei, I added a Xen command line toggle to disable IOMMU and P2M table sharing. Tim, I verified that changes to p2m-pt.c don''t break shadow mode (64bit hypervisor and Win 7 guest). Hongkaixing, I incorporated your suggestion in patch 2, so I should add your Signed-off-by eventually.
2011 Sep 28
2
event channel in xenpaging
...n instead of IDC between dom0 and domU? 2)In your latest patch,[PATCH 9 of 9] xenpaging: watch the domains /xenpaging/num_pages xenstore value.I found some problems. a、In main(),you put the page_out process into while(1) and make the following change. if ( interrupted ) victims[i].gfn = INVALID_MFN; - else - evict_victim(paging, &victims[i], fd, i); I think the" if ( interrupted )" should remove here.Because once handling page_in requests,the related victims slot should clear, then in evict_pages we can populate new page in this slot because in evict_page() the condition...
2012 Oct 20
15
xen-unstable, winxp32 very poor performance on AMD FX-8150, I bisected and changeset is 24770:7f79475d3de7
I ran a bisect to find out when Windows XP 32 bit becomes unusably slow. And I found the changeset that caused it. ========== The problem: ========== Windows 8 64 bit and 32 bit run fast and fine in the newest xen versions. Windows XP 32 bit runs unusably slow in anything new that I built from xen-unstable, but runs fast in 4.1.2 and 4.1.3 stable. While it is running slow, "xm top" or
2011 Jul 18
1
Re: trip to shanghai
...t; > in your case? > > > Hi, > > I continue to invesgate this bug and find hypercall_mmu_update in > qemu_remap_bucket(xc_map_foreign_bulk) is failing: > > do_mmu_update > ->mod_l1_entry > -> if ( !p2m_is_ram(p2mt) || unlikely(mfn == INVALID_MFN) ) > return -EINVAL; > > mfn==INVALID_MFN, because : > > mod_l1_entry > ->gfn_to_mfn(p2m_get_hostp2m(pg_dom), l1e_get_pfn(nl1e), &p2mt)); > ->p2m->get_entry > ->p2m_gfn_to_mfn >...
2011 Jul 21
51
Linux Stubdom Problem
...gt; >> Hi, >> >>    I continue to invesgate this bug and find hypercall_mmu_update in >> qemu_remap_bucket(xc_map_foreign_bulk) is failing: >> >> do_mmu_update >>       ->mod_l1_entry >>              ->  if ( !p2m_is_ram(p2mt) || unlikely(mfn == INVALID_MFN) ) >>                          return -EINVAL; >> >>    mfn==INVALID_MFN, because : >> >> mod_l1_entry >>       ->gfn_to_mfn(p2m_get_hostp2m(pg_dom), l1e_get_pfn(nl1e), &p2mt)); >>               ->p2m->get_entry >>                        ...
2012 Jun 08
18
[PATCH 0 of 4 RFC] Populate-on-demand: Check pages being returned by the balloon driver
Populate-on-demand: Check pages being returned by the balloon driver This patch series is the second result of my work last summer on decreasing fragmentation of superpages in a guests'' p2m when using populate-on-demand. This patch series is against 4.1; I''m posting it to get feedback on the viability of getting a ported version of this patch into 4.2. As with the previous
2012 Jan 05
13
[PATCH] xenpaging:add a new array to speed up page-in in xenpaging
...- goto out; - } + i = page_out_index[req.gfn].index ; if ( req.flags & MEM_EVENT_FLAG_DROP_PAGE ) { @@ -733,7 +736,11 @@ if ( interrupted ) victims[i].gfn = INVALID_MFN; else + { evict_victim(paging, &victims[i], fd, i); + if( victims[i].gfn !=INVALID_MFN ) + page_out_index[victims[i].gfn].index = i; + } } else {...
2011 Sep 08
5
[PATCH 0 of 2] v2: memshare/xenpaging/xen-access fixes for xen-unstable
The following two patches allow the parallel use of memsharing, xenpaging and xen-access by using an independent ring buffer for each feature. Please review. v2: - update mem_event_check_ring arguments, check domain rather than domain_id - check ring_full first because its value was just evaluated - check if ring buffer is initialized before calling mem_access_domctl/mem_paging_domctl
2011 Sep 23
2
Some problems about xenpaging
...struct page_info *page; + /* alloc a new page to compensate the pod list */ + page = alloc_domheap_page(d, 0); + if ( unlikely(page == NULL) ) + { + goto out_entry_check; + } + set_p2m_entry(d, gpfn + i, _mfn(INVALID_MFN), 0, p2m_invalid); + p2m_mem_paging_drop_page(d, gpfn+i); + p2m_pod_cache_add(d, page, 0); + steal_for_cache = ( p2md->pod.entry_count > p2md->pod.count ); + nonpod--; + ram--; + } + /* for other ram types */ el...
2011 Nov 29
10
[PATCH 0 of 2] Fix correctness race in xc_mem_paging_prep
ging_prep ensures that an mfn is backing the paged-out gfn, and transitions to the next state in the paging state machine for this page. Foreign mappings of the gfn will now succeed. This is the key idea, as it allows the pager to now map the gfn and fill in its contents. Unfortunately, it also allows any other foreign mapper to map the gfn and read its contents. This is particularly dangerous
2013 Mar 21
27
[PATCH 0/4] xen/arm: guest SMP support
Hi all, this small patch series implement guest SMP support for ARM, using the ARM PSCI interface for secondary cpu bringup. Stefano Stabellini (4): xen/arm: basic PSCI support, implement cpu_on xen/arm: support for guest SGI xen/arm: support vcpu_op hypercalls xen: move VCPUOP_register_vcpu_info to common code xen/arch/arm/domain.c | 66 ++++++++++++++++++++++++
2009 Jan 31
2
Re: Debugging Xen via serial console
Hi, kdb: to debug xen hypervisor, could also debug guests gdbsx: to debug PV/HVM linux guests The tree is : http://xenbits.xensource.com/ext/debuggers.hg See README-dbg. You''ll need to setup serial access for kdb. Thanks, Mukesh > > Hi Dan, > > I''m currently using your version of ssplitd as it is. I haven''t tried > kdb. For some reason I
2007 Dec 05
13
[PATCH] unshadow the page table page which are used as data page
The patch deals with the situation which guest OS uses unused page table pages as data pages and write data to them. The pages will still be grabbed by Xen as page table pages, and lots of unnecessary page faults occur. The patch will check if the data guest writes to the page table contains valid mfn or not, if not, we believe it is a data page now and then unshadow the page. The patch
2012 Jan 05
3
[PATCH 0 of 2] xenpaging:speed up page-in
The following two patches are about how to speed up in xenpaging when page in pages. On suse11-64 with 4G memory,if we page out 2G pages,it will cost about 15.5 seconds, but take 2088 seconds to finish paging in.If page-in costs too much time,it will cause unmesurable problems when vm or dom0 access the paged_out page,such as BSOD,crash. What鈥檚 more,the dom0 is always in high I/O pressure.
2007 Jan 18
13
[PATCH 0/5] dump-core take 2:
The following dump-core patches changes its format into ELF, adds PFN-GMFN table, HVM support, and adds experimental IA64 support. - ELF format Program header and note section are adopted. - HVM domain support To know the memory area to dump, XENMEM_set_memory_map is added. XENMEM_memory_map hypercall is for current domain, so new one is created. and hvm domain builder tell xen its
2009 Aug 06
2
[PATCH] hvm emul: fix cmpxchg emulation to use an atomic operation
...MAPPING_EXCEPTION : + (mfn_x(mfn1) == READONLY_GFN) ? + MAPPING_SILENT_FAIL : MAPPING_UNHANDLEABLE); + + if ( likely(((vaddr + bytes - 1) & PAGE_MASK) == (vaddr & PAGE_MASK)) ) + { + /* Whole write fits on a single page */ + mfn2 = _mfn(INVALID_MFN); + map = map_domain_page(mfn_x(mfn1)) + (vaddr & ~PAGE_MASK); + } + else + { + /* Cross-page emulated writes are only supported for HVM guests; + * PV guests ought to know better */ + mfn2 = emulate_gva_to_mfn(v, (vaddr + bytes - 1) & PAGE_MASK); +...
2012 Dec 10
26
[PATCH 00/11] Add virtual EPT support Xen.
From: Zhang Xiantao <xiantao.zhang@intel.com> With virtual EPT support, L1 hyerpvisor can use EPT hardware for L2 guest''s memory virtualization. In this way, L2 guest''s performance can be improved sharply. According to our testing, some benchmarks can show > 5x performance gain. Signed-off-by: Zhang Xiantao <xiantao.zhang@intel.com> Zhang Xiantao (11):
2010 Jan 30
20
"Iomem mapping not permitted" during windows crash dump under GPLPV
I''ve recently noticed that my windows crash dumps fail at around 40-50% under GPLPV. ''xm dmesg'' shows the following: (XEN) grant_table.c:350:d0 Iomem mapping not permitted ffffffffffffffff (domain 865) At first I thought that the cause was just a bug in my grant ref code but it just occurred to me that this could be happening when Windows tries to write out the
2008 Jun 19
0
[PATCH] ia64/xen: introduce definitions necessary for ia64/xen hypercalls.
...upt */ +#define VIRQ_MCA_CPE VIRQ_ARCH_2 /* MCA cpe interrupt */ + +/* Maximum number of virtual CPUs in multi-processor guests. */ +/* keep sizeof(struct shared_page) <= PAGE_SIZE. + * this is checked in arch/ia64/xen/hypervisor.c. */ +#define MAX_VIRT_CPUS 64 + +#ifndef __ASSEMBLY__ + +#define INVALID_MFN (~0UL) + +union vac { + unsigned long value; + struct { + int a_int:1; + int a_from_int_cr:1; + int a_to_int_cr:1; + int a_from_psr:1; + int a_from_cpuid:1; + int a_cover:1; + int a_bsw:1; + long reserved:57; + }; +}; + +union vdc { + unsigned long value; + struct { + int d_vmsw:1; + int...
2008 Jun 19
0
[PATCH] ia64/xen: introduce definitions necessary for ia64/xen hypercalls.
...upt */ +#define VIRQ_MCA_CPE VIRQ_ARCH_2 /* MCA cpe interrupt */ + +/* Maximum number of virtual CPUs in multi-processor guests. */ +/* keep sizeof(struct shared_page) <= PAGE_SIZE. + * this is checked in arch/ia64/xen/hypervisor.c. */ +#define MAX_VIRT_CPUS 64 + +#ifndef __ASSEMBLY__ + +#define INVALID_MFN (~0UL) + +union vac { + unsigned long value; + struct { + int a_int:1; + int a_from_int_cr:1; + int a_to_int_cr:1; + int a_from_psr:1; + int a_from_cpuid:1; + int a_cover:1; + int a_bsw:1; + long reserved:57; + }; +}; + +union vdc { + unsigned long value; + struct { + int d_vmsw:1; + int...