similar to: RE: Assertion ''! IS_INVALID_M2P_ENTRY(gpfn)'' failedkillsmachine

Displaying 20 results from an estimated 1000 matches similar to: "RE: Assertion ''! IS_INVALID_M2P_ENTRY(gpfn)'' failedkillsmachine"

2006 Mar 29
0
Assertion ''! IS_INVALID_M2P_ENTRY(gpfn)'' failed kills machine
I''ve triggered this a few times now, so time to report. Unfortuantely, I don''t have a good recipe for reproducing this one, and this is a reasonably hacked up branch of (domU) XenLinux. So, while I can''t discount bugs in this domU, a domU bug shouldn''t crash Xen like this: (I belive this one was from simply typing halt in domU) (XEN) Assertion ''!
2008 Aug 21
2
doubt on releasing domain pages
Hi, I am trying to release domU pages from page_list and xenpage_list after domU shutdown while retaining the rest of the domain information. To achieve this in __domain_finalise_shutdown i call domain_relinquish_resources. This is failing to release pages from page_list for type PGT_l2_page_tables and crashing dom0. To be specific, while testing on mini-os i saw that when
2013 Dec 06
36
[V6 PATCH 0/7]: PVH dom0....
Hi, V6: The only change from V5 is in patch #6: - changed comment to reflect autoxlate - removed a redundant ASSERT - reworked logic a bit so that get_page_from_gfn() is called with NULL for p2m type as before. arm has ASSERT wanting it to be NULL. Tim: patch 4 needs your approval. Daniel: patch 5 needs your approval. These patches implement PVH dom0. Patches 1 and 2
2012 Jun 27
0
Re: [PATCH 2 of 4] xen, pod: Zero-check recently populated pages (checklast)
> # HG changeset patch > # User George Dunlap <george.dunlap@eu.citrix.com> > # Date 1340815812 -3600 > # Node ID ea827c449088a1017b6e5a9564eb33df70f8a9c6 > # Parent b4e1fec1c98f6cbad666a972f473854518c25500 > xen,pod: Zero-check recently populated pages (checklast) > > When demand-populating pages due to guest accesses, check recently > populated > pages to see
2007 Jan 08
6
Xen 3.0.4 - Ballooning
Hi, Maybe it''s a dumb question, but I''m actually trying to understand how the memory allocation works within Xen. I try to give 128MB to a domU and see if it increases for example when I "nano" a 500mb file, but the process just get killed when it reachs the 128MB memory limit. How do I configure the guests so they can ask for more memory until a limit is reached
2011 Sep 23
2
Some problems about xenpaging
Hi, Olaf we have tested the xenpaging feature and found some problems. (1) the test case like this : when we start a VM with POD enable, the xenpaging is started at the same time. this case will cause many problems ,finally, we fixed the BUG, the patch is attached below. (2) there is a very serious problem. we have observed many VM crash examples, the error code is not always the same.
2011 May 06
14
[PATCH 0 of 4] Use superpages on restore/migrate
This patch series restores the use of superpages when restoring or migrating a VM, while retaining efficient batching of 4k pages when superpages are not appropriate or available. Signed-off-by: George Dunlap <george.dunlap@eu.citrix.com> _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
2011 Jul 21
51
Linux Stubdom Problem
2011/7/19 Stefano Stabellini <stefano.stabellini@eu.citrix.com>: > CC''ing Tim and xen-devel > > On Mon, 18 Jul 2011, Jiageng Yu wrote: >> 2011/7/16 Stefano Stabellini <stefano.stabellini@eu.citrix.com>: >> > On Fri, 15 Jul 2011, Jiageng Yu wrote: >> >> 2011/7/15 Jiageng Yu <yujiageng734@gmail.com>: >> >> > 2011/7/15
2013 Sep 06
2
[PATCH] xen: arm: improve VMID allocation.
The VMID field is 8 bits. Rather than allowing only up to 256 VMs per host reboot before things start "acting strange" instead maintain a simple bitmap of used VMIDs and allocate them statically to guests upon creation. This limits us to 256 concurrent VMs which is a reasonable improvement. Eventually we will want a proper scheme to allocate VMIDs on context switch. The existing code
2012 Jun 08
18
[PATCH 0 of 4 RFC] Populate-on-demand: Check pages being returned by the balloon driver
Populate-on-demand: Check pages being returned by the balloon driver This patch series is the second result of my work last summer on decreasing fragmentation of superpages in a guests'' p2m when using populate-on-demand. This patch series is against 4.1; I''m posting it to get feedback on the viability of getting a ported version of this patch into 4.2. As with the previous
2007 Aug 28
6
[PATCH] Make XEN_DOMCTL_destroydomain hypercall continuable.
# HG changeset patch # User yamahata@valinux.co.jp # Date 1188274001 -32400 # Node ID 2c9db26f1d0e0fdd4757d76a67f4b37ba0e40351 # Parent 58d131f1fb35977ff2d8682f553391c8a866d52c Make XEN_DOMCTL_destroydomain hypercall continuable. XEN_DOMCTL_destroydomain hypercall frees domain resources, especially it frees all pages of the domain. When domain memory is very large, it takes too long resulting in
2007 Jul 02
3
Walking an HVM''s shadow page tables and other memory management questions.
Hello, I''m new to Xen and especially to the hypervisor code. I''m working off a 3.0.4.1 base and have the following questions regarding the memory management code for an x86, 32-bit platform (capable of supporting PAE). I''m doing some research into providing grant table hypercall support from a Windows 2003 HVM. I have made all the necessary changes to allow the
2006 Feb 09
11
X86_64 "assert" when booting 64-bit image.
I can boot a 64-bit HVM image on my SVM-system, and it goes quite a long way before it crashes (due to not finding the hard-disk image and thus no root FS - which I think may well be the configuration file). But if I set debug=y in the Rules.mk, I don''t get very far at all - it crashes well before it gets to any user-mode (Ring3) - in fact, I think it''s crashing pretty much as
2012 Mar 09
4
Max gfn?
I have a 12GB 64-bit Linux HVM guest (CentOS 5.5). When I look at the e820 map in the guest, I see the following: BIOS-provided physical RAM map: BIOS-e820: 0000000000010000 - 000000000009e000 (usable) BIOS-e820: 000000000009e000 - 00000000000a0000 (reserved) BIOS-e820: 00000000000e0000 - 0000000000100000 (reserved) BIOS-e820: 0000000000100000 - 00000000f0000000 (usable) BIOS-e820:
2007 Jan 18
13
[PATCH 0/5] dump-core take 2:
The following dump-core patches changes its format into ELF, adds PFN-GMFN table, HVM support, and adds experimental IA64 support. - ELF format Program header and note section are adopted. - HVM domain support To know the memory area to dump, XENMEM_set_memory_map is added. XENMEM_memory_map hypercall is for current domain, so new one is created. and hvm domain builder tell xen its
2013 Nov 06
0
[PATCH v5 5/6] xen/arm: Implement hypercall for dirty page tracing
Add hypercall (shadow op: enable/disable and clean/peek dirtied page bitmap). It consists of two parts: dirty page detecting and saving. For detecting, we setup the guest p2m''s leaf PTE read-only and whenever the guest tries to write something, permission fault happens and traps into xen. The permission-faulted GPA should be saved for the toolstack (when it wants to see which pages
2012 Nov 15
1
[RFC/PATCH v4] XENMEM_claim_pages (subop of existing) hypercall
This is a fourth cut of the hypervisor patch of the proposed XENMEM_claim_pages hypercall/subop, taking into account feedback from Jan and Keir and IanC, plus some fixes found via runtime debugging (using privcmd only) and some added comments/cleanup. [Logistical note: I will be out tomorrow (Friday) plus US holidays next week so my responsiveness to comments may be slower for awhile. --djm]
2007 Mar 20
62
RFC: [0/2] Remove netloop by lazy copying in netback
Hi Keir: These two patches remove the need for netloop by performing the copying in netback and only if it is necessary. The rationale is that most packets will be processed without delay allowing them to be freed without copying at all. So instead of copying every packet destined to dom0 we''ll only copy those that linger longer than a specified amount of time (currently 0.5s). As it
2012 Oct 04
49
[RFC 00/14] arm: implement ballooning and privcmd foreign mappings based on x86 PVH
This series implements ballooning for Xen on ARM and builds and Mukesh''s PVH privcmd stuff to implement foreign page mapping on ARM, replacing the old "HACK: initial (very hacky) XENMAPSPACE_gmfn_foreign" patch. The baseline is a bit complex, it is basically Stefano''s xenarm-forlinus branch (commit bbd6eb29214e) merged with Konrad''s linux-next-pvh branch
2008 May 09
14
[PATCH] patch to support super page (2M) with EPT
Attached are the patches to support super page with EPT. We only support 2M size. And shadow may still work fine with 4K pages. The patches can be split into 3 parts. Apply order is as attached. tool.diff To allocate 2M physical contiguous memory in guest except the first 2M and the last 2M. The first 2M covers special memory, and Xen use the last few pages in guest memory to do special