Displaying 20 results from an estimated 100 matches similar to: "[PATCH V2] mem_event: Add support for MEM_EVENT_REASON_MSR"
2013 Jan 03
2
[PATCH V4] mem_event: Add support for MEM_EVENT_REASON_MSR
Add the new MEM_EVENT_REASON_MSR event type. Works similarly
to the other register events, except event.gla always contains
the MSR address (in addition to event.gfn, which holds the value).
MEM_EVENT_REASON_MSR does not honour the HVMPME_onchangeonly bit,
as doing so would complicate the hvm_msr_write_intercept()
switch-based handling of writes for different MSR addresses,
with little added
2012 Nov 29
14
Mem_event API and MEM_EVENT_REASON_SINGLESTEP
Hello,
I''m interested in using the mem_event API (or it''s LibVMI wrapper), but
I''m interested in capturing write events anywhere in a domU guest''s
memory, _and_ not just once for each page write, but every time a page
is being written to.
To this end, I''ve looked at the xen-access.c example, where at first all
the pages are being monitored:
2013 Feb 25
1
xc_domain_hvm_getcontext_partial(... CPU ...), the RIP register, and mem_event
Hello,
assuming that the VCPU is paused while waiting for userspace to respond
to a page fault mem_event, and userspace calls libxc''s
xc_domain_hvm_getcontext_partial(... CPU ...) and tries to inspect RIP,
is this a safe thing to do? Is RIP''s value guaranteed to be safe to
inspect in this context?
Thanks,
Razvan Cojocaru
2012 Apr 24
1
[PATCH] [resend] xen-access: Check return values and clean up on errors during init
Check the return values of the libxc mem_access calls.
Free allocated structures (platform_info, domain_info) on errors during initialization and exit.
Unbind VIRQ, close event channel and connection to Xen on errors during initialization
Signed-off-by: Aravindh Puthiyaparambil <aravindh@virtuata.com>
Acked-by: Ian Jackson <ian.jackson@eu.citrix.com>
diff -r ea47068fa7a2 -r
2012 Sep 10
10
[PATCH] mem_event: fix regression affecting CR3, CR4 memory events
This is a patch repairing a regression in code previously functional in 4.1.x. It appears that, during some refactoring work, calls to hvm_memory_event_cr3 and hvm_memory_event_cr4 were lost.
These functions were originally called in mov_to_cr() of vmx.c, but the commit http://xenbits.xen.org/hg/xen-unstable.hg/rev/1276926e3795 abstracted the original code into generic functions up a level in
2011 Sep 08
5
[PATCH 0 of 2] v2: memshare/xenpaging/xen-access fixes for xen-unstable
The following two patches allow the parallel use of memsharing, xenpaging and
xen-access by using an independent ring buffer for each feature.
Please review.
v2:
- update mem_event_check_ring arguments, check domain rather than domain_id
- check ring_full first because its value was just evaluated
- check if ring buffer is initialized before calling
mem_access_domctl/mem_paging_domctl
2012 Mar 30
3
[PATCH] xenpaging: add error code to indicate iommem passthrough
# HG changeset patch
# User Olaf Hering <olaf@aepfle.de>
# Date 1333120147 -7200
# Node ID 34d9828185501f6e7ea2c3c2a397176a8e54ef0a
# Parent 80653c8ea1d19dfe6130396bbc77f766eb9f9fab
xenpaging: add error code to indicate iommem passthrough
Similar to the existing ENODEV and EXDEV error codes, add EMDEV to
indicate that iommu passthrough is not compatible with paging.
All error codes are
2012 Mar 01
14
[PATCH 0 of 3] RFC Paging support for AMD NPT V2
There has been some progress, but still no joy. Definitely not intended for
inclusion at this point.
Tim, Wei, I added a Xen command line toggle to disable IOMMU and P2M table
sharing.
Tim, I verified that changes to p2m-pt.c don''t break shadow mode (64bit
hypervisor and Win 7 guest).
Hongkaixing, I incorporated your suggestion in patch 2, so I should add your
Signed-off-by eventually.
2012 Feb 17
1
[PATCH] x86/mm: Make sure the event channel is released accurately
# HG changeset patch
# User h00166998@h00166998.china.huawei.com
# Date 1329462865 -28800
# Node ID 9fd12f919ddbd15927117eff42149664dba698ca
# Parent b75664e5390583c5d2075c82a14245bc941b3aaf
x86/mm: Make sure the event channel is released accurately
In xenpaging source code,there is an interdomain communication between dom0
and domU. In mem_event_enable(),the function
2013 Aug 16
7
[PATCH v2] xen/console: buffer and show origin of guest PV writes
Guests other than domain 0 using the console output have previously been
controlled by the VERBOSE define, but with no designation of which
guest''s output was on the console. This patch converts the HVM output
buffering to be used by all domains, line buffering their output and
prefixing it with the domain ID. This is especially useful for debugging
stub domains.
Signed-off-by: Daniel De
2013 Sep 09
0
[PATCH v3] xen/console: buffer and show origin of guest PV writes
Guests other than domain 0 using the console output have previously been
controlled by the VERBOSE #define, but with no designation of which
guest''s output was on the console. This patch converts the HVM output
buffering to be used by all domains except the hardware domain (dom0):
stripping non-printable characters, line buffering the output, and
prefixing it with the domain ID. This is
2011 Nov 29
10
[PATCH 0 of 2] Fix correctness race in xc_mem_paging_prep
ging_prep ensures that an mfn is backing the paged-out gfn, and
transitions to the next state in the paging state machine for this page.
Foreign mappings of the gfn will now succeed. This is the key idea, as it
allows the pager to now map the gfn and fill in its contents.
Unfortunately, it also allows any other foreign mapper to map the gfn and read
its contents. This is particularly dangerous
2012 Feb 24
0
[xen-unstable test] 12043: regressions - FAIL
flight 12043 xen-unstable real [real]
http://www.chiark.greenend.org.uk/~xensrcts/logs/12043/
Regressions :-(
Tests which did not succeed and are blocking,
including tests which could not be run:
test-amd64-i386-rhel6hvm-amd 7 redhat-install fail REGR. vs. 12031
test-amd64-i386-rhel6hvm-intel 7 redhat-install fail REGR. vs. 12031
test-amd64-i386-xl-winxpsp3-vcpus1 7
2012 Jan 05
3
[PATCH 0 of 2] xenpaging:speed up page-in
The following two patches are about how to speed up in xenpaging when page in pages.
On suse11-64 with 4G memory,if we page out 2G pages,it will cost about 15.5 seconds,
but take 2088 seconds to finish paging in.If page-in costs too much time,it will cause
unmesurable problems when vm or dom0 access the paged_out page,such as BSOD,crash.
What鈥檚 more,the dom0 is always in high I/O pressure.
2012 Aug 10
18
[PATCH v2 0/5] ARM hypercall ABI: 64 bit ready
Hi all,
this patch series makes the necessary changes to make sure that the
current ARM hypercall ABI can be used as-is on 64 bit ARM platforms:
- it defines xen_ulong_t as uint64_t on ARM;
- it introduces a new macro to handle guest pointers, called
XEN_GUEST_HANDLE_PARAM (that has size 4 bytes on aarch and is going to
have size 8 bytes on aarch64);
- it replaces all the occurrences of
2011 Dec 24
3
mapping virtual address to corresponding shadow page in xen
hi,
How can I get corresponding shadowed page number of a virtual address in vm?
--
Best Regards,
Mohamad Rezaei
2011 Mar 25
2
[RFC PATCH 2/3] AMD IOMMU: Implement p2m sharing
--
Advanced Micro Devices GmbH
Sitz: Dornach, Gemeinde Aschheim,
Landkreis München Registergericht München,
HRB Nr. 43632
WEEE-Reg-Nr: DE 12919551
Geschäftsführer:
Alberto Bozzo, Andrew Bowd
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
http://lists.xensource.com/xen-devel
2013 Oct 30
3
[PATCH 4/4] XSA-60 security hole: flush cache when vmentry back to UC guest
From 159251a04afcdcd8ca08e9f2bdfae279b2aa5471 Mon Sep 17 00:00:00 2001
From: Liu Jinsong <jinsong.liu@intel.com>
Date: Thu, 31 Oct 2013 06:38:15 +0800
Subject: [PATCH 4/4] XSA-60 security hole: flush cache when vmentry back to UC guest
This patch flush cache when vmentry back to UC guest, to prevent
cache polluted by hypervisor access guest memory during UC mode.
The elegant way to do this
2012 Aug 16
27
[PATCH v3 0/6] ARM hypercall ABI: 64 bit ready
Hi all,
this patch series makes the necessary changes to make sure that the
current ARM hypercall ABI can be used as-is on 64 bit ARM platforms:
- it defines xen_ulong_t as uint64_t on ARM;
- it introduces a new macro to handle guest pointers, called
XEN_GUEST_HANDLE_PARAM (that has size 4 bytes on aarch and is going to
have size 8 bytes on aarch64);
- it replaces all the occurrences of
2011 Nov 08
48
Need help with fixing the Xen waitqueue feature
The patch ''mem_event: use wait queue when ring is full'' I just sent out
makes use of the waitqueue feature. There are two issues I get with the
change applied:
I think I got the logic right, and in my testing vcpu->pause_count drops
to zero in p2m_mem_paging_resume(). But for some reason the vcpu does
not make progress after the first wakeup. In my debugging there is one