search for: vcpu_pause_nosync

Displaying 4 results from an estimated 4 matches for "vcpu_pause_nosync".

2007 Feb 14
0
[PATCH 11/12] Allow vcpu to pause self
Allow one vcpu to pause self, with full context sync-ed. Signed-off-by Kevin Tian <kevin.tian@intel.com> diff -r f6443af464be xen/common/domain.c --- a/xen/common/domain.c Tue Feb 13 13:37:12 2007 +0800 +++ b/xen/common/domain.c Tue Feb 13 13:41:44 2007 +0800 @@ -375,6 +375,18 @@ void vcpu_pause_nosync(struct vcpu *v) vcpu_sleep_nosync(v); } +/* This is not a code path called from guest, since execution is paused + * after next context switch. ACPI PM requires this to pause vcpu0 of + * dom0 itself, in case that vcpu not running on cpu 0. + */ +void vcpu_pause_self(void) +{ + struct vc...
2007 Jun 27
10
[PATCH 6/10] Allow vcpu to pause self
...ing interface, since the new flag also serves as a sync point. Signed-off-by Kevin Tian <kevin.tian@intel.com> diff -r d5315422dbc8 xen/common/domain.c --- a/xen/common/domain.c Mon May 14 18:35:31 2007 -0400 +++ b/xen/common/domain.c Mon May 14 20:21:04 2007 -0400 @@ -530,6 +530,17 @@ void vcpu_pause_nosync(struct vcpu *v) vcpu_sleep_nosync(v); } +/* _VPF_need_sync serves not only as flag for sync pause, but also + * as hint for other cpu waiting for this pause. + */ +void vcpu_pause_self(void) +{ + struct vcpu *v = current; + + set_bit(_VPF_need_sync, &v->pause_flags); + vcpu_...
2012 Jan 05
3
[PATCH 0 of 2] xenpaging:speed up page-in
The following two patches are about how to speed up in xenpaging when page in pages. On suse11-64 with 4G memory,if we page out 2G pages,it will cost about 15.5 seconds, but take 2088 seconds to finish paging in.If page-in costs too much time,it will cause unmesurable problems when vm or dom0 access the paged_out page,such as BSOD,crash. What鈥檚 more,the dom0 is always in high I/O pressure.
2011 Sep 08
5
[PATCH 0 of 2] v2: memshare/xenpaging/xen-access fixes for xen-unstable
The following two patches allow the parallel use of memsharing, xenpaging and xen-access by using an independent ring buffer for each feature. Please review. v2: - update mem_event_check_ring arguments, check domain rather than domain_id - check ring_full first because its value was just evaluated - check if ring buffer is initialized before calling mem_access_domctl/mem_paging_domctl