search for: node_affin

Displaying 4 results from an estimated 4 matches for "node_affin".

Did you mean: node_admin
2013 Sep 17
1
[PATCH] xen: numa-sched: leave node-affinity alone if not in "auto" mode
If the domain''s NUMA node-affinity is being specified by the user/toolstack (instead of being automatically computed by Xen), we really should stick to that. This means domain_update_node_affinity() is wrong when it filters out some stuff from there even in "!auto" mode. This commit fixes that. Of course, this does not mean node-affinity is always honoured (e.g., a vcpu won''t run on a pcpu of a different cpupool) but the necessary logic for taking into account all the...
2013 Oct 16
4
[PATCH 1/7] xen: vNUMA support for PV guests
...+}; + +void domain_vnuma_destroy(struct domain_vnuma_info *v); + #endif /* __XEN_DOMAIN_H__ */ diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h index 25bf637..6693b35 100644 --- a/xen/include/xen/sched.h +++ b/xen/include/xen/sched.h @@ -408,6 +408,7 @@ struct domain nodemask_t node_affinity; unsigned int last_alloc_node; spinlock_t node_affinity_lock; + struct domain_vnuma_info vnuma; }; struct domain_setup_info diff --git a/xen/include/xen/vnuma.h b/xen/include/xen/vnuma.h new file mode 100644 index 0000000..2637476 --- /dev/null +++ b/xen/include/xen/vnuma.h @@ -...
2012 Jan 04
98
RFC: Still TODO for 4.2?
What are the outstanding things to do before we think we can start on the 4.2 -rc''s? Does anyone have a timetable in mind? hypervisor: * ??? - Keir, Tim, Jan? tools: * libxl stable API -- we would like 4.2 to define a stable API which downstream''s can start to rely on not changing. Aspects of this are: * event handling (IanJ working
2011 Sep 08
5
[PATCH 0 of 2] v2: memshare/xenpaging/xen-access fixes for xen-unstable
The following two patches allow the parallel use of memsharing, xenpaging and xen-access by using an independent ring buffer for each feature. Please review. v2: - update mem_event_check_ring arguments, check domain rather than domain_id - check ring_full first because its value was just evaluated - check if ring buffer is initialized before calling mem_access_domctl/mem_paging_domctl