similar to: Making a hypercall in DomU

Displaying 20 results from an estimated 400 matches similar to: "Making a hypercall in DomU"

2011 Oct 26
3
Hypercall by DomU Application
Hi, I am working on a security architecture. In this architecture, the application in DomU has to communicate directly with the hypervisor. But as I can see, the xen architecture allows only DomU kernel to raise a hypercall. I am planning to enable application to communicate with xen directly. I am assuming, setting up a trap gate with Ring-3 access should do the trick. I have few questions
2012 Sep 14
1
[PATCH] xenpm: make argument parsing and error handling more consistent
Specifically, what values are or aren''t accepted as CPU identifier, and how the values get interpreted should be consistent across sub-commands (intended behavior now: non-negative values are okay, and along with omitting the argument, specifying "all" will also be accepted). For error handling, error messages should get consistently issued to stderr, and the tool should now
2006 Aug 14
0
[PATCH] remove another xc_dom0_op() user
# HG changeset patch # User john.levon@sun.com # Date 1155596263 -3600 # Node ID dc12f0942649ee27ff2aec3e3d9c2d99746b9e58 # Parent 09bc238664cb7b6d0f960f5e135ac80a75b342f1 Remove another external-to-libxc user of xc_dom0_op(). Signed-off-by: John Levon <john.levon@sun.com> diff -r 09bc238664cb -r dc12f0942649 tools/xenmon/xenbaked.c --- a/tools/xenmon/xenbaked.c Mon Aug 14 15:40:36 2006
2010 Oct 15
0
Raising VIRQ from hypervisor
Hi, How can I raise VIRQ from hypervisor? -- Srujan D. Kotikela _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
2006 Sep 29
4
[PATCH 4/6] xen: export NUMA topology in physinfo hcall
This patch modifies the physinfo hcall to export NUMA CPU and Memory topology information. The new physinfo hcall is integrated into libxc and xend (xm info specifically). Included in this patch is a minor tweak to xm-test''s xm info testcase. The new fields in xm info are: nr_nodes : 4 mem_chunks : node0:0x0000000000000000-0x0000000190000000
2006 Jul 31
1
[PATCH 5/6] xen, tools: calculate nr_cpus via num_online_cpus
Once Xen calculates nr_nodes properly, all nr_cpu calculations based on nr_nodes * sockets_per_node * cores_per_socket * threads_per_core are broken. The easy fix is to replace those calculations with a new field, nr_cpus in physinfo which is calculated by num_online_cpus(). This patch does so and attempts to change all users over to nr_cpus field in physinfo. This patch touches
2007 Aug 28
6
[PATCH] Make XEN_DOMCTL_destroydomain hypercall continuable.
# HG changeset patch # User yamahata@valinux.co.jp # Date 1188274001 -32400 # Node ID 2c9db26f1d0e0fdd4757d76a67f4b37ba0e40351 # Parent 58d131f1fb35977ff2d8682f553391c8a866d52c Make XEN_DOMCTL_destroydomain hypercall continuable. XEN_DOMCTL_destroydomain hypercall frees domain resources, especially it frees all pages of the domain. When domain memory is very large, it takes too long resulting in
2011 Dec 21
2
[PATCH] xenpm: assorted adjustments
- use consistent error values (stop mixing of [positive] errno values with literal -E... ones) - properly format output - don''t use leading zeros in decimal output - move printing of average frequency into P-state conditional (rather than a C-state one) - don''t print some C-state related info when CPU idle management is disabled in the hypervisor - use calloc() for array
2012 Apr 24
1
[PATCH] [resend] xen-access: Check return values and clean up on errors during init
Check the return values of the libxc mem_access calls. Free allocated structures (platform_info, domain_info) on errors during initialization and exit. Unbind VIRQ, close event channel and connection to Xen on errors during initialization Signed-off-by: Aravindh Puthiyaparambil <aravindh@virtuata.com> Acked-by: Ian Jackson <ian.jackson@eu.citrix.com> diff -r ea47068fa7a2 -r
2008 Mar 07
6
where is the location of definition of "do_xen_version"?
hi, my friends: Currently, i am studying the way of hypercall's implementation. i have already known the flow of hypercall's execuation, and i decided to add a new hypercall into the Xen. first, i want to know the detail of one hypercall function, for example, "do_xen_version", but i can not find the location of definition of "do_xen_version". who can help me? i have
2012 Jul 23
2
[PATCH V2] qemu-xen-traditionnal, Fix dirty logging during migration.
This moves the xen_modified_memory call from cpu_physical_memory_map to cpu_physical_memory_unmap because the memory could be migrated before the device model have written to it. But because we need to know the guest address and to avoid rewriting a new function, the call is moved to qemu_invalidate_entry. So this later has to new parameters, the length of the mapping and if it was a write.
2008 May 24
2
Use of XEN_GUEST_HANDLE
Hi all, I have recently started to go through xen source. I want to know the usage of XEN_GUEST_HANDLE. I see thats its just a macro which will prefix each data type with __guest_handle_ . And DEFINE_XEN_GUEST_HANDLE(name) will just typedefs ''__guest_handle_name'' to be a pointer to a data type ''name '' . What is the reason for such abstraction? And how
2004 Oct 18
3
potential bug in "xm atropos" implementation
>From tools/libxc/xc_atropos.c: int xc_atropos_domain_set(int xc_handle, u32 domid, u64 period, u64 slice, u64 latency, int xtratime) which takes 6 arguments >From tools/python/xen/xm/main.py: class ProgAtropos(Prog): <snip> def main(self, args): if len(args) != 5: self.err("%s: Invalid argument(s)" % args[0])
2008 Jul 10
2
question about mapping memory from Dom0 to domU
Hi: I wanna write code about sharing memory between dom0 and domU, and I decide to use void *xc_map_foreign_range(int xc_handle, unint32_t dom, int size, int port, unsigned long mfn) (this is a function in xenctl.h). My question is : If I wanna map a range of memory in dom0 to domU, how can I get the value of mfn? and I do not know the port here. Can anybody help me ? Thanks .
2012 Jan 31
26
[PATCH 00/10] FLASK updates: MSI interrupts, cleanups
This patch set adds XSM security labels to useful debugging output locations, and fixes some assumptions that all interrupts behaved like GSI interrupts (which had useful non-dynamic IDs). It also cleans up the policy build process and adds an example of how to use the user field in the security context. Debug output: [PATCH 01/10] xsm: Add security labels to event-channel dump [PATCH 02/10] xsm:
2007 Oct 19
4
[PATCH] nr_cpus calculation problem due to incorrect sockets_per_node
Testing on an 8-node 128-way NUMA machine has exposed a problem with Xen''s nr_cpus calculation. In this case, since Xen cuts off recognized CPUs at 32, the machine appears to have 16 CPUs on the first and second nodes and none on the remaining nodes. Given this asymmetry, the calculation of sockets_per_node (which is later used to calculate nr_cpus) is incorrect:
2009 Jan 14
5
[PATCH] Support cross-bitness guest when core-dumping
This patch allows core-dumping to work on a cross-bit host/guest configuration, whereas previously that was not supported. It supports both PV and FV guests. The core file format generated by the host, needs to match that of the guest, so an alignment issue is addressed, along with the p2m frame list handling being done according to the guest size. Signed-off-by: Bruce Rogers
2007 Jan 11
0
[PATCH 6/8] HVM save restore: guest memory handling
[PATCH 6/8] HVM save restore: guest memory handling Signed-off-by: Zhai Edwin <edwin.zhai@intel.com> add support for save/restore HVM guest memory diff -r bb1c450b2739 tools/libxc/xc_hvm_restore.c --- a/tools/libxc/xc_hvm_restore.c Thu Jan 11 21:03:11 2007 +0800 +++ b/tools/libxc/xc_hvm_restore.c Thu Jan 11 21:05:45 2007 +0800 @@ -31,6 +31,40 @@ #include <xen/hvm/ioreq.h>
2006 Apr 14
8
[rfc] [patch] 32/64-bit hypercall interface revisited
Last year we had a discussion[1] about how the hypercall ABI unfortunately contains fields that change width between 32- and 64-bit builds. This is a huge problem as we come up on the python management stack for ppc64, since the distributions ship 32-bit python. A 32-bit python/libxc cannot currently manage a 64-bit hypervisor. I had a patch but was unable to test it, and some other things were
2007 Jan 18
13
[PATCH 0/5] dump-core take 2:
The following dump-core patches changes its format into ELF, adds PFN-GMFN table, HVM support, and adds experimental IA64 support. - ELF format Program header and note section are adopted. - HVM domain support To know the memory area to dump, XENMEM_set_memory_map is added. XENMEM_memory_map hypercall is for current domain, so new one is created. and hvm domain builder tell xen its