similar to: [rfc] [patch] 32/64-bit hypercall interface revisited

Displaying 20 results from an estimated 300 matches similar to: "[rfc] [patch] 32/64-bit hypercall interface revisited"

2007 Jan 18
13
[PATCH 0/5] dump-core take 2:
The following dump-core patches changes its format into ELF, adds PFN-GMFN table, HVM support, and adds experimental IA64 support. - ELF format Program header and note section are adopted. - HVM domain support To know the memory area to dump, XENMEM_set_memory_map is added. XENMEM_memory_map hypercall is for current domain, so new one is created. and hvm domain builder tell xen its
2010 Jan 09
3
101th domU fails to start with "SETVCPUCONTEXT failed"
Hello there, We (a small hosting community) are running a steadily growing number of Xen domUs on a quad dualcore Xeon server with 64GB ram. We''ve got 100 running domUs at the moment. Trying to create a new one results in this error: Error: (1, ''Internal error'', ''launch_vm: SETVCPUCONTEXT failed (rc=-1)\n'') If I shut down another domain, I can
2012 Sep 04
2
[PATCH] valgrind: Support for ioctls used by Xen toolstack processes.
Please CC as I''m not subscribed to valgrind-developers. Under Xen the toolstack is responsible for managing the domains in the system, e.g. creating, destroying, and otherwise manipulating them. To do this it uses a number of ioctls on the /proc/xen/privcmd device. Most of these (the MMAPBATCH ones) simply set things up such that a subsequenct mmap call will map the desired guest
2008 Apr 22
4
[PATCH 0/3] xen: more portability patches
Hi Jeremy. Here are the 3 patches for ia64/xen support. The first one is just to move manage.c under drivers/xen. This is trivial. The second and third ones are for compilation fix on ia64. You may want to postpone those 2 patches to make your merge task easy as you said before. thanks, Diffstat arch/x86/xen/Makefile | 2 +- drivers/xen/Makefile | 2 +-
2008 Apr 22
4
[PATCH 0/3] xen: more portability patches
Hi Jeremy. Here are the 3 patches for ia64/xen support. The first one is just to move manage.c under drivers/xen. This is trivial. The second and third ones are for compilation fix on ia64. You may want to postpone those 2 patches to make your merge task easy as you said before. thanks, Diffstat arch/x86/xen/Makefile | 2 +- drivers/xen/Makefile | 2 +-
2012 Jan 11
12
[PATCH 00/11] Btrfs: some patches for 3.3
The biggest one is a fix for fstrim, and there''s a fix for on-disk free space cache. Others are small fixes and cleanups. The last three have been sent weeks ago. The patchset is also available in this repo: git://repo.or.cz/linux-btrfs-devel.git for-chris Note there''s a small confict with Al Viro''s vfs changes. Li Zefan (11): Btrfs: add pinned extents to
2011 Jan 23
1
SETVCPUCONTEXT failed
Hi guys, I cannot get my domU''s to boot. My current setup is as follows: Xen: 4.0.1 Dom0: Debian with custom linux 2.6.32.27 Xen kernel DomU: Same kernel as Dom0 When I try to start the VM, I just get: Error: (1, ''Internal error'', ''launch_vm: SETVCPUCONTEXT failed (rc=-1)\n'') And /var/log/xend.log says: [2011-01-23 15:36:01 3201] DEBUG
2011 Jan 23
1
SETVCPUCONTEXT failed
Hi guys, I cannot get my domU''s to boot. My current setup is as follows: Xen: 4.0.1 Dom0: Debian with custom linux 2.6.32.27 Xen kernel DomU: Same kernel as Dom0 When I try to start the VM, I just get: Error: (1, ''Internal error'', ''launch_vm: SETVCPUCONTEXT failed (rc=-1)\n'') And /var/log/xend.log says: [2011-01-23 15:36:01 3201] DEBUG
2012 Aug 10
18
[PATCH v2 0/5] ARM hypercall ABI: 64 bit ready
Hi all, this patch series makes the necessary changes to make sure that the current ARM hypercall ABI can be used as-is on 64 bit ARM platforms: - it defines xen_ulong_t as uint64_t on ARM; - it introduces a new macro to handle guest pointers, called XEN_GUEST_HANDLE_PARAM (that has size 4 bytes on aarch and is going to have size 8 bytes on aarch64); - it replaces all the occurrences of
2010 Sep 15
15
xenpaging fixes for kernel and hypervisor
Patrick, there following patches fix xenpaging for me. Granttable handling is incomplete. If a page is gone, a GNTST_eagain should be returned to the caller to inidcate the hypercall has to be retried after a while, until the page is available again. Please review. Olaf _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com
2010 Jun 01
31
XCP
Hi there, We are using latest version of XCP on 6 hosts. While issuing VM.start or VM.start_on xmlrpc functional call , it says : {''Status'': ''Failure'', ''ErrorDescription'': [''SESSION_INVALID'', ''OpaqueRef:cfb6df14-387d-40a1-cc27-d5962cba7712'']} However if I put VM.start in a loop maybe after
2009 Jan 14
5
[PATCH] Support cross-bitness guest when core-dumping
This patch allows core-dumping to work on a cross-bit host/guest configuration, whereas previously that was not supported. It supports both PV and FV guests. The core file format generated by the host, needs to match that of the guest, so an alignment issue is addressed, along with the p2m frame list handling being done according to the guest size. Signed-off-by: Bruce Rogers
2012 Sep 14
1
[PATCH] xenpm: make argument parsing and error handling more consistent
Specifically, what values are or aren''t accepted as CPU identifier, and how the values get interpreted should be consistent across sub-commands (intended behavior now: non-negative values are okay, and along with omitting the argument, specifying "all" will also be accepted). For error handling, error messages should get consistently issued to stderr, and the tool should now
2007 Aug 28
6
[PATCH] Make XEN_DOMCTL_destroydomain hypercall continuable.
# HG changeset patch # User yamahata@valinux.co.jp # Date 1188274001 -32400 # Node ID 2c9db26f1d0e0fdd4757d76a67f4b37ba0e40351 # Parent 58d131f1fb35977ff2d8682f553391c8a866d52c Make XEN_DOMCTL_destroydomain hypercall continuable. XEN_DOMCTL_destroydomain hypercall frees domain resources, especially it frees all pages of the domain. When domain memory is very large, it takes too long resulting in
2011 Jan 06
3
Offline Deduplication for Btrfs V2
Just a quick update, I''ve dropped the hashing stuff in favor of doing a memcmp in the kernel to make sure the data is still the same. The thing that takes a while is reading the data up from disk, so doing a memcmp of the entire buffer isn''t that big of a deal, not to mention there''s a possiblity for malicious users if there is a problem with the hashing algorithms we
2008 Mar 07
6
where is the location of definition of "do_xen_version"?
hi, my friends: Currently, i am studying the way of hypercall's implementation. i have already known the flow of hypercall's execuation, and i decided to add a new hypercall into the Xen. first, i want to know the detail of one hypercall function, for example, "do_xen_version", but i can not find the location of definition of "do_xen_version". who can help me? i have
2010 Dec 02
1
Making a hypercall in DomU
Hi, I have *implemented a new hypercall* and it is working fine when tested from the Dom0 user-space. I want to invoke this hypercall from DomU user-space. I copied all the /usr/lib/libxen* and /usr/include/xen* (recursively) to the DomU Here''s the code I wrote to invoke hypercall: #include <stdio.h> #include <xenctrl.h> int main(void){ int xc_handle, rc;
2011 Dec 21
2
[PATCH] xenpm: assorted adjustments
- use consistent error values (stop mixing of [positive] errno values with literal -E... ones) - properly format output - don''t use leading zeros in decimal output - move printing of average frequency into P-state conditional (rather than a C-state one) - don''t print some C-state related info when CPU idle management is disabled in the hypervisor - use calloc() for array
2007 Feb 26
2
[PATCH 0 of 2] Parse image elfnotes, write them to xenstore, save and load via image sxpr
Here are two patches that let xm create, save and restore extract and preserve elfnotes read by the domain builder. This is handy for a few things. In particular, I''d like it so that xm can decide whether or not guest domains support fast resume (if save fails, or for checkpointing). _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com
2013 Feb 13
4
[PATCH 0/3] FLASK policy build rework
These patches update the example FLASK policy shipped with Xen and enable its build if the required tools are present. The third patch requires rerunning autoconf to update tools/configure. [PATCH 1/3] flask/policy: sort dom0 accesses [PATCH 2/3] flask/policy: rework policy build system [PATCH 3/3] tools/flask: add FLASK policy to build