Displaying 20 results from an estimated 5000 matches similar to: "changing max vcpus threshold while a domain is running"
2006 Mar 10
12
[PATCH] Add SCHEDOP_block_on
This patch adds a facility to block on a particular event channel, for
when a domain needs to block, but cannot enable interrupts. A domain
uses the call something like this:
while (1) {
clear_evtchn_pending(evtchn);
if (check_for_data(evtchn))
break;
HYPERVISOR_block_on(evtchn);
}
The clear of the pending is needed to ensure that any subsequent calls
to block_on() don''t
2007 Feb 14
0
[PATCH 11/12] Allow vcpu to pause self
Allow one vcpu to pause self, with full context sync-ed.
Signed-off-by Kevin Tian <kevin.tian@intel.com>
diff -r f6443af464be xen/common/domain.c
--- a/xen/common/domain.c Tue Feb 13 13:37:12 2007 +0800
+++ b/xen/common/domain.c Tue Feb 13 13:41:44 2007 +0800
@@ -375,6 +375,18 @@ void vcpu_pause_nosync(struct vcpu *v)
vcpu_sleep_nosync(v);
}
+/* This is not a code path called from
2006 Oct 04
0
[PATCH,RFC 6/17] 32-on-64 shared info handling
Index: 2006-10-04/xen/arch/x86/domain.c
===================================================================
--- 2006-10-04.orig/xen/arch/x86/domain.c 2006-10-04 15:03:07.000000000 +0200
+++ 2006-10-04/xen/arch/x86/domain.c 2006-10-04 15:11:03.000000000 +0200
@@ -550,10 +550,10 @@ static void load_segments(struct vcpu *n
/* CS longword also contains full evtchn_upcall_mask. */
2005 Oct 10
13
[PATCH] 0/2 VCPU creation and allocation
I''ve put together two patches. The first introduces a new dom0_op,
set_max_vcpus, which with an associated variable and a check in the
VCPUOP handler fixes [1]bug 288. Also included is a new VCPUOP,
VCPUOP_create, which handles all of the vcpu creation tasks and leaves
initialization and unpausing to VCPUOP_initialize. The separation
allows for build-time allocation of vcpus which
2011 Dec 20
0
sedf: remove useless tracing printk and harmonize comments style.
sched_sedf.c used o have its own mechanism for producing tracing-alike
kind of information (domain block, wakeup, etc.). Nowadays, with an even
not so high number of pCPUs/vCPUs, just trying to enable this makes
the serial console completely unusable, produces tons of very hard to
parse and interpreet logging and can easily livelock Dom0. Moreover,
pretty much the same result this is struggling to
2005 Oct 07
1
[patch] testing needed: "xenif" dom0_ops
This patch changes the dom0_ops structures as discussed in the thread
"32/64-bit hypercall interface". Keir, I added a struct inside XENIF_PTR() to
catch direct users in the general code; it was quite useful to have the
compiler identify those spots.
I have compiled x86_32 and run it with xm-test[1] under qemu. There are 63
passed tests, so that''s good. I still need to
2006 Jul 31
1
[PATCH 5/6] xen, tools: calculate nr_cpus via num_online_cpus
Once Xen calculates nr_nodes properly, all nr_cpu calculations based on
nr_nodes * sockets_per_node * cores_per_socket * threads_per_core are
broken. The easy fix is to replace those calculations with a new field,
nr_cpus in physinfo which is calculated by num_online_cpus(). This
patch does so and attempts to change all users over to nr_cpus field in
physinfo. This patch touches
2005 Jun 17
0
RE: [Patch 1/2] Re-org dom0_ops.h to allow arch specificdefinition
>-----Original Message-----
>From: Ian Pratt [mailto:m+Ian.Pratt@cl.cam.ac.uk]
>Sent: Friday, June 17, 2005 4:25 PM
>To: Tian, Kevin; xen-devel@lists.xensource.com
>
>It probably makes sense not to re-use arch-specific dom0_op numbers
>(though there wouldn''t be any real abiguity), but I''m not sure its
worth
>trying to carve up the number space like this.
2005 Nov 24
1
[PATCH] dom0_ops -EPERM not -EACCES?
Hi all,
dom0_ops return -EACCES when the interface version in wrong. I think
that an acm failure is closer to being non-priv''ed, which returns
-EPERM. Wouldn''t want dom0 tools to report they were the wrong version
just because ACM limited some operation, surely?
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> (authored)
diff -r e2d1c88d50f7 xen/common/dom0_ops.c
2006 Feb 08
2
[PATCH] make x86_64 vcpu hotplug work like i386
hi,
i386 vcpu hotplug seems to work reliably but x86_64 does not and i think
i have discovered why. in x86-64, a cpu within a domu can be removed with
vcpu-set but subsequent calls do nothing.
after xenwatch_thread grabs the event triggered by the write to the
store, it calls the registered handler and never comes back.
eventually, __cpu_die in drivers/xen/core/smpboot.c spins while waiting
2006 Sep 14
0
[PATCH][RFC] credit scheduler enhancement, eliminate unnecessary vcpu migrations
Hi Keir,
csched_load_balance is used to see if there are higher privileged vcpu
in
other physical processors'' runnable queue, if there is, then migrate
this vcpu to this physical processor.
But at following scenario, this vcpu migration is unnecessary.
1. idle_vcpu0 is running on lp0, hvm_vcpu is in lp0''s runnable queue,
this
happens when hvm_vcpu is just being woken up,
2.
2007 Aug 10
2
jinfo, jboss and ExtendedDTraceProbes
Hi,
I would like to use dtrace to observe some behaviour in my J2EE app
running in JBoss. I start JBoss normally in a 1.6 VM. When I try to
enable set the ExtendedDTraceProbes flag using jinfo, I get a thread
dump in the JBoss console and the message: "Unable to open door: target
process not responding or HotSpot VM not loaded" in the jinfo console.
I have tried this with both
2014 Jun 11
1
Recent (6.02) PXELINUX on Soekris net4501
* Ady, 2014-06-11 :
> Perhaps it would be worth testing with official pre-built binaries of
> Syslinux 6.03-pre14 using the following config:
Different behaviour.
* pxelinux.0:
After TFTP load of pxelinus.0, spits out gibberish for ~1 minute
(as though it had set an incorrect default serial port speed),
then loads config and switches back to correct baud rate.
"Hello
2013 Feb 21
2
[PATCH v3] x86/nhvm: properly clean up after failure to set up all vCPU-s
Otherwise we may leak memory when setting up nHVM fails half way.
This implies that the individual destroy functions will have to remain
capable (in the VMX case they first need to be made so, following
26486:7648ef657fe7 and 26489:83a3fa9c8434) of being called for a vCPU
that the corresponding init function was never run on.
Once at it, also remove a redundant check from the corresponding
2013 Nov 20
2
[PATCH] hvm: reset TSC to 0 after domain resume from S3
Host S3 implicitly resets the host TSC to 0, but the tsc offset for hvm domains is not recalculated
when they resume, causing it to go into negative values. In Linux guest using tsc clocksource, this
results in a hang after wrap back to positive values since the tsc clocksource implementation
expects it reset.
Signed-off-by: Tomasz Wroblewski <tomasz.wroblewski@citrix.com>
---
2005 Jul 03
11
[PATCH] xm info
This patch makes "xm info" show information on xen version, compile
info, number of socket/core, etc...
In order to do that, it extends physinfo hypercall to return number of
socket, adds few functions to libxc and extends python wrapper
correspondingly.
Here is the output of new "xm info":
--
system : Linux
host : ubuntu
xen_release
2007 Jan 11
6
[PATCH 4/8] HVM save restore: vcpu context support
[PATCH 4/8] HVM save restore: vcpu context support
Signed-off-by: Zhai Edwin <edwin.zhai@intel.com>
save/restore HVM vcpu context such as vmcs
diff -r ee20d1905bde xen/arch/x86/domain.c
--- a/xen/arch/x86/domain.c Thu Jan 11 16:40:55 2007 +0800
+++ b/xen/arch/x86/domain.c Thu Jan 11 16:46:59 2007 +0800
@@ -573,6 +573,7 @@ int arch_set_info_guest(
else
{
2013 Mar 27
2
[PATCH] x86/S3: Restore broken vcpu affinity on resume (v3)
When in SYS_STATE_suspend, and going through the cpu_disable_scheduler
path, save a copy of the current cpu affinity, and mark a flag to
restore it later.
Later, in the resume process, when enabling nonboot cpus restore these
affinities.
v2:
Fix formatting problems.
remove early return in cpu_disable_scheduler() path.
v3:
Fix remaining errant tab
Move restore_vcpu_affinity() to thaw_domains(),
2013 Dec 14
0
[PATCH v2] xen: sched: introduce hard and soft affinity in credit 2 scheduler
Modified function runq_candidate in the credit 2 scheduler to
have it consider hard and soft affinity when choosing the next
vCPU from the run queue to run on the given pCPU.
Function now chooses the vCPU with the most credit that has hard affinity
and maybe soft affinity for the given pCPU. If it does not have soft affinity
and there is another vCPU that prefers to run on the given pCPU, then as
2006 Dec 01
1
[PATCH 2/10] Add support for netfront/netback acceleration drivers
This set of patches adds the support for acceleration plugins to the
netfront/netback drivers. These plugins are intended to support
virtualisable network hardware that can be directly accessed from the
guest, bypassing dom0.
This is in response to the RFC we posted to xen-devel with an outline
of our approach at the end of September.
To follow will be another set of patches to provide our