Displaying 12 results from an estimated 12 matches similar to: "[PATCH RFC v3 5/6] tick: get duration of the last idle loop"
2013 Jun 03
0
[PATCH] xen/smp: Fixup NOHZ per cpu data when onlining an offline CPU.
The xen_play_dead is an undead function. When the vCPU is told to
offline it ends up calling xen_play_dead wherin it calls the
VCPUOP_down hypercall which offlines the vCPU. However, when the
vCPU is onlined back, it resumes execution right after
VCPUOP_down hypercall.
That was OK (albeit the API for play_dead assumes that the CPU
stays dead and never returns) but with commit 4b0c0f294
(tick:
2017 Nov 13
0
[PATCH RFC v3 6/6] KVM guest: introduce smart idle poll algorithm
From: Yang Zhang <yang.zhang.wz at gmail.com>
using smart idle poll to reduce the useless poll when system is idle.
Signed-off-by: Quan Xu <quan.xu0 at gmail.com>
Signed-off-by: Yang Zhang <yang.zhang.wz at gmail.com>
Cc: Paolo Bonzini <pbonzini at redhat.com>
Cc: Thomas Gleixner <tglx at linutronix.de>
Cc: Ingo Molnar <mingo at redhat.com>
Cc: "H. Peter
2017 Nov 13
0
[PATCH RFC v3 6/6] KVM guest: introduce smart idle poll algorithm
From: Yang Zhang <yang.zhang.wz at gmail.com>
using smart idle poll to reduce the useless poll when system is idle.
Signed-off-by: Quan Xu <quan.xu0 at gmail.com>
Signed-off-by: Yang Zhang <yang.zhang.wz at gmail.com>
Cc: Paolo Bonzini <pbonzini at redhat.com>
Cc: Thomas Gleixner <tglx at linutronix.de>
Cc: Ingo Molnar <mingo at redhat.com>
Cc: "H. Peter
2017 Nov 13
0
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
2017 Nov 13
0
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
2006 Feb 09
0
I need help on VICIDIAL and auto dial
Vicidial can't call and transfer to my softphone.
I get some line that says
Spawn Extension....exited on non zero....
Here's some of the CLI output. I am using Asterisk 1.2.4 and astguiclient
1.1.8
...thanks for the help
|SELECT count(*) FROM vicidial_auto_calls where status = 'LIVE' and
server_ip='127.0.0.1' and campaign_id = '' and call_time <
2017 Nov 13
2
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Quan Xu <quan.xu0 at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
don't
2017 Nov 13
2
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Quan Xu <quan.xu0 at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
don't
2006 Feb 09
1
Re: Help on Vicidial
Here is another log from the * server CLI, I reall hope some one can help me
out on this one. thanks
|SELECT count(*) FROM vicidial_auto_calls where status = 'LIVE' and
server_ip='127.0.0.1' and
campaign_id = '' and call_time < "" and lead_id != '';|
-- VDAD get agent: |0|update of vla table: |127.0.0.1
|UPDATE vicidial_live_agents set
2017 Nov 13
7
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
2017 Nov 13
7
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com>
Some latency-intensive workload have seen obviously performance
drop when running inside VM. The main reason is that the overhead
is amplified when running inside VM. The most cost I have seen is
inside idle path.
This patch introduces a new mechanism to poll for a while before
entering idle state. If schedule is needed during poll, then we
2009 Mar 02
0
[PATCH 4 of 13] DisplayState interface change
Import "DisplayState interface change" from qemu mainstream: the patch
has been adapted to qemu-xen and merged with several following fixes.
The original qemu svn commit is the following:
git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@6336 c046a42c-6fe2-441c-8c8c-71466251a162
Signed-off-by: Stefano Stabellini <stefano.stabellini@eu.citrix.com>
---
diff --git a/console.c