search for: vmresum

Displaying 6 results from an estimated 6 matches for "vmresum".

Did you mean: vmresume
2012 Nov 22
41
[PATCH V3] vmx/nmi: Do not use self_nmi() in VMEXIT handler
The self_nmi() code cause''s an NMI to be triggered by sending an APIC message to the local processor. However, NMIs are blocked by the VMEXIT, until the next iret or VMENTER. Volume 3 Chapter 27 Section 1 of the Intel SDM states: An NMI causes subsequent NMIs to be blocked, but only after the VM exit completes. As a result, as soon as the VMENTER happens, an immediate VMEXIT happens
2007 Feb 15
8
Communication between guest OS and VMM
Hi all, Please can anybody explain how communication(direct or indirect) happens between xen and guest os(windows) in full virtualization mode. Both from VMM to guest and from guest to VMM. Thanks and regards, Aditya. __________________________________________________________ Yahoo! India Answers: Share what you know. Learn something new http://in.answers.yahoo.com/
2017 Nov 15
0
[Xen-devel] [PATCH RFC v3 0/6] x86/idle: add halt poll support
...sive workload have seen obviously performance > drop when running inside VM. The main reason is that the overhead > is amplified when running inside VM. The most cost I have seen is > inside idle path. Meaning an VMEXIT b/c it is an 'halt' operation ? And then going back in guest (VMRESUME) takes time. And hence your latency gets all whacked b/c of this? So if I understand - you want to use your _full_ timeslice (of the guest) without ever (or as much as possible) to go in the hypervisor? Which means in effect you don't care about power-saving or CPUfreq savings, you just want...
2017 Nov 13
7
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com> Some latency-intensive workload have seen obviously performance drop when running inside VM. The main reason is that the overhead is amplified when running inside VM. The most cost I have seen is inside idle path. This patch introduces a new mechanism to poll for a while before entering idle state. If schedule is needed during poll, then we
2017 Nov 13
7
[PATCH RFC v3 0/6] x86/idle: add halt poll support
From: Yang Zhang <yang.zhang.wz at gmail.com> Some latency-intensive workload have seen obviously performance drop when running inside VM. The main reason is that the overhead is amplified when running inside VM. The most cost I have seen is inside idle path. This patch introduces a new mechanism to poll for a while before entering idle state. If schedule is needed during poll, then we
2013 Mar 12
14
vpmu=1 and running 'perf top' within a PVHVM guest eventually hangs dom0 and hypervisor has stuck vCPUS. Romley-EP (model=45, stepping=2)
...t;: test %eax,%eax 0xffff82c4c01d3888 <vmx_vmexit_handler+584>: jne 0xffff82c4c01d4d5a <vmx_vmexit_handler+5914> Which I think translates to roughly: 2323 local_irq_enable(); 2324 2325 /* XXX: This looks ugly, but we need a mechanism to ensure 2326 * any pending vmresume has really happened 2327 */ 2328 vcpu_nestedhvm(v).nv_vmswitch_in_progress = 0; 2329 if ( nestedhvm_vcpu_in_guestmode(v) ) 2330 { 2331 if ( nvmx_n2_vmexit_handler(regs, exit_reason) ) 2332 goto out; 2333 } Which does not make much sense. There is no loop h...