similar to: DomU crash during migration when suspending source domain

Displaying 20 results from an estimated 100 matches similar to: "DomU crash during migration when suspending source domain"

2008 Jan 09
4
[PATCH/RFC 0/2] CPU hotplug virtio driver
I'm sending a first draft of my proposed cpu hotplug driver for kvm/virtio The first patch is the kernel module, while the second, the userspace pci device. The host boots with the maximum cpus it should ever use, through the -smp parameter. Due to real machine constraints (which qemu copies), i386 does not allow for any addition of cpus after boot, so this is the most general way. I do
2008 Jan 09
4
[PATCH/RFC 0/2] CPU hotplug virtio driver
I'm sending a first draft of my proposed cpu hotplug driver for kvm/virtio The first patch is the kernel module, while the second, the userspace pci device. The host boots with the maximum cpus it should ever use, through the -smp parameter. Due to real machine constraints (which qemu copies), i386 does not allow for any addition of cpus after boot, so this is the most general way. I do
2009 Nov 12
20
VIRQ_CON_RING
Is there any real user for this vIRQ? I''m asking because, while debugging a problem that required to add some printk()s that could temporarily issue at a high rate, I ran into the situation that this extra output prevented guest/dom0 boot from making any progress. As I then realized this was due to the tasklet_schedule() called from inside the console handling code, which resulted in
2009 Nov 12
20
VIRQ_CON_RING
Is there any real user for this vIRQ? I''m asking because, while debugging a problem that required to add some printk()s that could temporarily issue at a high rate, I ran into the situation that this extra output prevented guest/dom0 boot from making any progress. As I then realized this was due to the tasklet_schedule() called from inside the console handling code, which resulted in
2009 Apr 22
3
Error on running any controller- Noob question
Hi , I am running on Vista with webrick ... When I generate controllers , models and views I get an error as under. C:/Ruby/lib/ruby/gems/1.8/gems/activesupport-2.3.2/lib/active_support/ dependencies.rb:105:in `const_missing'' C:/ROR_lpapp042109/app/controllers/afflnetworks_controller.rb:5:in `index'' I tried reinstalling but kept stumbling on this. gem -v 1.3.2 rails -v 2.3.2
2005 Aug 04
7
[PATCH] Convert shutdown to use xenstore
The attached patch: 1. Converts the shutdown driver and xend to use the store instead of control messages, 2. Includes Anthony''s xenstore notification code, and 3. Changes xend so that sysrq''s are no longer sent as "special case" shutdown messages. Store keys are cheap, so making the sysrq delivery less obscure is good. I think I have made all of the
2009 Apr 22
2
Nova-s2 pci-passthrough not properly working
Dear all, i''m having problems passing a Hauppage Nova-S2 card to my DomU (Mythtv Backend). Hardware is: HP XW8200 2x 3.0GHz Xeon 8Gb ECC Ram 1x TT3200-S2 card 1x DVBWordl S2 card 1x Nova-S2 card [many other card for other domUs...] Dom0 is Ubuntu 8.04 server LTS. Xen 3.2 from repo. DomU (Mythtv) is a 8.04 (using LVM) both using 2.6.24-23-xen kernels Dom0 configs: lspci: ... 02:01.0
2009 Nov 23
18
OpenSuSE 11.2 bug, dom0-cpus limit causes xenwatch_cb running 100% and xm command freeze and xend dead
Hi all, I have upgraded a test machine from OpenSuSE 11.1 to 11.2. I have found following bug: the server is a 2x quadcore intel box also 2x4=8cpu If you limit the dom0 cpu with dom0-cpus= [1-7]: - [xenwatch_cb] is running 100% cpu and makes var log entry every 65 sec BUG: soft lockup - CPU#X stuck for 61s! - xm commands not work - xend is dead if set dom0-cpus to 0 or 8: - everything looks
2005 Jun 07
8
[PATCH] add dom0 vcpu hotplug control
This patch adds new control messages for vcpu hotplug events. Via the xm vcpu_hotplug sub-program, VCPUS in domains can be enabled/disabled when CONFIG_HOTPLUG_CPU is enabled in the target domain''s kernel. Currently there is nothing that tracks whether a VCPU is up or down. My previous [1]patch added a new per-VCPU flag (VCPUF_down) which could be used to keep track of which VCPUS are
2009 Jan 09
1
[PULL]: tip/cpus4096 updates
Hi Ingo, Please pull a few more cpus4096 changes. Most are fairly lightweight with the biggest change being to use cpumask_var_t's for the domain and pending_mask cpumask's in irq_desc. It also deals with this cpu_evtchn_mask bss space bump: 65536 +2031616 2097152 +3100% cpu_evtchn_mask(.bss) Everything merges into tip/master cleanly. Thanks! Mike The following changes since
2009 Jan 09
1
[PULL]: tip/cpus4096 updates
Hi Ingo, Please pull a few more cpus4096 changes. Most are fairly lightweight with the biggest change being to use cpumask_var_t's for the domain and pending_mask cpumask's in irq_desc. It also deals with this cpu_evtchn_mask bss space bump: 65536 +2031616 2097152 +3100% cpu_evtchn_mask(.bss) Everything merges into tip/master cleanly. Thanks! Mike The following changes since
2012 Sep 11
0
[PATCH 1/3] x86/hvm: don't use indirect calls without need
Direct calls perform better, so we should prefer them and use indirect ones only when there indeed is a need for indirection. Signed-off-by: Jan Beulich <jbeulich@suse.com> --- a/xen/arch/x86/apic.c +++ b/xen/arch/x86/apic.c @@ -1373,7 +1373,7 @@ void error_interrupt(struct cpu_user_reg void pmu_apic_interrupt(struct cpu_user_regs *regs) { ack_APIC_irq(); -
2007 Feb 20
0
[PATCH] Xenconsoled should ignore spurious watch event. Otherwise, it can rebind to the same evtchn of a dying domU during suspending and cause below error message:
# HG changeset patch # User john.levon@sun.com # Date 1171946682 28800 # Node ID e8a7491ed4c57d4d82642dd954b8909d9065b377 # Parent 24a6bc1d1b93405558741a2a8fa7ba2f7c44964a Xenconsoled should ignore spurious watch event. Otherwise, it can rebind to the same evtchn of a dying domU during suspending and cause below error message: (XEN) event_channel.c:177:d0 EVTCHNOP failure: domain 0, error -22,
2008 Apr 16
1
Wine suspending itself? (WoW on laptop with i945GM graphics)
I'm experiencing some real weirdness trying to get WoW to work in Wine on this laptop I have. It works pretty well (if a bit slow, but I'm willing to attribute that to the Intel graphics hardware) except that, at unpredictable intervals, the Wine process will suddenly freeze -- graphics stop updating and sound goes into a loop. The thing is that, if I run it from a terminal window, I can
2012 May 26
2
CentOS 6.2: suspending vim with ctrl-z and resuming with fg - stopped working
Hello, pardon my chaotic question, but does anybody have an idea, why can't I suspend vim-enhanced-7.2.411-1.6.el6.x86_64 with a CTRL-Z, then execute few commands at my bash prompt and then get back to the vim session again with "fg"? It has stopped working at my CentOS 6.2 machine (I haven't noticed exactly when) but works fine with CentOS 5.x. I'm using PuTTY to login to
2020 May 18
0
how does autofs deal with stuck NFS mounts and suspending to RAM?
On May 18, 2020, at 5:13 AM, hw <hw at gc-24.de> wrote: > > Is there a better alternative for mounting remote file systems over unreliable > connections? I don?t have a good answer for you, because if you?d asked me without all this backstory whether NFS or SSHFS is more tolerant of bad connections, I?d have told you SSHFS. NFS comes out of the "Unix lab? world, where all
2020 May 19
0
how does autofs deal with stuck NFS mounts and suspending to RAM?
> That's what I thought. Should I make a bug report? Sshfs is clearly intended > to reconnect automatically when mounted like that, and it doesn't do that. Not so clearly. Look at the sshfs reconnect option, and also ssh/ssfs ServerAliveInterval/ServerAliveCountMax.
2020 May 19
1
how does autofs deal with stuck NFS mounts and suspending to RAM?
On Mon, May 18, 2020 at 05:36:03PM -0600, Warren Young wrote: > On May 18, 2020, at 5:13 AM, hw <hw at gc-24.de> wrote: > > > > Is there a better alternative for mounting remote file systems > > over unreliable > > connections? > > I don?t have a good answer for you, because if you?d asked me > without all this backstory whether NFS or SSHFS is more
2013 Apr 08
1
[PATCH] drm/nouveau: idle all channels before suspending
Seems to make suspend slightly more reliable on my system. Cc: stable at vger.kernel.org [3.7+] Signed-off-by: Maarten Lankhorst <maarten.lankhorst at canonical.com> --- diff --git a/drivers/gpu/drm/nouveau/nouveau_drm.c b/drivers/gpu/drm/nouveau/nouveau_drm.c index b6bdc9f..5032c31 100644 --- a/drivers/gpu/drm/nouveau/nouveau_drm.c +++ b/drivers/gpu/drm/nouveau/nouveau_drm.c @@ -464,6
2020 May 19
2
how does autofs deal with stuck NFS mounts and suspending to RAM?
On Tuesday, May 19, 2020 1:36:03 AM CEST Warren Young wrote: > On May 18, 2020, at 5:13 AM, hw <hw at gc-24.de> wrote: > > Is there a better alternative for mounting remote file systems over > > unreliable connections? > > I don?t have a good answer for you, because if you?d asked me without all > this backstory whether NFS or SSHFS is more tolerant of bad