> -----Original Message-----
> From: Konrad Rzeszutek Wilk [mailto:konrad.wilk@oracle.com]
> Sent: Friday, September 07, 2012 9:56 PM
> To: Ren, Yongjie
> Cc: Konrad Rzeszutek Wilk; ''xen-devel''; ''Keir
Fraser''; ''Ian Campbell''; ''Jan
> Beulich''
> Subject: Re: [Xen-devel] Xen4.2-rc3 test result
>
> On Fri, Sep 07, 2012 at 08:01:37AM +0000, Ren, Yongjie wrote:
> > > -----Original Message-----
> > > From: Konrad Rzeszutek Wilk [mailto:konrad.wilk@oracle.com]
> > > Sent: Thursday, September 06, 2012 7:12 PM
> > > To: Ren, Yongjie
> > > Cc: Konrad Rzeszutek Wilk; ''xen-devel'';
''Keir Fraser''; ''Ian Campbell'';
> ''Jan
> > > Beulich''
> > > Subject: Re: [Xen-devel] Xen4.2-rc3 test result
> > >
> > > On Thu, Sep 06, 2012 at 08:18:16AM +0000, Ren, Yongjie wrote:
> > > > > -----Original Message-----
> > > > > From: Konrad Rzeszutek [mailto:ketuzsezr@gmail.com] On
Behalf
> Of
> > > > > Konrad Rzeszutek Wilk
> > > > > Sent: Saturday, September 01, 2012 1:24 AM
> > > > > To: Ren, Yongjie
> > > > > Cc: ''xen-devel''; ''Keir
Fraser''; ''Ian Campbell''; ''Jan
Beulich''; ''Konrad
> > > > > Rzeszutek Wilk''
> > > > > Subject: Re: [Xen-devel] Xen4.2-rc3 test result
> > > >
> > > > > > 5. Dom0 cannot be shutdown before PCI detachment
from guest
> and
> > > > > when pci assignment conflicts
> > > > > >
http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1826
> > > > >
> > > > > Um, so you are assigning the same VF to two guests. I
am surprised
> > > that
> > > > > the tools even allowed you to do that. Was
''xm'' allowing you to do
> > > that?
> > > > >
> > > > No, ''xl'' doesn''t allow me to do
that. We can''t assignment a device to
> > > different guests.
> > > > Sorry, the description of this bug is not accurate. I
changed its title to
> > > "Dom0 cannot be shut down before PCI device detachment from
a
> guest".
> > > > If a guest (with a PCI device assigned) is running, Dom0
will panic
> when
> > > shutting down.
> > >
> > > And does it panic if you use the ''irqpoll''
option it asks for?
> > >
> > Adding ''irqpoll'' makes no change.
> >
> > It should be a regression for Xen from 4.1 to 4.2.
> > I didn''t meet this issue with 4.1 Xen and 3.5.3 Dom0.
>
> Aaah. That was not clear from the bugzilla.
>
> >
> > > Xen-pciback has no involvment as:
> > >
> > >
> > > [ 283.747488] xenbus_dev_shutdown: backend/pci/1/0: Initialised
!> > > Connected, skipping
> > >
> > > and the guest still keeps on getting interrupts.
> > >
> > > What is the stack at the hang?
> > >
> > [ 283.747488] xenbus_dev_shutdown: backend/pci/1/0: Initialised !>
Connected,
> > skipping
> > [ 283.747505] xenbus_dev_shutdown: backend/vkbd/2/0: Initialising
!> > Connected, skipping
> > [ 283.747515] xenbus_dev_shutdown: backend/console/2/0:
> Initialising !> > Connected, skipping
> > [ 380.236571] irq 16: nobody cared (try booting with the
"irqpoll"
> option)
> > [ 380.236588] Pid: 0, comm: swapper/0 Not tainted 3.4.4 #1
> > [ 380.236596] Call Trace:
> > [ 380.236601] <IRQ> [<ffffffff8110b538>]
> __report_bad_irq+0x38/0xd0
> > [ 380.236626] [<ffffffff8110b72c>] note_interrupt+0x15c/0x210
> > [ 380.236637] [<ffffffff81108ffa>]
> handle_irq_event_percpu+0xca/0x230
> > [ 380.236648] [<ffffffff811091b6>] handle_irq_event+0x56/0x90
> > [ 380.236658] [<ffffffff8110bde3>]
handle_fasteoi_irq+0x63/0x120
> > [ 380.236671] [<ffffffff8131c8f1>]
> __xen_evtchn_do_upcall+0x1b1/0x280
> > [ 380.236703] [<ffffffff8131d51a>]
xen_evtchn_do_upcall+0x2a/0x40
> > [ 380.236716] [<ffffffff816bc36e>]
> xen_do_hypervisor_callback+0x1e/0x30
> > [ 380.236723] <EOI> [<ffffffff810013aa>] ?
> hypercall_page+0x3aa/0x1000
> > [ 380.236742] [<ffffffff810013aa>] ?
hypercall_page+0x3aa/0x1000
> > [ 380.236754] [<ffffffff810538d0>] ? xen_safe_halt+0x10/0x20
> > [ 380.236768] [<ffffffff81067c9a>] ? default_idle+0x6a/0x1d0
> > [ 380.236778] [<ffffffff81067256>] ? cpu_idle+0x96/0xf0
> > [ 380.236789] [<ffffffff81689a18>] ? rest_init+0x68/0x70
> > [ 380.236800] [<ffffffff81ca9e33>] ? start_kernel+0x407/0x414
> > [ 380.236810] [<ffffffff81ca984a>] ? kernel_init+0x1e1/0x1e1
> > [ 380.236821] [<ffffffff81ca9346>] ?
> x86_64_start_reservations+0x131/0x136
> > [ 380.236833] [<ffffffff81cade7a>] ?
xen_start_kernel+0x621/0x628
> > [ 380.236841] handlers:
> > [ 380.236850] [<ffffffff81428d80>] usb_hcd_irq
> > [ 380.236860] Disabling IRQ #16
>
> That is not the hang stack. That is the kernel telling you that something
> has gone astray with an interrupt. But its unclear what happend _after_
> that.
>
> It might be that the system called the proper shutdown hypercall and its
> waiting for the hypervisor to do its stuff. Can you try using the
''q'' to get
> a stack dump of the dom0 and see where its spinning/sitting please?
>
(XEN) ''q'' pressed -> dumping domain info
(now=0x43F:E9B06F08)
(XEN) General information for domain 0:
(XEN) refcnt=3 dying=0 pause_count=0
(XEN) nr_pages=1048576 xenheap_pages=6 shared_pages=0 paged_pages=0
dirty_cpus={1-24,26-30} max_pages=4294967295
(XEN) handle=00000000-0000-0000-0000-000000000000 vm_assist=0000000d
(XEN) Rangesets belonging to domain 0:
(XEN) I/O Ports { 0-1f, 22-3f, 44-60, 62-9f, a2-3f7, 400-407, 40c-cfb,
d00-ffff }
(XEN) Interrupts { 0-315, 317-327 }
(XEN) I/O Memory { 0-febff, fec01-fec3e, fec40-fec7e, fec80-fedff,
fee01-ffffffffffffffff }
(XEN) Memory pages belonging to domain 0:
(XEN) DomPage list too long to display
(XEN) XenPage 000000000043c8ad: caf=c000000000000002, taf=7400000000000002
(XEN) XenPage 000000000043c8ac: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 000000000043c8ab: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 000000000043c8aa: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000000bb48e: caf=c000000000000002, taf=7400000000000002
(XEN) XenPage 0000000000418d84: caf=c000000000000002, taf=7400000000000002
(XEN) VCPU information and callbacks for domain 0:
(XEN) VCPU0: CPU14 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={14} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU1: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={18} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU2: CPU15 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={15} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU3: CPU24 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={24} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU4: CPU27 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={27} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU5: CPU30 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={30} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU6: CPU24 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU7: CPU11 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={11} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU8: CPU19 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={19} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU9: CPU23 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU10: CPU28 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={28} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU11: CPU29 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={29} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU12: CPU10 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={10} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU13: CPU17 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={17} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU14: CPU8 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={8} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU15: CPU22 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={22} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU16: CPU7 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={7} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU17: CPU16 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={16} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU18: CPU26 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={26} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU19: CPU13 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={13} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU20: CPU3 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={3} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU21: CPU12 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={12} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU22: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU23: CPU23 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={23} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU24: CPU20 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={20} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU25: CPU2 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={2} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU26: CPU1 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={1} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU27: CPU9 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={9} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU28: CPU5 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={5} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU29: CPU4 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={4} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU30: CPU21 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={21} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) VCPU31: CPU6 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={6} cpu_affinity={0-127}
(XEN) pause_count=0 pause_flags=1
(XEN) No periodic timer
(XEN) General information for domain 1:
(XEN) refcnt=3 dying=0 pause_count=0
(XEN) nr_pages=1047549 xenheap_pages=6 shared_pages=0 paged_pages=0
dirty_cpus={} max_pages=1048832
(XEN) handle=dd1ee134-05ef-4a65-a2fe-163bce610687 vm_assist=00000000
(XEN) paging assistance: hap refcounts translate external
(XEN) Rangesets belonging to domain 1:
(XEN) I/O Ports { }
(XEN) Interrupts { 101-103 }
(XEN) I/O Memory { ebb41-ebb43, ebb60-ebb63 }
(XEN) Memory pages belonging to domain 1:
(XEN) DomPage list too long to display
(XEN) PoD entries=0 cachesize=0
(XEN) XenPage 000000000044000e: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 000000000044000d: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 000000000044000c: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004a04f7: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000000bd4f8: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004a0400: caf=c000000000000001, taf=7400000000000001
(XEN) VCPU information and callbacks for domain 1:
(XEN) VCPU0: CPU1 [has=F] poll=0 upcall_pend = 01, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=4
(XEN) paging assistance: hap, 4 levels
(XEN) No periodic timer
(XEN) VCPU1: CPU10 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU2: CPU11 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU3: CPU12 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU4: CPU13 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU5: CPU14 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU6: CPU15 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU7: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU8: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU9: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU10: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU11: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU12: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU13: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU14: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU15: CPU0 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={0-15}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) General information for domain 2:
(XEN) refcnt=3 dying=0 pause_count=0
(XEN) nr_pages=523261 xenheap_pages=6 shared_pages=0 paged_pages=0
dirty_cpus={31} max_pages=524544
(XEN) handle=b10a1595-a5b0-46e2-a67a-3c77c4755761 vm_assist=00000000
(XEN) paging assistance: hap refcounts translate external
(XEN) Rangesets belonging to domain 2:
(XEN) I/O Ports { }
(XEN) Interrupts { }
(XEN) I/O Memory { }
(XEN) Memory pages belonging to domain 2:
(XEN) DomPage list too long to display
(XEN) PoD entries=0 cachesize=0
(XEN) XenPage 00000000004b44f3: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004b44f2: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004b44f1: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004b44f0: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000000bb08f: caf=c000000000000001, taf=7400000000000001
(XEN) XenPage 00000000004b27a7: caf=c000000000000001, taf=7400000000000001
(XEN) VCPU information and callbacks for domain 2:
(XEN) VCPU0: CPU31 [has=T] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={31} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=0
(XEN) paging assistance: hap, 4 levels
(XEN) No periodic timer
(XEN) VCPU1: CPU16 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=4
(XEN) paging assistance: hap, 4 levels
(XEN) No periodic timer
(XEN) VCPU2: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU3: CPU19 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU4: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU5: CPU19 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU6: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU7: CPU19 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU8: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU9: CPU16 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU10: CPU17 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU11: CPU18 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU12: CPU19 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU13: CPU20 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU14: CPU21 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) VCPU15: CPU22 [has=F] poll=0 upcall_pend = 00, upcall_mask = 00
dirty_cpus={} cpu_affinity={16-31}
(XEN) pause_count=0 pause_flags=2
(XEN) paging assistance: hap, 1 levels
(XEN) No periodic timer
(XEN) Notifying guest 0:0 (virq 1, port 5, stat 0/0/0)
(XEN) Notifying guest 0:1 (virq 1, port 12, stat 0/0/0)
[ 4672.954495] (XEN) Notifying guest 0:2 (virq 1, port 19, stat 0/0/0)
vcpu 0
(XEN) Notifying guest 0:3 (virq 1, port 26, stat 0/0/0)
0: masked=0 pend(XEN) Notifying guest 0:4 (virq 1, port 33, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:5 (virq 1, port 40, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:6 (virq 1, port 47, stat 0/0/0)
(XEN) Notifying guest 0:7 (virq 1, port 54, stat 0/0/0)
1: masked=1 pend(XEN) Notifying guest 0:8 (virq 1, port 61, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:9 (virq 1, port 68, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:10 (virq 1, port 75, stat 0/0/0)
(XEN) Notifying guest 0:11 (virq 1, port 82, stat 0/0/0)
2: masked=1 pend(XEN) Notifying guest 0:12 (virq 1, port 89, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:13 (virq 1, port 96, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:14 (virq 1, port 103, stat 0/0/0)
(XEN) Notifying guest 0:15 (virq 1, port 110, stat 0/0/0)
3: masked=1 pend(XEN) Notifying guest 0:16 (virq 1, port 117, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:17 (virq 1, port 124, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:18 (virq 1, port 131, stat 0/0/0)
(XEN) Notifying guest 0:19 (virq 1, port 138, stat 0/0/0)
4: masked=1 pend(XEN) Notifying guest 0:20 (virq 1, port 145, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:21 (virq 1, port 152, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:22 (virq 1, port 159, stat 0/0/0)
(XEN) Notifying guest 0:23 (virq 1, port 166, stat 0/0/0)
5: masked=1 pend(XEN) Notifying guest 0:24 (virq 1, port 173, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:25 (virq 1, port 180, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:26 (virq 1, port 187, stat 0/0/0)
(XEN) Notifying guest 0:27 (virq 1, port 194, stat 0/0/0)
6: masked=1 pend(XEN) Notifying guest 0:28 (virq 1, port 201, stat 0/0/0)
ing=0 event_sel (XEN) Notifying guest 0:29 (virq 1, port 208, stat 0/0/0)
0000000000000000(XEN) Notifying guest 0:30 (virq 1, port 215, stat 0/0/0)
(XEN) Notifying guest 0:31 (virq 1, port 222, stat 0/0/0)
7: masked=1 pend(XEN) Notifying guest 1:0 (virq 1, port 0, stat 0/-1/-1)
ing=0 event_sel (XEN) Notifying guest 1:1 (virq 1, port 0, stat 0/-1/0)
0000000000000000(XEN) Notifying guest 1:2 (virq 1, port 0, stat 0/-1/0)
(XEN) Notifying guest 1:3 (virq 1, port 0, stat 0/-1/0)
8: masked=1 pend(XEN) Notifying guest 1:4 (virq 1, port 0, stat 0/-1/0)
ing=0 event_sel (XEN) Notifying guest 1:5 (virq 1, port 0, stat 0/-1/0)
0000000000000000(XEN) Notifying guest 1:6 (virq 1, port 0, stat 0/-1/0)
(XEN) Notifying guest 1:7 (virq 1, port 0, stat 0/-1/0)
9: masked=1 pend(XEN) Notifying guest 1:8 (virq 1, port 0, stat 0/-1/0)
ing=1 event_sel (XEN) Notifying guest 1:9 (virq 1, port 0, stat 0/-1/0)
0000000000000002(XEN) Notifying guest 1:10 (virq 1, port 0, stat 0/-1/0)
(XEN) Notifying guest 1:11 (virq 1, port 0, stat 0/-1/0)
10: masked=1 pen(XEN) Notifying guest 1:12 (virq 1, port 0, stat 0/-1/0)
ding=0 event_sel(XEN) Notifying guest 1:13 (virq 1, port 0, stat 0/-1/0)
000000000000000(XEN) Notifying guest 1:14 (virq 1, port 0, stat 0/-1/0)
0
(XEN) Notifying guest 1:15 (virq 1, port 0, stat 0/-1/0)
11: masked=1 pen(XEN) Notifying guest 2:0 (virq 1, port 0, stat 0/-1/0)
ding=0 event_sel(XEN) Notifying guest 2:1 (virq 1, port 0, stat 0/-1/0)
000000000000000(XEN) Notifying guest 2:2 (virq 1, port 0, stat 0/-1/0)
0
(XEN) Notifying guest 2:3 (virq 1, port 0, stat 0/-1/0)
12: masked=1 pen(XEN) Notifying guest 2:4 (virq 1, port 0, stat 0/-1/0)
ding=0 event_sel(XEN) Notifying guest 2:5 (virq 1, port 0, stat 0/-1/0)
000000000000000(XEN) Notifying guest 2:6 (virq 1, port 0, stat 0/-1/0)
0
(XEN) Notifying guest 2:7 (virq 1, port 0, stat 0/-1/0)
13: masked=1 pen(XEN) Notifying guest 2:8 (virq 1, port 0, stat 0/-1/0)
ding=0 event_sel(XEN) Notifying guest 2:9 (virq 1, port 0, stat 0/-1/0)
000000000000000(XEN) Notifying guest 2:10 (virq 1, port 0, stat 0/-1/0)
0
(XEN) Notifying guest 2:11 (virq 1, port 0, stat 0/-1/0)
14: masked=1 pen(XEN) Notifying guest 2:12 (virq 1, port 0, stat 0/-1/0)
ding=0 event_sel(XEN) Notifying guest 2:13 (virq 1, port 0, stat 0/-1/0)
000000000000000(XEN) Notifying guest 2:14 (virq 1, port 0, stat 0/-1/0)
0
(XEN) Notifying guest 2:15 (virq 1, port 0, stat 0/-1/0)
15: masked=1 pen(XEN) Shared frames 0 -- Saved frames 0
ding=(XEN) *** Serial input -> DOM0 (type ''CTRL-a'' three
times to switch input to Xen)