Dear xen-users! I am running Xen 4.0.1 with 2.6.32.x pvops dom0 kernels from the git stable branch for several months now. Every now and then I gave the migration feature a try, but I failed to migrate a machine up to now. I did not have the time to look into the errors closely, but since I didn''t read about migration beeing impossible on this list, I obviously must do something wrong :( My setup: - 2 identical servers with ubuntu 10.04 server and self packaged kernel and xen (inspired by [1]) - the storage for the dom U is a DRBD volume on top of lvm2 - the servers are connected via a dedicated gbit link The domU config: name = "test1"; memory = 1024; vcpus = 1; builder = ''hvm'' kernel = "/usr/lib/xen/boot/hvmloader" boot = "c" pae = 1 acpi = 1 apic = 1 localtime = 1 on_poweroff = "destroy" on_reboot = "restart" on_crash = "destroy" device_model = "/usr/lib/xen/bin/qemu-dm" vnc = 1 vfb = [ "type=vnc,vncunused=1,vnclisten=0.0.0.0,keymap=en-us" ] disk = [ ''phy:/dev/drbd9,xvda,w''] vif = [ "bridge=br0, mac=00:16:3e:00:01:0d" ]; serial = "pty" I start the vm on server xen02 and do a xm migrate --live test1 xen01 What I get in the xen02 xend.log: [2011-04-11 12:58:04 2612] DEBUG (XendCheckpoint:124) [xc_save]: /usr/lib/xen/bin/xc_save 29 7 0 0 5 [2011-04-11 12:58:04 2612] INFO (XendCheckpoint:423) xc_save: failed to get the suspend evtchn port [2011-04-11 12:58:04 2612] INFO (XendCheckpoint:423) [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 1 0%^H^H^H^H 5%^H^H^H^H 10%^H^H^H^H 15%^H^H^H^H 20%^H^H^H^H 25%^H^H^H^H 30%^H^H^H^H 35%^H^H^H^H 40%^H^H^H^H 45%^H^H^H^H 50%^H^H^H^H 55%^H^H^H^H 60%^H^H^H^H 65%^H^H^H^H 70%^H^H^H^H 75%^H^H^H^H 80%^H^H^H^H 85%^H^H^H^H 90%^H^H^H^H 95%^M 1: sent 265216, skipped 1355, delta 13794ms, dom0 0%, target 0%, sent 630Mb/s, dirtied 3Mb/s 1431 pages [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 2 0%^H^H^H^H 23%^M 2: sent 1416, skipped 15, delta 40ms, dom0 0%, target 0%, sent 1159Mb/s, dirtied 123Mb/s 151 pages [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 3 0%^M 3: sent 151, skipped 0, delta 12ms, dom0 0%, target 0%, sent 412Mb/s, dirtied 49Mb/s 18 pages [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 4 0%^M 4: sent 18, skipped 0, Start last iteration [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:394) suspend [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:127) In saveInputHandler suspend [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:129) Suspending 7 ... [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:519) XendDomainInfo.shutdown(suspend) [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:1891) XendDomainInfo.handleShutdownWatch [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:1891) XendDomainInfo.handleShutdownWatch [2011-04-11 12:58:18 2612] INFO (XendDomainInfo:2088) Domain has shutdown: name=migrating-test1 id=7 reason=suspend. [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:135) Domain 7 suspended. [2011-04-11 12:58:18 2612] INFO (image:538) signalDeviceModel:restore dm state to running [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) SUSPEND shinfo 00001d11 [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) delta 260ms, dom0 3%, target 1%, sent 2Mb/s, dirtied 43Mb/s 343 pages [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:144) Written done [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 5 0%^M 5: sent 343, skipped 0, delta 10ms, dom0 0%, target 0%, sent 1123Mb/s, dirtied 1123Mb/s 343 pages [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Total pages sent267144 (0.25x) [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) (of which 0 were fixups) [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) All memory is saved [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Save exit rc=0 [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:3053) XendDomainInfo.destroy: domid=7 [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2411) Destroying device model [2011-04-11 12:58:18 2612] INFO (image:615) migrating-test1 device model terminated [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2418) Releasing devices [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2424) Removing vif/0 The migration fails somewhere halfway through, since the receiving host xen01 has a domU test1 after the procedure, but this domU doesn''t respond. Thats the log output of xen01: [2011-04-11 12:58:05 2550] INFO (image:822) Need to create platform device.[domid:28] [2011-04-11 12:58:05 2550] DEBUG (XendCheckpoint:286) restore:shadow=0x9, _static_max=0x40000000, _static_min=0x0, [2011-04-11 12:58:05 2550] DEBUG (XendCheckpoint:305) [xc_restore]: /usr/lib/xen/bin/xc_restore 27 28 2 3 1 1 1 0 [2011-04-11 12:58:05 2550] INFO (XendCheckpoint:423) xc_domain_restore start: p2m_size = 100000 [2011-04-11 12:58:05 2550] INFO (XendCheckpoint:423) Reloading memory pages: 0% [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Read 5792 bytes of QEMU data [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Read 888 bytes of QEMU data [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) ERROR Internal error: Error when reading batch size [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) ERROR Internal error: error when buffering batch, finishing [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Writing 6680 bytes of QEMU data [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Restore exit with rc=0 [2011-04-11 12:58:18 2550] DEBUG (XendCheckpoint:394) store-mfn 1044476 [2011-04-11 12:58:18 2550] DEBUG (XendDomainInfo:2992) XendDomainInfo.completeRestore Since I did not read about other people having this issue, I didn''t file a bug report right away. Does somebody have an advice? Thanks, Mark [1] http://bderzhavets.wordpress.com/2010/06/02/setup-libvirt-0-8-0-xen-4-0-on-top-of-ubuntu-10-04-server-via-daniel-baumann-virtualization-ppa/ _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On Tue, Apr 12, 2011 at 10:39 AM, Mark Dokter <dokter@icg.tugraz.at> wrote:> Dear xen-users! > > I am running Xen 4.0.1 with 2.6.32.x pvops dom0 kernels from the git > stable branch for several months now. > Every now and then I gave the migration feature a try, but I failed to > migrate a machine up to now. I did not have the time to look into the > errors closely, but since I didn''t read about migration beeing > impossible on this list, I obviously must do something wrong :( > > My setup: > - 2 identical servers with ubuntu 10.04 server and self packaged kernel > and xen (inspired by [1]) > - the storage for the dom U is a DRBD volume on top of lvm2 > - the servers are connected via a dedicated gbit link > > The domU config: > name = "test1"; > memory = 1024; > vcpus = 1; > builder = ''hvm'' > kernel = "/usr/lib/xen/boot/hvmloader" > boot = "c" > pae = 1 > acpi = 1 > apic = 1 > localtime = 1 > on_poweroff = "destroy" > on_reboot = "restart" > on_crash = "destroy" > device_model = "/usr/lib/xen/bin/qemu-dm" > vnc = 1 > vfb = [ "type=vnc,vncunused=1,vnclisten=0.0.0.0,keymap=en-us" ] > disk = [ ''phy:/dev/drbd9,xvda,w''] > vif = [ "bridge=br0, mac=00:16:3e:00:01:0d" ]; > serial = "pty" > > I start the vm on server xen02 and do a > xm migrate --live test1 xen01 > > What I get in the xen02 xend.log: > > [2011-04-11 12:58:04 2612] DEBUG (XendCheckpoint:124) [xc_save]: > /usr/lib/xen/bin/xc_save 29 7 0 0 5 > [2011-04-11 12:58:04 2612] INFO (XendCheckpoint:423) xc_save: failed to > get the suspend evtchn port > [2011-04-11 12:58:04 2612] INFO (XendCheckpoint:423) > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory > pages: iter 1 0%^H^H^H^H 5%^H^H^H^H 10%^H^H^H^H 15%^H^H^H^H > 20%^H^H^H^H 25%^H^H^H^H 30%^H^H^H^H 35%^H^H^H^H 40%^H^H^H^H 45%^H^H^H^H > 50%^H^H^H^H 55%^H^H^H^H 60%^H^H^H^H 65%^H^H^H^H 70%^H^H^H^H 75%^H^H^H^H > 80%^H^H^H^H 85%^H^H^H^H 90%^H^H^H^H 95%^M 1: sent 265216, skipped 1355, > delta 13794ms, dom0 0%, target 0%, sent 630Mb/s, dirtied 3Mb/s 1431 pages > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory > pages: iter 2 0%^H^H^H^H 23%^M 2: sent 1416, skipped 15, delta 40ms, > dom0 0%, target 0%, sent 1159Mb/s, dirtied 123Mb/s 151 pages > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory > pages: iter 3 0%^M 3: sent 151, skipped 0, delta 12ms, dom0 0%, target > 0%, sent 412Mb/s, dirtied 49Mb/s 18 pages > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory > pages: iter 4 0%^M 4: sent 18, skipped 0, Start last iteration > [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:394) suspend > [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:127) In > saveInputHandler suspend > [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:129) Suspending 7 ... > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:519) > XendDomainInfo.shutdown(suspend) > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:1891) > XendDomainInfo.handleShutdownWatch > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:1891) > XendDomainInfo.handleShutdownWatch > [2011-04-11 12:58:18 2612] INFO (XendDomainInfo:2088) Domain has > shutdown: name=migrating-test1 id=7 reason=suspend. > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:135) Domain 7 suspended. > [2011-04-11 12:58:18 2612] INFO (image:538) signalDeviceModel:restore dm > state to running > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) SUSPEND shinfo 00001d11 > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) delta 260ms, dom0 > 3%, target 1%, sent 2Mb/s, dirtied 43Mb/s 343 pages > [2011-04-11 12:58:18 2612] DEBUG (XendCheckpoint:144) Written done > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Saving memory > pages: iter 5 0%^M 5: sent 343, skipped 0, delta 10ms, dom0 0%, target > 0%, sent 1123Mb/s, dirtied 1123Mb/s 343 pages > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Total pages sent> 267144 (0.25x) > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) (of which 0 were > fixups) > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) All memory is saved > [2011-04-11 12:58:18 2612] INFO (XendCheckpoint:423) Save exit rc=0 > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:3053) > XendDomainInfo.destroy: domid=7 > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2411) Destroying device > model > [2011-04-11 12:58:18 2612] INFO (image:615) migrating-test1 device model > terminated > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2418) Releasing devices > [2011-04-11 12:58:18 2612] DEBUG (XendDomainInfo:2424) Removing vif/0 > > > > The migration fails somewhere halfway through, since the receiving host > xen01 has a domU test1 after the procedure, but this domU doesn''t respond. > Thats the log output of xen01: > > > [2011-04-11 12:58:05 2550] INFO (image:822) Need to create platform > device.[domid:28] > [2011-04-11 12:58:05 2550] DEBUG (XendCheckpoint:286) > restore:shadow=0x9, _static_max=0x40000000, _static_min=0x0, > [2011-04-11 12:58:05 2550] DEBUG (XendCheckpoint:305) [xc_restore]: > /usr/lib/xen/bin/xc_restore 27 28 2 3 1 1 1 0 > [2011-04-11 12:58:05 2550] INFO (XendCheckpoint:423) xc_domain_restore > start: p2m_size = 100000 > [2011-04-11 12:58:05 2550] INFO (XendCheckpoint:423) Reloading memory > pages: 0% > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Read 5792 bytes of > QEMU data > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Read 888 bytes of > QEMU data > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) ERROR Internal > error: Error when reading batch size > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) ERROR Internal > error: error when buffering batch, finishing > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Writing 6680 bytes > of QEMU data > [2011-04-11 12:58:18 2550] INFO (XendCheckpoint:423) Restore exit with rc=0 > [2011-04-11 12:58:18 2550] DEBUG (XendCheckpoint:394) store-mfn 1044476 > [2011-04-11 12:58:18 2550] DEBUG (XendDomainInfo:2992) > XendDomainInfo.completeRestore > > > > Since I did not read about other people having this issue, I didn''t file > a bug report right away. Does somebody have an advice? >Can you reproduce this on a newer version of Xen (4.1 or xen-unstable)? Thanks, Todd> Thanks, > Mark > > [1] > http://bderzhavets.wordpress.com/2010/06/02/setup-libvirt-0-8-0-xen-4-0-on-top-of-ubuntu-10-04-server-via-daniel-baumann-virtualization-ppa/-- Todd Deshane http://www.linkedin.com/in/deshantm http://www.xen.org/products/cloudxen.html _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
hi all, I have just ported the pvovhvm drivers to ubuntu 10.10 ,the domU can boot successfully, but when the domU running, intermittence, the system will give me such warnings: INFO: task login:1073 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. INFO: task flush-202:0:457 blocked for more than 120seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. INFO: task jbd2/sda1-8:422 blocked for more than 120 seconds."echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. what''s wrong with the domU? is it means there are some errors inxen-blkfront? how should I debug this error??? Thanks a lot. _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
Todd Deshane
2011-Apr-13 05:31 UTC
Re: [Xen-users] some errors about pvonhvm : xen-blkfront
2011/4/13 立新 牛 <nlxswig@hotmail.com>:> hi all, > I have just ported the pvovhvm drivers to ubuntu 10.10 ,the domU can > boot successfully, but when the domU running, intermittence, the system > will give me such warnings: > INFO: task login:1073 blocked for more than 120 seconds. > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disa bles this > message. > INFO: task flush-202:0:457 blocked for more than 120seconds. > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this > message. > INFO: task jbd2/sda1-8:422 blocked for more than 120 seconds. > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > > what''s wrong with the domU? is it means there are some errors in > xen-blkfront? how should I debug this error??? > Thanks a lot. >What version of the PVonHVM drivers did you port? There were some recent patches that fixed some things (not sure of the details), but you can find them on xen-devel see for example: http://xen.markmail.org/search/?q=rc3+from:%22Stefano+Stabellini%22#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results Thanks, Todd -- Todd Deshane http://www.linkedin.com/in/deshantm http://www.xen.org/products/cloudxen.html _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
hi, thanks for your time. I ported the pvonhvm with this method: Ubuntu 10.10 was comply to the pvops, so it has it''s own blkfront, so I just add platform-pci to ubuntu 10.10 source, and then compile, that''s ok, the platform-pci was based on the source linux 2.6.32, which come from debian 6.0, the debian 6.0 support pvonhvm.> From: todd.deshane@xen.org > Date: Wed, 13 Apr 2011 05:31:09 +0000 > Subject: Re: [Xen-users] some errors about pvonhvm : xen-blkfront > To: nlxswig@hotmail.com > CC: xen-users@lists.xensource.com > > 2011/4/13 立新 牛 <nlxswig@hotmail.com>: > > hi all, > > I have just ported the pvovhvm drivers to ubuntu 10.10 ,the domU can > > boot successfully, but when the domU running, intermittence, the system > > will give me such warnings: > > INFO: task login:1073 blocked for more than 120 seconds. > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disa bles this > > message. > > INFO: task flush-202:0:457 blocked for more than 120seconds. > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this > > message. > > INFO: task jbd2/sda1-8:422 blocked for more than 120 seconds. > > "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > > > > what''s wrong with the domU? is it means there are some errors in > > xen-blkfront? how should I debug this error??? > > Thanks a lot. > > > > What version of the PVonHVM drivers did you port? > > There were some recent patches that fixed some things (not sure of the > details), but you can find them on xen-devel see for example: > > http://xen.markmail.org/search/?q=rc3+from:%22Stefano+Stabellini%22#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results > > Thanks, > Todd > > > -- > Todd Deshane > http://www.linkedin.com/in/deshantm > http://www.xen.org/products/cloudxen.html > > _______________________________________________ > Xen-users mailing list > Xen-users@lists.xensource.com > http://lists.xensource.com/xen-users_______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On 04/13/2011 07:13 AM, Todd Deshane wrote:> On Tue, Apr 12, 2011 at 10:39 AM, Mark Dokter <dokter@icg.tugraz.at> wrote: >> Dear xen-users! >> [...] >> >> >> Since I did not read about other people having this issue, I didn''t file >> a bug report right away. Does somebody have an advice? >> > > Can you reproduce this on a newer version of Xen (4.1 or xen-unstable)? >I have one spare machine for testing. I can set up xen 4.1 there (I''ll try a debian unstable, since I failed packaging xen 4.1 for ubuntu). With only one machine, I can only test migrating from xen 4.0.1 to xen 4.1 - you think this makes sense for sorting out the problem? Btw, the domU is running ubuntu 10.10 with a 2.6.38 kernel and pv-on-hvm drivers.> Thanks, > Todd >kind regards, Mark _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On Wed, Apr 13, 2011 at 7:35 AM, Mark Dokter <dokter@icg.tugraz.at> wrote:> On 04/13/2011 07:13 AM, Todd Deshane wrote: >> On Tue, Apr 12, 2011 at 10:39 AM, Mark Dokter <dokter@icg.tugraz.at> wrote: >>> Dear xen-users! >>> [...] >>> >>> >>> Since I did not read about other people having this issue, I didn''t file >>> a bug report right away. Does somebody have an advice? >>> >> >> Can you reproduce this on a newer version of Xen (4.1 or xen-unstable)? >> > > I have one spare machine for testing. I can set up xen 4.1 there (I''ll > try a debian unstable, since I failed packaging xen 4.1 for ubuntu). > With only one machine, I can only test migrating from xen 4.0.1 to xen > 4.1 - you think this makes sense for sorting out the problem? > > Btw, the domU is running ubuntu 10.10 with a 2.6.38 kernel and pv-on-hvm > drivers. >There have been some fixes (not sure if they are related) to the PV on HVM drivers recently. Take a look at: http://xen.markmail.org/search/?q=rc3+from:"Stefano+Stabellini"#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results Also the releases of Xen 4.1 and 4.0.2 (likely final RC) might have some developments that change things too. http://xen.markmail.org/search/?q=from:keir#query:from%3Akeir%20date%3A201104%20+page:1+mid:4cj4itzfe34jay4f+state:results http://wiki.xensource.com/xenwiki/Xen4.1 We just need to narrow down where things where things don''t and find out if it is specific to your setup or kernel, or xen version, etc. In general things are being tested so migration not working is likely some combination of your setup. Thanks, Todd -- Todd Deshane http://www.linkedin.com/in/deshantm http://www.xen.org/products/cloudxen.html _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On 04/13/2011 08:11 PM, Todd Deshane wrote:> > There have been some fixes (not sure if they are related) to the PV on > HVM drivers recently. Take a look at: > > http://xen.markmail.org/search/?q=rc3+from:"Stefano+Stabellini"#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results >I''ll try a 2.6.39rc4 domU kernel next.> Also the releases of Xen 4.1 and 4.0.2 (likely final RC) might have > some developments that change things too. > > http://xen.markmail.org/search/?q=from:keir#query:from%3Akeir%20date%3A201104%20+page:1+mid:4cj4itzfe34jay4f+state:results > > http://wiki.xensource.com/xenwiki/Xen4.1 > > > We just need to narrow down where things where things don''t and find > out if it is specific to your setup or kernel, or xen version, etc. In > general things are being tested so migration not working is likely > some combination of your setup. >I have set up a xen 4.1 box now (Debian kernel image linux-image-2.6.32-5-xen-amd64 from squeeze and xen 4.1 from source). I tried to migrate a Ubuntu Natty test domU from the xen 4.0.1 box to the 4.1 box and unfortunately it failed again. The domU was created on the target machine, but it doesn''t respond. The xend.log from the source (4.0.1) machine: [2011-04-26 12:41:40 2612] INFO (XendDomain:1206) Domain Sandbox-Natty (13) unpaused. [2011-04-26 12:42:37 2612] DEBUG (XendCheckpoint:124) [xc_save]: /usr/lib/xen/bin/xc_save 29 13 0 0 5 [2011-04-26 12:42:37 2612] INFO (XendCheckpoint:423) xc_save: failed to get the suspend evtchn port [2011-04-26 12:42:37 2612] INFO (XendCheckpoint:423) [2011-04-26 12:42:49 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 1 0%^H^H^H^H 5%^H^H^H^H 10%^H^H^H^H 15%^H^H^H^H 20%^H^H^H^H 25%^H^H^H^H 30%^H^H^H^H 35%^H^H^H^H 40%^H^H^H^H 45%^H^H^H^H 50%^H^H^H^H 55%^H^H^H^H 60%^H^H^H^H 65%^H^H^H^H 70%^H^H^H^H 75%^H^H^H^H 80%^H^H^H^H 85%^H^H^H^H 90%^H^H^H^H 95%^M 1: sent 265216, skipped 1807, delta 12446ms, dom0 0%, target 1%, sent 698Mb/s, dirtied 5Mb/s 1910 pages [2011-04-26 12:42:49 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 2 0%^H^H^H^H 22%^M 2: sent 1886, skipped 24, delta 52ms, dom0 0%, target 0%, sent 1188Mb/s, dirtied 35Mb/s 56 pages [2011-04-26 12:42:49 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 3 0%^M 3: sent 56, skipped 0, delta 12ms, dom0 0%, target 0%, sent 152Mb/s, dirtied 133Mb/s 49 pages [2011-04-26 12:42:49 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 4 0%^M 4: sent 49, skipped 0, Start last iteration [2011-04-26 12:42:49 2612] DEBUG (XendCheckpoint:394) suspend [2011-04-26 12:42:49 2612] DEBUG (XendCheckpoint:127) In saveInputHandler suspend [2011-04-26 12:42:49 2612] DEBUG (XendCheckpoint:129) Suspending 13 ... [2011-04-26 12:42:49 2612] DEBUG (XendDomainInfo:519) XendDomainInfo.shutdown(suspend) [2011-04-26 12:42:49 2612] DEBUG (XendDomainInfo:1891) XendDomainInfo.handleShutdownWatch [2011-04-26 12:42:49 2612] DEBUG (XendDomainInfo:1891) XendDomainInfo.handleShutdownWatch [2011-04-26 12:42:49 2612] INFO (XendDomainInfo:2088) Domain has shutdown: name=migrating-Sandbox-Natty id=13 reason=suspend. [2011-04-26 12:42:49 2612] INFO (XendCheckpoint:135) Domain 13 suspended. [2011-04-26 12:42:50 2612] INFO (image:538) signalDeviceModel:restore dm state to running [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) SUSPEND shinfo 00001d05 [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) delta 276ms, dom0 1%, target 1%, sent 5Mb/s, dirtied 38Mb/s 325 pages [2011-04-26 12:42:50 2612] DEBUG (XendCheckpoint:144) Written done [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) Saving memory pages: iter 5 0%^M 5: sent 325, skipped 0, delta 10ms, dom0 20%, target 0%, sent 1064Mb/s, dirtied 1064Mb/s 325 pages [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) Total pages sent267532 (0.26x) [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) (of which 0 were fixups) [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) All memory is saved [2011-04-26 12:42:50 2612] INFO (XendCheckpoint:423) Save exit rc=0 [2011-04-26 12:42:50 2612] DEBUG (XendDomainInfo:3053) XendDomainInfo.destroy: domid=13 The xend.log from the target (xen 4.1) machine: [...] [2011-04-26 12:42:25 29344] INFO (image:822) Need to create platform device.[domid:26] [2011-04-26 12:42:25 29344] DEBUG (XendCheckpoint:278) restore:shadow=0xa, _static_max=0x40000000, _static_min=0x0, [2011-04-26 12:42:25 29344] DEBUG (XendCheckpoint:305) [xc_restore]: /usr/lib/xen/bin/xc_restore 18 26 3 4 1 1 1 0 [2011-04-26 12:42:39 29344] INFO (XendCheckpoint:423) xc: error: 0-length read: Internal error [2011-04-26 12:42:39 29344] INFO (XendCheckpoint:423) xc: error: read_exact_timed failed (read rc: 0, errno: 0): Internal error [2011-04-26 12:42:39 29344] INFO (XendCheckpoint:423) xc: error: Error when reading batch size (0 = Success): Internal error [2011-04-26 12:42:39 29344] INFO (XendCheckpoint:423) xc: error: error when buffering batch, finishing (0 = Success): Internal error [2011-04-26 12:42:39 29344] DEBUG (XendCheckpoint:394) store-mfn 1044476 [2011-04-26 12:42:39 29344] DEBUG (XendDomainInfo:3010) XendDomainInfo.completeRestore [...]> Thanks, > Todd >-- Mark Dokter - System Administrator Institute for Computergraphics and Vision - http://www.icg.tugraz.at Graz University of Technology, Inffeldgasse 16/II, 8010 Graz, Austria Phone: +43/316/873-5036, Fax: +43/316/873-5050 _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On 04/26/2011 01:14 PM, Mark Dokter wrote:> On 04/13/2011 08:11 PM, Todd Deshane wrote: >> >> There have been some fixes (not sure if they are related) to the PV on >> HVM drivers recently. Take a look at: >> >> http://xen.markmail.org/search/?q=rc3+from:"Stefano+Stabellini"#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results >> > > I''ll try a 2.6.39rc4 domU kernel next. >The 2.6.39-rc4 kernel from [1] hangs at boot after the message [ 3.323128] Initialising Xen virtual ethernet driver. I tried to play with xen_emul_unplug, but that didn''t help :( [1] http://kernel.ubuntu.com/~kernel-ppa/mainline/v2.6.39-rc4-natty/ -- Mark Dokter - System Administrator Institute for Computergraphics and Vision - http://www.icg.tugraz.at Graz University of Technology, Inffeldgasse 16/II, 8010 Graz, Austria Phone: +43/316/873-5036, Fax: +43/316/873-5050 _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On Tue, Apr 26, 2011 at 7:53 AM, Mark Dokter <dokter@icg.tugraz.at> wrote:> On 04/26/2011 01:14 PM, Mark Dokter wrote: >> On 04/13/2011 08:11 PM, Todd Deshane wrote: >>> >>> There have been some fixes (not sure if they are related) to the PV on >>> HVM drivers recently. Take a look at: >>> >>> http://xen.markmail.org/search/?q=rc3+from:"Stefano+Stabellini"#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results >>> >> >> I''ll try a 2.6.39rc4 domU kernel next. >> > > The 2.6.39-rc4 kernel from [1] hangs at boot after the message > [ 3.323128] Initialising Xen virtual ethernet driver. > > I tried to play with xen_emul_unplug, but that didn''t help :( > >Is this still an issue or was it resolved with the natty release? If it still isn''t working you may want to post detailed setup and errors to the xen-devel mailing list.> [1] > http://kernel.ubuntu.com/~kernel-ppa/mainline/v2.6.39-rc4-natty/ > > -- > Mark Dokter - System Administrator > Institute for Computergraphics and Vision - http://www.icg.tugraz.at > Graz University of Technology, Inffeldgasse 16/II, 8010 Graz, Austria > Phone: +43/316/873-5036, Fax: +43/316/873-5050 > > _______________________________________________ > Xen-users mailing list > Xen-users@lists.xensource.com > http://lists.xensource.com/xen-users >-- Todd Deshane http://www.linkedin.com/in/deshantm http://www.xen.org/products/cloudxen.html _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
On 05/02/2011 05:47 AM, Todd Deshane wrote:> On Tue, Apr 26, 2011 at 7:53 AM, Mark Dokter <dokter@icg.tugraz.at> wrote: >> On 04/26/2011 01:14 PM, Mark Dokter wrote: >>> On 04/13/2011 08:11 PM, Todd Deshane wrote: >>>> >>>> There have been some fixes (not sure if they are related) to the PV on >>>> HVM drivers recently. Take a look at: >>>> >>>> http://xen.markmail.org/search/?q=rc3+from:"Stefano+Stabellini"#query:rc3%20from%3A%22Stefano%20Stabellini%22+page:1+mid:fnbwmyqwsspi46gi+state:results >>>> >>> >>> I''ll try a 2.6.39rc4 domU kernel next. >>> >> >> The 2.6.39-rc4 kernel from [1] hangs at boot after the message >> [ 3.323128] Initialising Xen virtual ethernet driver. >> >> I tried to play with xen_emul_unplug, but that didn''t help :( >> >> > > Is this still an issue or was it resolved with the natty release? > > If it still isn''t working you may want to post detailed setup and > errors to the xen-devel mailing list. >Unfortunately, it''s still an issue. I''ll continue discussing this issue on xen-devel when I have my second 4.1 box up and running (then I can test migration from 4.0.1 <=> 4.1 and 4.1 <=> 4.1) The natty 2.6.39rc4 and rc5 don''t boot at all in my domU. Standard natty works fine. kind regards, Mark _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
Mark Dokter
2011-May-16 09:30 UTC
Re: [Xen-users] Live migration on xen 4.0.1 fails (works on 4.1)
On 05/02/2011 12:34 PM, Mark Dokter wrote:> On 05/02/2011 05:47 AM, Todd Deshane wrote: >> >> Is this still an issue or was it resolved with the natty release? >> >> If it still isn''t working you may want to post detailed setup and >> errors to the xen-devel mailing list. >> > > Unfortunately, it''s still an issue. I''ll continue discussing this issue > on xen-devel when I have my second 4.1 box up and running (then I can > test migration from 4.0.1 <=> 4.1 and 4.1 <=> 4.1) >Whatever caused my earlier migration attemts to fail is now gone with two xen 4.1 boxes and a recent domU kernel (debian unstable 2.6.38). My test setup is now: 1 debian squeeze dom0 box with xen 4.1 from debian unstable and linux-image-2.6.32-5-xen-amd64 1 ubuntu natty dom0 box with recompiled xen 4.1 debian unstable packages and a self compiled dom0 kernel from the 2.6.32-stable git. With this setup I tested migration with file:/<pathto>/domU.img configuration and nfs shared storage and with a phy:/<pathto-drbd-device> (dual primary) setup. Both methods seem to work fine with my debian-unstable test-domU. With a debian squeeze (2.6.32 distribution kernel) the migration works on the xen-dom0 part, but the domU kernel doesn''t seem to handle the save/restore too well and freezes.> The natty 2.6.39rc4 and rc5 don''t boot at all in my domU. Standard natty > works fine.Didn''t have time to test any new 2.6.39rc* images... cheers, Mark _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users