Andre Konopka
2007-Aug-09 06:52 UTC
[Xen-users] migrate a VM to another machine (XEN on SLES10SP1)
Hi
I''m trying to build a HA Xen Cluster on SLES10SP1
I have two machines running a XEN dom0, the two machines share a SAN
disk, installed as OCFS2 partition.
First I installed SLES10 as a guest machine. The image is stored on my
shared partion.
I can start and stop the VM on the machine where I created it without
any problem...
Okay, as next step I changed the xend-config.sxp config file and enabled
relocation.
First I tried an offline migration. Without success...
In the log I found the following information (on the target machine).
Maybe someone can explain it a little bit:
[2007-08-08 14:31:58 xend 4730] ERROR (XendDomain:1011) Restore failed
Traceback (most recent call last):
File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomain.py",
line 1006, in domain_restore_fd
return XendCheckpoint.restore(self, fd, paused=paused)
File
"/usr/lib64/python2.4/site-packages/xen/xend/XendCheckpoint.py",
line 173, in restore
dominfo.waitForDevices() # Wait for backends to set up
File
"/usr/lib64/python2.4/site-packages/xen/xend/XendDomainInfo.py",
line 503, in waitForDevices
self.getDeviceController(devclass).waitForDevices()
File
"/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py",
line 149, in waitForDevices
return map(self.waitForDevice, self.deviceIDs())
File
"/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py",
line 168, in waitForDevice
raise VmError("Device %s (%s) could not be connected. "
VmError: Device 51728 (vbd) could not be connected. Backend device not
found.
For me it seems that the second machine can''t activate the disc???
On the first machine my shared partion is mounted as /dev/sdc
Device File: /dev/sdc (/dev/sg3)
Device Files: /dev/sdc,
/dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000,
/dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:3,
/dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5,
/dev/disk/by-label/virtmach
on the second machine as /dev/sdb
Device File: /dev/sdb (/dev/sg2)
Device Files: /dev/sdb,
/dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000,
/dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:2,
/dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5,
/dev/disk/by-label/virtmach
This is the config file, created by the graphical
''virt-manager''.
ostype="sles10"
name="sles10-1"
memory=512
vcpus=1
uuid="55adf3ee-9f23-f3d4-6cad-48d4c43cdf84"
on_crash="destroy"
on_poweroff="destroy"
on_reboot="restart"
localtime=0
builder="linux"
bootloader="/usr/lib/xen/boot/domUloader.py"
bootargs="--entry=xvda2:/boot/vmlinuz-xen,/boot/initrd-xen"
extra="TERM=xterm "
disk=[ ''file:/virtmach/images/sles10ak/disk0,xvda,w'',
''file:/isoimages/SLES-10-SP1-x86_64-DVD1.iso,xvdb,r'', ]
vif=[ ''mac=00:16:3e:6f:70:d0'', ]
vfb=["type=vnc,vncunused=1"]
After booting both dom0 machines my fresh installed domU isn''t visible
any longer.
The image ist still on my shared partition but ''xm list'' only
reports
the dom0 instance????
HOW can I ''import'' my ''lost'' domU???
I think this ''import'' is also necessary for my HA solution. If
my first
machine crashes, the second one must be able to ''import'' the
DOMU to
start it???
Best regards
Andre
_______________________________________________
Xen-users mailing list
Xen-users@lists.xensource.com
http://lists.xensource.com/xen-users
trilok nuwal
2007-Aug-09 11:38 UTC
Re: [Xen-users] migrate a VM to another machine (XEN on SLES10SP1)
What i think is disk0 is yr domU image on SAN. /virtmach/images/sles10ak/disk0 Did u mount yr SAN disk partition at common mount point at both the machies. mount /dev/sdc to common mount point like /virtmach on both machine. Beacuse from the log, it looks like on destination machine u r not able to access the disk0. Simple thing is just copy the domU onfig file to other machine and try to craete the domU. If it goes fine then u could try migration. --Triok On 8/9/07, Andre Konopka <andre.konopka@presse-data.de> wrote:> > > Hi > > I''m trying to build a HA Xen Cluster on SLES10SP1 > > I have two machines running a XEN dom0, the two machines share a SAN > disk, installed as OCFS2 partition. > > First I installed SLES10 as a guest machine. The image is stored on my > shared partion. > I can start and stop the VM on the machine where I created it without > any problem... > > Okay, as next step I changed the xend-config.sxp config file and enabled > relocation. > > First I tried an offline migration. Without success... > > In the log I found the following information (on the target machine). > Maybe someone can explain it a little bit: > > [2007-08-08 14:31:58 xend 4730] ERROR (XendDomain:1011) Restore failed > Traceback (most recent call last): > File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomain.py", > line 1006, in domain_restore_fd > return XendCheckpoint.restore(self, fd, paused=paused) > File "/usr/lib64/python2.4/site-packages/xen/xend/XendCheckpoint.py", > line 173, in restore > dominfo.waitForDevices() # Wait for backends to set up > File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomainInfo.py", > line 503, in waitForDevices > self.getDeviceController(devclass).waitForDevices() > File > "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", > line 149, in waitForDevices > return map(self.waitForDevice, self.deviceIDs()) > File > "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", > line 168, in waitForDevice > raise VmError("Device %s (%s) could not be connected. " > VmError: Device 51728 (vbd) could not be connected. Backend device not > found. > > For me it seems that the second machine can''t activate the disc??? > > On the first machine my shared partion is mounted as /dev/sdc > > Device File: /dev/sdc (/dev/sg3) > Device Files: /dev/sdc, > /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, > /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:3, > /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, > /dev/disk/by-label/virtmach > > on the second machine as /dev/sdb > > Device File: /dev/sdb (/dev/sg2) > Device Files: /dev/sdb, > /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, > /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:2, > /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, > /dev/disk/by-label/virtmach > > > This is the config file, created by the graphical ''virt-manager''. > > ostype="sles10" > name="sles10-1" > memory=512 > vcpus=1 > uuid="55adf3ee-9f23-f3d4-6cad-48d4c43cdf84" > on_crash="destroy" > on_poweroff="destroy" > on_reboot="restart" > localtime=0 > builder="linux" > bootloader="/usr/lib/xen/boot/domUloader.py" > bootargs="--entry=xvda2:/boot/vmlinuz-xen,/boot/initrd-xen" > extra="TERM=xterm " > disk=[ ''file:/virtmach/images/sles10ak/disk0,xvda,w'', > ''file:/isoimages/SLES-10-SP1-x86_64-DVD1.iso,xvdb,r'', ] > vif=[ ''mac=00:16:3e:6f:70:d0'', ] > vfb=["type=vnc,vncunused=1"] > > > > After booting both dom0 machines my fresh installed domU isn''t visible > any longer. > The image ist still on my shared partition but ''xm list'' only reports > the dom0 instance???? > > HOW can I ''import'' my ''lost'' domU??? > > I think this ''import'' is also necessary for my HA solution. If my first > machine crashes, the second one must be able to ''import'' the DOMU to > start it??? > > > Best regards > > Andre > > > > > _______________________________________________ > Xen-users mailing list > Xen-users@lists.xensource.com > http://lists.xensource.com/xen-users >_______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
Andre Konopka
2007-Aug-09 13:11 UTC
Re: [Xen-users] migrate a VM to another machine (XEN on SLES10SP1)
Thank you for your hint, the problem was my "cd". On the second machine the path didn''t exist. Thanks Andre trilok nuwal wrote:> What i think is disk0 is yr domU image on SAN. > /virtmach/images/sles10ak/disk0 > > Did u mount yr SAN disk partition at common mount point at both the machies. > > mount /dev/sdc to common mount point like /virtmach on both machine. > > Beacuse from the log, it looks like on destination machine u r not able > to access the disk0. > > Simple thing is just copy the domU onfig file to other machine and try > to craete the domU. If it goes fine then u could try migration. > > --Triok > On 8/9/07, *Andre Konopka* <andre.konopka@presse-data.de > <mailto:andre.konopka@presse-data.de>> wrote: > > > Hi > > I''m trying to build a HA Xen Cluster on SLES10SP1 > > I have two machines running a XEN dom0, the two machines share a SAN > disk, installed as OCFS2 partition. > > First I installed SLES10 as a guest machine. The image is stored on my > shared partion. > I can start and stop the VM on the machine where I created it without > any problem... > > Okay, as next step I changed the xend-config.sxp config file and enabled > relocation. > > First I tried an offline migration. Without success... > > In the log I found the following information (on the target machine). > Maybe someone can explain it a little bit: > > [2007-08-08 14:31:58 xend 4730] ERROR (XendDomain:1011) Restore failed > Traceback (most recent call last): > File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomain.py", > line 1006, in domain_restore_fd > return XendCheckpoint.restore(self, fd, paused=paused) > File > "/usr/lib64/python2.4/site-packages/xen/xend/XendCheckpoint.py", > line 173, in restore > dominfo.waitForDevices() # Wait for backends to set up > File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomainInfo.py", > line 503, in waitForDevices > self.getDeviceController (devclass).waitForDevices() > File > "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", > line 149, in waitForDevices > return map(self.waitForDevice, self.deviceIDs()) > File > "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", > line 168, in waitForDevice > raise VmError("Device %s (%s) could not be connected. " > VmError: Device 51728 (vbd) could not be connected. Backend device not > found. > > For me it seems that the second machine can''t activate the disc??? > > On the first machine my shared partion is mounted as /dev/sdc > > Device File: /dev/sdc (/dev/sg3) > Device Files: /dev/sdc, > /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, > /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:3, > /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, > /dev/disk/by-label/virtmach > > on the second machine as /dev/sdb > > Device File: /dev/sdb (/dev/sg2) > Device Files: /dev/sdb, > /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, > /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:2, > /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, > /dev/disk/by-label/virtmach > > > This is the config file, created by the graphical ''virt-manager''. > > ostype="sles10" > name="sles10-1" > memory=512 > vcpus=1 > uuid="55adf3ee-9f23-f3d4-6cad-48d4c43cdf84" > on_crash="destroy" > on_poweroff="destroy" > on_reboot="restart" > localtime=0 > builder="linux" > bootloader="/usr/lib/xen/boot/domUloader.py" > bootargs="--entry=xvda2:/boot/vmlinuz-xen,/boot/initrd-xen" > extra="TERM=xterm " > disk=[ ''file:/virtmach/images/sles10ak/disk0,xvda,w'', > ''file:/isoimages/SLES-10-SP1-x86_64-DVD1.iso,xvdb,r'', ] > vif=[ ''mac=00:16:3e:6f:70:d0'', ] > vfb=["type=vnc,vncunused=1"] > > > > After booting both dom0 machines my fresh installed domU isn''t visible > any longer. > The image ist still on my shared partition but ''xm list'' only reports > the dom0 instance???? > > HOW can I ''import'' my ''lost'' domU??? > > I think this ''import'' is also necessary for my HA solution. If my first > machine crashes, the second one must be able to ''import'' the DOMU to > start it??? > > > Best regards > > Andre > > > > > _______________________________________________ > Xen-users mailing list > Xen-users@lists.xensource.com <mailto:Xen-users@lists.xensource.com> > http://lists.xensource.com/xen-users > > > > ------------------------------------------------------------------------ > > _______________________________________________ > Xen-users mailing list > Xen-users@lists.xensource.com > http://lists.xensource.com/xen-users_______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users