similar to: Xen hotplug question & issue

Displaying 20 results from an estimated 20000 matches similar to: "Xen hotplug question & issue"

2008 Aug 04
23
Xen and iSCSI - options and questions
Hello, I have a small Xen farm of 8 dom0 servers with 64 virtual machines running para-virtualized and this has been working great. Unfortunately, I''ve hit a limit: my iSCSI hardware supports only 512 concurrent connections and so I''m pretty much at the limit. (Wish I would have seen that problem sooner!) Of course, 87% of those connections are idle-- but necessary because I
2013 Apr 24
7
[PATCH] hotplug/Linux: add iscsi block hotplug script
This hotplug script has been tested with IET and NetBSD iSCSI targets, without authentication. This hotplug script will only work with PV guests not using pygrub. Signed-off-by: Roger Pau Monné <roger.pau@citrix.com> Cc: Ian Campbell <ian.campbell@citrix.com> Cc: Ian Jackson <ian.jackson@citrix.com> --- Changes due to 4.3 release freeze: * We can no longer provide a
2009 Mar 03
0
HEADS UP: iSCSI, zvol, and vdisk format support
Looks like I forgot to send this to the alias... MRJ ---- A quick heads up for the iscsi, zvol, and vdisk format putback I did into the 3.3 development gate. ISCSI ===== With this putback, you can now install onto and boot a guest using iSCSI disk(s). The iscsi formats supported are: phy:iscsi:/alias/<iscsi-alias> phy:iscsi:/static/<server IP>/<lun>/<target
2009 Jan 19
2
Error on xm create: VmError: (38, ''Function not implemented'')
Hi everyone, I generated my own DomU guest using opensuse''s yast dirinstall and stored it on a seperate iscsi target. I created it as a DomU on my Domain0 (2.6.25.16-0.1-xen kernel) successfully, using the following configuration: name = "vm01-opensuse11-base-LAMP" memory = 256 kernel = "/boot/vmlinuz-xen" ramdisk = "/boot/initrd-xen" root =
2007 Aug 13
0
[BUG] migration problem
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 first of all, thank you for a great project i really enjoy! ok, now here comes the bug report... i''m trying to test xen migration capabilities, with xen-3.1, hand compilled, fetched from mercurial repository. my setup looks like this - x86_64 dom0, x86_32p dom0, x86_32p domU. when trying to migrate from 64 to 32 bit dom0 i''m
2007 Feb 02
2
xen: when devices go away...
Hi there, I''ve been experimenting with iscsi and xen and have found an interesting behavior. I''d reasoned that using iscsi volumes as devices in a software RAID1 configuration would be a Good Idea since when one of the iscsi servers goes away the filesystem is stable. I have a dom0 which has access to two iscsi volumes on two seperate iscsi target servers. I configure a domU
2010 Oct 11
0
Ubuntu iSCSI install to COMSTAR zfs volume Howto
I apologize if this has been covered before. I have not seen a blow-by-blow installation guide for Ubuntu onto an iSCSI target. The install guides I have seen assume that you can make a target visible to all, which is a problem if you want multiple iSCSI installations on the same COMSTAR target. During install Ubuntu generates three random initiators and you have to deal with them to get things
2018 Jan 10
1
Whether libvirt can support all backing chain layer are iscsi network disk type
Hi, For backing chain that all images are iscsi network disk type , such as iscsi://ip/iqn../0(base image) <-iscsi://ip/iqn../1(active image). Currently, 'qemu-img info --backing-chain' command can display correct backing file info, but after starting guest with the active image in backing chain, it don't include the related <backingStore> element in dumpxml. So,
2018 Jan 03
0
questions about disk auth element for Backing Chain
Hi, I am a libvirt qe, I am testing the function about the new location of disk auth element (as sub-element of the source element) in Backing Chain Management. And I have a questions: When all the backing chain are iscsi network disk type(no matter if the authentication is exist), start the guest , it will only has the top level disk in xml and there are no other backingStore in it. The test
2011 Feb 28
2
can't disconnec iSCSI targets, please help
Hi, I'm trying to disconnect some iSCSI targets, but can't seem to. [root at localhost ~]# iscsiadm -m session tcp: [1] 192.168.2.202:3260,1 iqn.2011.01.22.freenas.nvr:500gb tcp: [3] 192.168.2.200:3260,1 iqn.2011-2.za.co.securehosting:RAID.thin3.vg0.1tba tcp: [4] 192.168.2.202:3260,1 iqn.2011.01.22.freenas.nvr:extent0 tcp: [5] 192.168.2.200:3260,1
2013 Dec 23
0
Re: Connect libvirt to iSCSI target
On 12/22/2013 10:09 AM, Marco wrote: > On 2013–12–21 John Ferlan wrote: > >> On 12/17/2013 07:13 PM, Marco wrote: >>> Hi! >>> >>> I'm new to libvirt and face problems connecting to an iSCSI target. >>> What I intend to do is to connect libvirt (I tried virt-manager and >>> virsh) to an iSCSI target and then boot from the LUNs which
2013 Jul 22
1
Bug#717573: XCP: Cannot connect to iSCSI target
Package: xcp-storage-managers Version: 0.1.1-3 Severity: normal XCP cannot use iscsi block devices. Fails with iscsiadm: No active sessions. I ran the following commands trying to diagnose the problem: First introducing a SR and then creating a PBD with the needed configuration for the iscsi block device. These steps work fine but when I tried to plug the PBD I got the error: root at
2007 Feb 24
1
zfs received vol not appearing on iscsi target list
Just installed Nexenta and I''ve been playing around with zfs. root at hzsilo:/tank# uname -a SunOS hzsilo 5.11 NexentaOS_20070105 i86pc i386 i86pc Solaris root at hzsilo:/tank# zfs list NAME USED AVAIL REFER MOUNTPOINT home 89.5K 219G 32K /export/home tank 330K 1.78T 51.9K /tank tank/iscsi_luns 147K
2013 Dec 22
2
Re: Connect libvirt to iSCSI target
On 2013–12–21 John Ferlan wrote: > On 12/17/2013 07:13 PM, Marco wrote: > > Hi! > > > > I'm new to libvirt and face problems connecting to an iSCSI target. > > What I intend to do is to connect libvirt (I tried virt-manager and > > virsh) to an iSCSI target and then boot from the LUNs which contain > > the VMs. > > > > I followed the
2014 Jul 04
2
iSCSI initiator iqn
Hi, I could not find any option to set iSCSI initiator iqn while using guestfish, although the underlying qemu command has this option. It appears that each time guestfish tries to connect to iSCSI LUN, a randomly generated initiator iqn is being used. This is preventing guestfish to connect to the iSCSI target in our environment as the target allows incoming connection based on the preconfigured
2014 Nov 16
2
Problem with Xen4CentOS
Hi folks, we (the company i am working for) are running several dozens of virtualisation servers using CentOS 6 + Xen4CentOS as the virtualisation infrastructure. With the latest versions of all packages installed [1], we see failures in live-migration of stock-CentOS6 HVM guests, leaving a "Domain-unnamed" on the source host, while the migrated guest runs fine on the target host.
2010 Jun 16
1
how to match the ID of a LUN in a storage pool with the GUID on the target server
I've configured a libvirt storage pool using an iscsi target from a Sun 7310 storage appliance and am using the LUNs in this target as volumes for my KVM guests. The setup is very similar to what Daniel covered in a recent blog posting: http://berrange.com/posts/2010/05/05/provisioning-kvm-virtual-machines-on-iscsi-the-hard-way-part-2-of-2/ It works great, but I can't figure out how
2016 Jan 12
0
Re: What are the advantages and disadvantages of running with or without libvirt?
There's two parts to this question. On Tue, Jan 12, 2016 at 05:26:10PM +0200, Yaniv Kaul wrote: > I didn't see what are the main differences in > http://libguestfs.org/guestfs.3.html#backend The basic concept of the backend is how do we run the libguestfs appliance (http://libguestfs.org/guestfs-internals.1.html#architecture). There are two ways we could run the appliance: either
2013 Dec 21
0
Re: Connect libvirt to iSCSI target
On 12/17/2013 07:13 PM, Marco wrote: > Hi! > > I'm new to libvirt and face problems connecting to an iSCSI target. > What I intend to do is to connect libvirt (I tried virt-manager and > virsh) to an iSCSI target and then boot from the LUNs which contain > the VMs. > > I followed the documentation¹ but got stuck at section 12.1.5.4.3. > > 1) > virsh
2017 Aug 02
2
Libvirt fails on network disk with ISCSI protocol
Hi, I am working on oVirt, and I am trying to run a VM with a network disk with ISCSI protocol ( the storage is on a Cinder server). Here is the disk XML I use: <disk device="disk" snapshot="no" type="network"> <address bus="0" controller="0" target="0" type="drive" unit="0" />