Displaying 20 results from an estimated 2000 matches similar to: "Problem booting guest with more than 8 disks"
2013 Nov 25
0
Re: Problem Connecting to RBD images using Sys::Guestfs Perl Module
On Mon, Nov 25, 2013 at 12:52:21PM -0700, Jon wrote:
> Hello,
>
> I'm having trouble connecting to rbd images. It seems like somewhere the
> name is getting chewed up. I wonder if this is related to my previous
> troubles [1] [2] with rbd images.
>
> I'm trying to add an rbd image, but when I launch the guestfs object I get
> an error:
>
> >>
2013 Nov 25
0
Re: Problem Connecting to RBD images using Sys::Guestfs Perl Module
Hello Richard,
I'm using:
>> virt-sysprep 1.25.6
I honestly don't remember if the currently installed version is from the
tarball or pulled from the github source... Now that you mention it, I
think it could even be the version with my "hack" to fix virt-sysprep...
Which version are you building? I see there's a:
>>
2013 Nov 25
2
Re: Problem Connecting to RBD images using Sys::Guestfs Perl Module
On Mon, Nov 25, 2013 at 09:58:50PM +0000, Richard W.M. Jones wrote:
> On Mon, Nov 25, 2013 at 12:52:21PM -0700, Jon wrote:
> > Hello,
> >
> > I'm having trouble connecting to rbd images. It seems like somewhere the
> > name is getting chewed up. I wonder if this is related to my previous
> > troubles [1] [2] with rbd images.
> >
> > I'm
2013 Nov 25
4
Problem Connecting to RBD images using Sys::Guestfs Perl Module
Hello,
I'm having trouble connecting to rbd images. It seems like somewhere the
name is getting chewed up. I wonder if this is related to my previous
troubles [1] [2] with rbd images.
I'm trying to add an rbd image, but when I launch the guestfs object I get
an error:
>> libguestfs: trace: launch = -1 (error)
I'm adding a single RBD
>> libguestfs: trace: add_drive
2013 Nov 05
0
Re: Problem using virt-sysprep with RBD images
On Mon, Nov 04, 2013 at 06:35:42PM -0700, Jon wrote:
> Hello,
>
> I'm having a problem when trying to use virt-sysprep against vms that have
> rbd disk images. When I run virt-sysprep I get the following error:
>
> >> root@kitt:~/libguestfs-1.22.4# virt-sysprep -d server-clone-test
> --firstboot firstboot.sh
> >> Examining the guest ...
> >>
2013 Nov 05
2
Problem using virt-sysprep with RBD images
Hello,
I'm having a problem when trying to use virt-sysprep against vms that have
rbd disk images. When I run virt-sysprep I get the following error:
>> root@kitt:~/libguestfs-1.22.4# virt-sysprep -d server-clone-test
--firstboot firstboot.sh
>> Examining the guest ...
>> Fatal error: exception Guestfs.Error("rbd: image name must begin with a
'/'")
My
2013 Nov 06
2
Re: Problem using virt-sysprep with RBD images
Hello Richard,
Thanks for the reply.
Indeed this behaviour exists in 1.25.6.
Grepping through the source [1], there are a number of files in
"./po/*.po[t]?" that contain this message, but I think it's ./src/drives.c
where the fail condition is actually detected / set.
On line 300 there is an if statement that checks the first character is a
slash:
if (exportname[0] !=
2013 Nov 06
3
Re: Problem using virt-sysprep with RBD images
Hello Richard,
Haha, ok, here's a good one: I commented that if statement out at line 300,
applied your patch (I see you updated the github of this code, perhaps
that's the best place to grab the code from), and when I run virt-sysprep,
I get the following parameter for my disk drive:
>> qemu-system-x86_64: -drive
2013 Nov 06
0
Re: Problem using virt-sysprep with RBD images
On Wed, Nov 06, 2013 at 02:25:54PM -0700, Jon wrote:
> Hello Richard,
>
> Haha, ok, here's a good one: I commented that if statement out at line 300,
> applied your patch (I see you updated the github of this code, perhaps
> that's the best place to grab the code from), and when I run virt-sysprep,
> I get the following parameter for my disk drive:
>
> >>
2013 Nov 06
1
Re: Problem using virt-sysprep with RBD images
Hello Rich,
Interesting. Thanks for the explanation.
When you specify an rbd on the command line for virt-sysprep, do you expect
the path to include the monitor address?
e.g.:
>> virt-sysprep -a rbd://host-name/pool-name/device-name
If I understand correctly, libvirt is able to understand the ceph
configuration, so when I create a device with qemu-img I only specify the
protocol and
2012 Sep 18
1
libvirt 0.10 and cephx
Hello,
Current version adding 'auth_supported=none' at the end of disk path
and failing to do an authentication with specified cephx key:
<source protocol='rbd'
name='rbd/vmxxxxxxxxx:id=qemukvm:key=[cut]==:auth_supported=cephx'>
<host name='10.0.0.13' port='6789'/>
<host name='10.0.0.10' port='6789'/>
2018 May 27
1
Using libvirt to access Ceph RBDs with Xen
Hi everybody,
my background: I'm doing Xen since 10++ years, many years with DRBD for
high availability, since some time I'm using preferable GlusterFS with
FUSE as replicated storage, where I place the image-files for the vms.
In my current project we started (successfully) with Xen/GlusterFS too,
but because the provider, where we placed the servers, uses widely CEPH,
we decided to
2015 Mar 31
0
Re: couple of ceph/rbd questions
On 03/31/2015 11:47 AM, Brian Kroth wrote:
> Hi, I've recently been working on setting up a set of libvirt compute
> nodes that will be using a ceph rbd pool for storing vm disk image
> files. I've got a couple of issues I've run into.
>
> First, per the standard ceph documentation examples [1], the way to add
> a disk is to create a block in the VM definition XML
2019 Apr 01
2
Re: guestfish Remote Images IPv6 Support
I believe the bug lies in libguestfs.
Taking out the commands being sent to QEMU and using qemu-img info I can
recreate the error:
# qemu-img info
"rbd:images/CentOS-7-x86_64-GenericCloud-1901:mon_host=[fd00::cefc:1]\:6789:auth_supported=none"
qemu-img: Could not open
'rbd:images/CentOS-7-x86_64-GenericCloud-1901:mon_host=[fd00::cefc:1]\:6789:auth_supported=none':
invalid
2019 Apr 01
1
Re: guestfish Remote Images IPv6 Support
This worked wonderfully! What are the odds of getting this upstream in
the near future? I'd rather not build from source in production.
# ./run guestfish --format=raw --ro -a
rbd://[fd00::cefc:1]:6789/images/CentOS-7-x86_64-GenericCloud-1901
libguestfs: trace: set_verbose true
libguestfs: trace: set_verbose = 0
libguestfs: trace: set_tmpdir "/root/libguestfs/tmp"
libguestfs:
2018 Aug 06
0
Re: ceph rbd pool and libvirt manageability (virt-install)
Hello everybody,
virt-install --version
1.4.0
How do I create a ceph network disk with virt-install without having to
edit it?
<disk type='network' device='disk'>
<driver name='qemu' type='raw'/>
<auth username='libvirt'>
<secret type='ceph' uuid='ec9be0c4-a60f-490e-af83-f0f27aaf48c9'/>
2015 Mar 31
2
couple of ceph/rbd questions
Hi, I've recently been working on setting up a set of libvirt compute
nodes that will be using a ceph rbd pool for storing vm disk image
files. I've got a couple of issues I've run into.
First, per the standard ceph documentation examples [1], the way to add a
disk is to create a block in the VM definition XML that looks something
like this:
<disk type='network'
2015 Oct 05
0
Storage statistics
Hi all,
I'd like to get a few storage statistics from a running vm.
I'd like to know read and write data about QCOW2 devices and about RBD
volumes.
I did try to execute a command similar to this:
virsh qemu-monitor-command vm-vnc '{ "execute": "query-block"}'
as described here:
2019 Mar 29
2
guestfish Remote Images IPv6 Support
I have scoured the web and can't find anything on the topic: Is IPv6
supported for remote image targets?
For example:
guestfish --format=raw --ro -a
rbd://[fd00::cefc:1]:6789/images/CentOS-7-x86_64-GenericCloud-1901
Does not work citing the following:
libguestfs: trace: set_verbose true
libguestfs: trace: set_verbose = 0
libguestfs: create: flags = 0, handle = 0x5560231bdfb0, program =
2019 Mar 30
0
Re: guestfish Remote Images IPv6 Support
On Fri, Mar 29, 2019 at 06:17:17PM -0500, Jonathan Wright wrote:
> I have scoured the web and can't find anything on the topic: Is IPv6
> supported for remote image targets?
It definitely should work, although I don't know if anyone has tried
using the address explicitly (rather than a hostname which resolves to
an AAAA record).
> For example:
>
> guestfish --format=raw