Displaying 20 results from an estimated 40000 matches similar to: "cannot attach device using libvirt"
2015 Sep 21
2
failure when trying to add a channel device via virsh attach-device
Hi,
On a KVM host installed with following packages versions :
[root@rhel6-roadster test_channels_OK]# rpm -qa | grep libvirt
libvirt-devel-0.10.2-29.el6.x86_64
libvirt-0.10.2-29.el6.x86_64
libvirt-client-0.10.2-29.el6.x86_64
libvirt-java-0.4.9-1.el6.noarch
libvirt-java-devel-0.4.9-1.el6.noarch
libvirt-python-0.10.2-29.el6.x86_64
[root@rhel6-roadster test_channels_OK]# rpm -qa | grep qemu
2013 Jul 16
2
Re: Hotplug of disk devices in LXC failed with libvirt of version 1.0.2
于 2013/7/15 17:32, Gao feng 写道:
> On 07/15/2013 05:18 PM, hzguanqiang wrote:
>> Hi, Gao Feng
>>
>> I've tried what you said, but still exists the problem:
>>
>> ubuntu@lxc:~$ vir attach-disk instance-0000002c /dev/dm-0 sdb
>> error: Failed to attach disk
>> error: Unable to create device /proc/10366/root/dev/sdb: Permission denied
>>
2013 Jul 17
1
Re: Hotplug of disk devices in LXC failed with libvirt of version 1.0.2
On 07/17/2013 04:15 PM, Gao feng wrote:
> On 07/16/2013 01:29 PM, Guan Qiang wrote:
>> 于 2013/7/15 17:32, Gao feng 写道:
>>> On 07/15/2013 05:18 PM, hzguanqiang wrote:
>>>> Hi, Gao Feng
>>>>
>>>> I've tried what you said, but still exists the problem:
>>>>
>>>> ubuntu@lxc:~$ vir attach-disk instance-0000002c
2015 May 08
2
Editing http://libvirt.org/sources/virshcmdref/html-single/#sect-attach-device
I just noticed the entries on attach-disk and attach-devices are
a bit on the empty size. As I am just attaching a usb device to one of
my vm clients, I was wondering if I could offer examples to those
entries.
2013 Jul 15
2
Re: Hotplug of disk devices in LXC failed with libvirt of version 1.0.2
Hi, Gao Feng
I've tried what you said, but still exists the problem:
ubuntu@lxc:~$ vir attach-disk instance-0000002c /dev/dm-0 sdb
error: Failed to attach disk
error: Unable to create device /proc/10366/root/dev/sdb: Permission denied
I think finding what it means by saying 'Unable to create device /proc/10366/root/dev/sdb: Permission denied' is the key.
But anyway, thanks for
2018 Mar 02
2
Re: Fail in virDomainUpdateDeviceFlags (libvirt-4.0.0 + Qemu-kvm 2.9.0 + Ceph 10.2.10)
On Tue, Feb 27, 2018 at 09:53:00 +0100, Michal Privoznik wrote:
> On 02/27/2018 03:06 AM, Star Guo wrote:
> > Hello Everyone,
> >
> >
> >
> > My pc run in CentOS 7.4 and install libvirt-4.0.0 + Qemu-kvm 2.9.0 + Ceph
> > 10.2.10 ALL-in-One.
> >
> >
> >
> > I use python-sdk with libvirt and run [self.domain.updateDeviceFlags(xml,
2015 Jan 19
2
Re: [libvirt] libvirt 1.2.10 and latest EL6 qemu-kvm
[Dropping libvir list and adding libvir-users list.]
On Mon, Jan 19, 2015 at 04:12:46PM +0100, liedekef@telenet.be wrote:
> Hi all,
>
> I had libvirt 1.2.10 running without issues together with the qemu package
> qemu-kvm-0.12.1.2-2.415.el6_5.14.x86_64.rpm
That's a really old QEMU and the libvirt you're using is relatively
newer.
> Now, after updating my server,
2013 Jul 15
2
Hotplug of disk devices in LXC failed with libvirt of version 1.0.2
Hi Daniel,
I noticed that the patch "Add support for hotplug/unplug of disk devices in LXC" you wrote had been merged into libvirt of version 1.0.2.
But when I used this function, it report an error with details as following:
ubuntu@lxc:~$ vir attach-device instance-0000002c disk.xml --config
error: Failed to attach device from disk.xml
error: Unable to create device
2019 Mar 07
1
why attach-disk can't be effective when the guest is booting
Hi,
When i use "virsh attach-device rhel disk.xml" when the guest is
booting,then i can't find the disk in the guest after booting.But i can
find it by "virsh domblklist rhel",and when i want to detach it from the
guest and use "virsh detach-disk rhel vdb", i still can find the disk by
"virsh domblklist rhel"
The disk.xml is like:
<disk
2018 Mar 05
2
Re: Fail in virDomainUpdateDeviceFlags (libvirt-4.0.0 + Qemu-kvm 2.9.0 + Ceph 10.2.10)
On Fri, Mar 02, 2018 at 15:32:44 -0500, John Ferlan wrote:
>
>
> On 03/02/2018 08:28 AM, Peter Krempa wrote:
> > On Tue, Feb 27, 2018 at 09:53:00 +0100, Michal Privoznik wrote:
> >> On 02/27/2018 03:06 AM, Star Guo wrote:
> >>> Hello Everyone,
> >>>
> >>>
> >>>
> >>> My pc run in CentOS 7.4 and install
2013 Oct 15
5
[PATCH net V2 1/2] virtio-net: don't respond to cpu hotplug notifier if we're not ready
We're trying to re-configure the affinity unconditionally in cpu hotplug
callback. This may lead the issue during resuming from s3/s4 since
- virt queues haven't been allocated at that time.
- it's unnecessary since thaw method will re-configure the affinity.
Fix this issue by checking the config_enable and do nothing is we're not ready.
The bug were introduced by commit
2013 Oct 15
5
[PATCH net V2 1/2] virtio-net: don't respond to cpu hotplug notifier if we're not ready
We're trying to re-configure the affinity unconditionally in cpu hotplug
callback. This may lead the issue during resuming from s3/s4 since
- virt queues haven't been allocated at that time.
- it's unnecessary since thaw method will re-configure the affinity.
Fix this issue by checking the config_enable and do nothing is we're not ready.
The bug were introduced by commit
2012 Dec 26
5
[RFC PATCH] virtio-net: reset virtqueue affinity when doing cpu hotplug
Add a cpu notifier to virtio-net, so that we can reset the
virtqueue affinity if the cpu hotplug happens. It improve
the performance through enabling or disabling the virtqueue
affinity after doing cpu hotplug.
Cc: Rusty Russell <rusty at rustcorp.com.au>
Cc: "Michael S. Tsirkin" <mst at redhat.com>
Cc: Jason Wang <jasowang at redhat.com>
Cc: virtualization at
2012 Dec 26
5
[RFC PATCH] virtio-net: reset virtqueue affinity when doing cpu hotplug
Add a cpu notifier to virtio-net, so that we can reset the
virtqueue affinity if the cpu hotplug happens. It improve
the performance through enabling or disabling the virtqueue
affinity after doing cpu hotplug.
Cc: Rusty Russell <rusty at rustcorp.com.au>
Cc: "Michael S. Tsirkin" <mst at redhat.com>
Cc: Jason Wang <jasowang at redhat.com>
Cc: virtualization at
2016 Oct 28
3
sttic vnet device for guest
Can i assign static vnet* device for some guests?
2013 Oct 17
2
[PATCH net V2 1/2] virtio-net: don't respond to cpu hotplug notifier if we're not ready
On Thu, Oct 17, 2013 at 09:57:41AM +1030, Rusty Russell wrote:
> Jason Wang <jasowang at redhat.com> writes:
> > We're trying to re-configure the affinity unconditionally in cpu hotplug
> > callback. This may lead the issue during resuming from s3/s4 since
> >
> > - virt queues haven't been allocated at that time.
> > - it's unnecessary since thaw
2013 Oct 17
2
[PATCH net V2 1/2] virtio-net: don't respond to cpu hotplug notifier if we're not ready
On Thu, Oct 17, 2013 at 09:57:41AM +1030, Rusty Russell wrote:
> Jason Wang <jasowang at redhat.com> writes:
> > We're trying to re-configure the affinity unconditionally in cpu hotplug
> > callback. This may lead the issue during resuming from s3/s4 since
> >
> > - virt queues haven't been allocated at that time.
> > - it's unnecessary since thaw
2013 Jul 17
0
Re: Hotplug of disk devices in LXC failed with libvirt of version 1.0.2
On 07/16/2013 01:29 PM, Guan Qiang wrote:
> 于 2013/7/15 17:32, Gao feng 写道:
>> On 07/15/2013 05:18 PM, hzguanqiang wrote:
>>> Hi, Gao Feng
>>>
>>> I've tried what you said, but still exists the problem:
>>>
>>> ubuntu@lxc:~$ vir attach-disk instance-0000002c /dev/dm-0 sdb
>>> error: Failed to attach disk
>>> error:
2012 Sep 10
1
Reg: Libvirt API does not return, after qemu-kvm is hungup.
Hi All,
I am facing the following issue, while using Libvirt version 0.8.4.
(1) Libvirt API does not return, since the qemu-kvm process is hung up.
Based on the information available on mail thread
*https://bugzilla.redhat.com/show_bug.cgi?id=676205* (comment 26 - scenario
1),
this has been fixed in later version of Libvirt, specifically in RHEL
6.2(libvirt 0.9.4).
I have checked the
2013 Oct 14
4
[PATCH net 1/2] virtio-net: don't respond to cpu hotplug notifier if we're not ready
We're trying to re-configure the affinity unconditionally in cpu hotplug
callback. This may lead the issue during resuming from s3/s4 since
- virt queues haven't been allocated at that time.
- it's unnecessary since thaw method will re-configure the affinity.
Fix this issue by checking the config_enable and do nothing is we're not ready.
The bug were introduced by commit