similar to: libvirt -- add virtio-scsi disk with scsi-generic

Displaying 20 results from an estimated 1000 matches similar to: "libvirt -- add virtio-scsi disk with scsi-generic"

2020 Oct 28
2
proper config for qemu's host_cdrom
Greetings, I was wondering what is the proper way to configure a scsi cdrom pass-through so in when the qemu line is generated, host_Cdrom will be used instead of host_device. looking at https://gitlab.com/libvirt/libvirt/-/blob/master/src/qemu/qemu_block.c#L1090, I see that hostcdrom must be true. in order for that to be true, the following must be (see
2014 Oct 19
2
HostDevev SCSI Devices
http://libvirt.org/formatdomain.html#elementsHostDevSubsys At this description used "drive"-typed address for SCSI. But when i get xml-description of scsi-devices by <code> unsigned int flags = VIR_CONNECT_LIST_NODE_DEVICES_CAP_SCSI_HOST ; int ret = virConnectListAllNodeDevices (currWorkConnect, &nodeDevices, flags); then get by
2016 Feb 08
2
Re: Getting "unknown device type" when attaching NPIV (vHBA) to vm
Thank you for your reply, I'm still getting hard time to attach the hostdev, I couldn't make it work out of the doc. I took the address details from: # ll /sys/bus/scsi/devices/7:* lrwxrwxrwx 1 root root 0 Feb 8 12:58 /sys/bus/scsi/devices/7:0:0:0 -> ../../../devices/pci0000:00/0000:00:03.0/0000:08:00.1/host7/rport-7:0-0/target7:0:0/7:0:0:0 lrwxrwxrwx 1 root root 0 Feb 8 12:58
2016 Feb 01
2
Getting "unknown device type" when attaching NPIV (vHBA) to vm
Hi, I'm trying to attach a host FC HBA to VM (NPIV). Running ovirt 3.5 on CentOS 6.7, libvirt-0.10.2-46.el6_6.6.x86_64, I created vHBA using nodedev-create command - see below HBA and vHBA. I wrote a vdsm hook which appends the vHBA xml stanza to the devices xml element but the vm start starts without the vHBA device. Also, when trying to attach the device manually as a live device - I get:
2019 Oct 14
2
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Hi Cole & Michal, I'm sorry for my late response, I just end my journey today. Thank your response, your suggestion is very helpful to me. I have added Michal in this mail, Michal helps me review my initial patchset. (https://www.spinics.net/linux/fedora/libvir/msg191339.html) All concern about this feature is the XML design. My original XML design exposes more details of Qemu.
2016 Mar 17
1
Questions regarding hostdev scsi
Hi! I'm oVirt developer responsible for most of 'hostdev' support. While working on SCSI passthrough (that is hostdev type='scsi'), I've encountered few issues I'm not sure how to solve somewhat effectively and nicely. Just a note - oVirt by default disables 'dynamic_ownership', meaning we have to handle endpoint ownership/labeling ourselves. This is not
2016 Sep 03
1
Re: attaching storage pool error
Hi again! Need to add here, since I been investigating a bit further, that it seems to be possible to add an NPIV vhba to the guest, using vdsm hooks. I've seen the reference to "host device assignments", which Dan pointed me too, but I didn't really understand the meaning of it at that moment. It would be interesting to hear from someone that have been doing it, how it would
2019 Oct 15
1
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Cole Robinson <crobinso@redhat.com> 于2019年10月15日周二 上午1:48写道: > > On 10/14/19 3:12 AM, Li Feng wrote: > > Hi Cole & Michal, > > > > I'm sorry for my late response, I just end my journey today. > > Thank your response, your suggestion is very helpful to me. > > > > I have added Michal in this mail, Michal helps me review my initial patchset.
2020 Nov 02
0
Re: proper config for qemu's host_cdrom
On Wed, Oct 28, 2020 at 15:06:51 +0100, daggs wrote: > Greetings, > > I was wondering what is the proper way to configure a scsi cdrom pass-through so in when the qemu line is generated, host_Cdrom will be used instead of host_device. > > looking at https://gitlab.com/libvirt/libvirt/-/blob/master/src/qemu/qemu_block.c#L1090, I see that hostcdrom must be true. > in order for
2013 Feb 12
6
[PATCH v3 0/5] virtio-scsi multiqueue
This series implements virtio-scsi queue steering, which gives performance improvements of up to 50% (measured both with QEMU and tcm_vhost backends). The patches build on top of the new virtio APIs at http://permalink.gmane.org/gmane.linux.kernel.virtualization/18431; the new API simplifies the locking of the virtio-scsi driver nicely, thus it makes sense to require them as a prerequisite.
2013 Feb 12
6
[PATCH v3 0/5] virtio-scsi multiqueue
This series implements virtio-scsi queue steering, which gives performance improvements of up to 50% (measured both with QEMU and tcm_vhost backends). The patches build on top of the new virtio APIs at http://permalink.gmane.org/gmane.linux.kernel.virtualization/18431; the new API simplifies the locking of the virtio-scsi driver nicely, thus it makes sense to require them as a prerequisite.
2020 Oct 14
2
scsi passthrough differs between guests
Greetings, I have two machines running the same distro, both running qemu 5.1.0, one runs libvirt 6.7.0, the other 6.8.0. I've decided to test the viability of passing through my sata cdrom into a vm, so I went to the libvirt docs, read a bit and added the following to a debian10 uefi vm running on libvirt 6.8.0: <controller type='scsi' index='0'
2016 Sep 03
2
Ang: Ang: Re: Ang: Ang: Re: Ang: Re: attaching storage pool error
Hi! Report from my multipath tests today. My test virtual machine, that runs from an NPIV pool, is not able to use multipath. When I pulled the cable from one of the targets, it crashed. But, strangely, it could boot up again on that other path, that it just crashed on. That tells me it can use both paths, and is not limited to one of them only, but because the multipath layer isn't
2010 Aug 31
3
Automatically detecting LUNs without a reboot
Hi everybody. I have an HP StorageWorks array connected to one of my servers, connected via an LSI controller (lspci reports: SCSI storage controller: LSI Logic / Symbios Logic SAS1068E PCI-Express Fusion-MPT SAS (rev 08)) on an HP DL385 running CentOS 5.5 x86_64. I have set up multipath and all is working but, when I create a snaphost and export it as a new LUN from the HP web interface, it
2014 Apr 14
2
Re: LXC + USB passthrough = Operation not permitted
Can you see the USB device in the lxc? My answering assumes you cannot see the USB device in the container. If you can see it, please ignore my answer. According to my understanding, “hostdev mode='subsystem' “ is used with KVM only. We use this if we want to passthrough a device to a KVM virtual machine. After doing that, the host machine will lose control of the device. If container is
2019 Apr 24
2
PCI passthrough and abstraction
When you pass a device in the pci chain (after virsh nodedev-dettach'ing it from host) to the guest, how much is passed without being emulated/abstracted?
2019 Aug 22
2
Re: RLIMIT_MEMLOCK in container environment
On Thu, Aug 22, 2019 at 2:24 AM Daniel P. Berrangé <berrange@redhat.com> wrote: > > On Wed, Aug 21, 2019 at 01:37:21PM -0700, Ihar Hrachyshka wrote: > > Hi all, > > > > KubeVirt uses libvirtd to manage qemu VMs represented as Kubernetes > > API resources. In this case, libvirtd is running inside an > > unprivileged pod, with some host mounts / capabilities
2015 Apr 14
1
HBA enumeration and multipath configuration
# cat /etc/redhat-release CentOS Linux release 7.1.1503 (Core) # uname -r 3.10.0-123.20.1.el7.x86_64 Hi, We use iSCSI over a 10G Ethernet Adapter and SRP over an Infiniband adapter to provide multipathing to our storage: # lspci | grep 10-Gigabit 81:00.0 Ethernet controller: Intel Corporation 82599ES 10-Gigabit SFI/SFP+ Network Connection (rev 01) 81:00.1 Ethernet controller: Intel Corporation
2017 Dec 21
2
Cannot Write xvga=yes in xml file
Hello,my system info: virsh verison Compiled against library: libvirt 3.9.0 Using library: libvirt 3.9.0 Using API: QEMU 3.9.0 Running hypervisor: QEMU 2.10.90 And i want to pass through my gpu to my window 7 guest with <hostdev>, next line is comand i use: <hostdev mode='subsystem' type='pci' managed='yes' xvga='yes'>
2013 Oct 28
5
[PATCH] virtio-scsi: Fix hotcpu_notifier use-after-free with virtscsi_freeze
vqs are freed in virtscsi_freeze but the hotcpu_notifier is not unregistered. We will have a use-after-free usage when the notifier callback is called after virtscsi_freeze. Signed-off-by: Asias He <asias at redhat.com> --- drivers/scsi/virtio_scsi.c | 15 ++++++++++++++- 1 file changed, 14 insertions(+), 1 deletion(-) diff --git a/drivers/scsi/virtio_scsi.c b/drivers/scsi/virtio_scsi.c