Displaying 20 results from an estimated 106 matches for "scsi_devices".
Did you mean:
scsi_device
2020 Mar 11
6
[PATCH RFC v2 02/24] scsi: allocate separate queue for reserved commands
On Tue, Mar 10, 2020 at 09:08:56PM +0000, John Garry wrote:
> On 10/03/2020 18:32, Christoph Hellwig wrote:
> > On Wed, Mar 11, 2020 at 12:25:28AM +0800, John Garry wrote:
> > > From: Hannes Reinecke <hare at suse.com>
> > >
> > > Allocate a separate 'reserved_cmd_q' for sending reserved commands.
> >
> > Why? Reserved command
2020 Mar 11
6
[PATCH RFC v2 02/24] scsi: allocate separate queue for reserved commands
On Tue, Mar 10, 2020 at 09:08:56PM +0000, John Garry wrote:
> On 10/03/2020 18:32, Christoph Hellwig wrote:
> > On Wed, Mar 11, 2020 at 12:25:28AM +0800, John Garry wrote:
> > > From: Hannes Reinecke <hare at suse.com>
> > >
> > > Allocate a separate 'reserved_cmd_q' for sending reserved commands.
> >
> > Why? Reserved command
2020 Apr 23
0
[PATCH RFC v2 02/24] scsi: allocate separate queue for reserved commands
On 4/23/20 4:13 PM, John Garry wrote:
> On 07/04/2020 17:30, Christoph Hellwig wrote:
>> On Tue, Apr 07, 2020 at 04:00:10PM +0200, Hannes Reinecke wrote:
>>> My concern is this:
>>>
>>> struct scsi_device *scsi_get_host_dev(struct Scsi_Host *shost)
>>> {
>>> ????[ .. ]
>>> ????starget = scsi_alloc_target(&shost->shost_gendev, 0,
2013 Nov 06
3
Calling 'virsh' from within a hook script - avoid deadlock?
...hile attempting to do so.
Is there any workaround around this? Any better way, forking virsh execution somehow?
Could anyone explain why this is creating a deadlock in the first place? Isn't libvirt multithreaded?
Thanks for the help!
Shlomi
### hook script:
#!/bin/bash
enum_scsi() {
scsi_devices=`cd /dev;ls sd[^a]*`
}
gen_hba_xml() {
cat > ./hba.xml << EOF
<controller type='scsi' model='virtio-scsi'/>
EOF
}
passthrough_start() {
# create virtio-scsi HBA
gen_hba_xml
virsh attach-device --persistent $domain_name ./hba.xml
rm -f ./hba.x...
2020 Apr 07
0
[PATCH RFC v2 02/24] scsi: allocate separate queue for reserved commands
On 4/7/20 1:54 PM, John Garry wrote:
> On 06/04/2020 10:05, Hannes Reinecke wrote:
>> On 3/11/20 7:22 AM, Christoph Hellwig wrote:
>>> On Tue, Mar 10, 2020 at 09:08:56PM +0000, John Garry wrote:
>>>> On 10/03/2020 18:32, Christoph Hellwig wrote:
>>>>> On Wed, Mar 11, 2020 at 12:25:28AM +0800, John Garry wrote:
>>>>>> From: Hannes
2008 May 30
5
[PATCH 1/4] pvSCSI driver
pvSCSI backend driver
Signed-off-by: Tomonari Horikoshi <t.horikoshi@jp.fujitsu.com>
Signed-off-by: Jun Kamada <kama@jp.fujitsu.com>
-----
Jun Kamada
_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xensource.com
http://lists.xensource.com/xen-devel
2019 Dec 11
0
[PATCH 15/24] compat_ioctl: scsi: move ioctl handling into drivers
Each driver calling scsi_ioctl() gets an equivalent compat_ioctl()
handler that implements the same commands by calling scsi_compat_ioctl().
The scsi_cmd_ioctl() and scsi_cmd_blk_ioctl() functions are compatible
at this point, so any driver that calls those can do so for both native
and compat mode, with the argument passed through compat_ptr().
With this, we can remove the entries from
2019 Dec 11
3
[PATCH 00/24] block, scsi: final compat_ioctl cleanup
Hi Jens, James and Martin,
This series concludes the work I did for linux-5.5 on the compat_ioctl()
cleanup, killing off fs/compat_ioctl.c and block/compat_ioctl.c by moving
everything into drivers.
Overall this would be a reduction both in complexity and line count, but
as I'm also adding documentation the overall number of lines increases
in the end.
My plan was originally to keep the
2011 Nov 16
4
not using LVM for Linux VM guests?
I came across an old post comment yesterday (from http://echenh.blogspot.com/2010/04/how-to-extend-lvm-on-vmware-guest-os.html ) discussing the "hack" of LVM on Linux VM guests and whether it's better not to use it to simplify disk management.
I've re-posted the comment below, does it sound reasonable? Is it better to not use LVM on Linux VM guests?
--Russell
2020 Jan 02
1
[PATCH v3 13/22] compat_ioctl: scsi: move ioctl handling into drivers
Each driver calling scsi_ioctl() gets an equivalent compat_ioctl()
handler that implements the same commands by calling scsi_compat_ioctl().
The scsi_cmd_ioctl() and scsi_cmd_blk_ioctl() functions are compatible
at this point, so any driver that calls those can do so for both native
and compat mode, with the argument passed through compat_ptr().
With this, we can remove the entries from
2014 May 24
0
[PATCH] virtio-scsi: Implement change_queue_depth for virtscsi targets
change_queue_depth allows changing per-target queue depth via sysfs.
It also allows the SCSI midlayer to ramp down the number of concurrent
inflight requests in response to a SCSI BUSY status response and allows
the midlayer to ramp the count back up to the device maximum when the
BUSY condition has resolved.
Signed-off-by: Venkatesh Srinivas <venkateshs at google.com>
---
2014 May 24
0
[PATCH] virtio-scsi: Implement change_queue_depth for virtscsi targets
change_queue_depth allows changing per-target queue depth via sysfs.
It also allows the SCSI midlayer to ramp down the number of concurrent
inflight requests in response to a SCSI BUSY status response and allows
the midlayer to ramp the count back up to the device maximum when the
BUSY condition has resolved.
Signed-off-by: Venkatesh Srinivas <venkateshs at google.com>
---
2020 Apr 07
0
[PATCH RFC v2 02/24] scsi: allocate separate queue for reserved commands
On 4/7/20 4:35 PM, John Garry wrote:
> On 07/04/2020 15:00, Hannes Reinecke wrote:
>> On 4/7/20 1:54 PM, John Garry wrote:
>>> On 06/04/2020 10:05, Hannes Reinecke wrote:
[ .. ]
>>>> This would be okay if 'this_id' would have been defined by the driver;
>>>> sadly, most drivers which are affected here do set 'this_id' to -1.
2006 Aug 02
0
[PATCH 0/6] SCSI frontend and backend drivers
This patchset includes an updated version of the SCSI frontend and
backend drivers.
The frontend and backend drivers exchange SCSI RDMA Protocol (SRP)
messages via a ring buffer. The backend driver sends SCSI commands to
the user-space daemon, which performs SCSI commands and I/O
operations. The backend driver uses VM_FOREIGN feature like the blktap
driver for zero-copy of data pages.
Like the
2016 Mar 25
2
Odd behavior of a CentOS 7 box after repair of an external RAID
I don't think I've seen this with CentOS 6 or 5, but I had to repair an
external RAID box this morning. The server, running CentOS 7, has an LSI
HBA card in it, and it's presented to the system as /dev/sdb. I shut off
the RAID controller, powered it off, did the repair, then plugged
everything back in (that includes the two fiber cables and the ethernet),
and brought the RAID back up.
2014 Oct 04
2
Mounting LUNs from a SAN array - LUN mappings to devices in /dev/ - are they static?
Hi All :)
I am currently involved in a project in which there is a SAN array (Sun
Storagetek 2540) which exports LUNs for some servers with Centos 5.2 x86. I
will be performing a migration to Centos 5.9 x86_64 in some time and am
gathering needed info now :)
I am trying to find the place in the OS where there is the information
about LUN mappings to /dev/ devices.
For example on array level I
2008 Jul 03
3
[PATCH 2/4] pvSCSI : Fix many points of backend/frontend driver
Please refer following Mr. Steven''s mail posted on June 24th.
Message-ID: <20080624131313.GB18379@weybridge.uk.xensource.com>
Message-ID: <20080624131256.GA18379@weybridge.uk.xensource.com>
Signed-off-by: Tomonari Horikoshi <t.horikoshi@jp.fujitsu.com>
Signed-off-by: Jun Kamada <kama@jp.fujitsu.com>
-----
Jun Kamada
2016 Mar 25
2
Odd behavior of a CentOS 7 box after repair of an external RAID
Valeri Galtsev wrote:
>
> On Fri, March 25, 2016 9:55 am, m.roth at 5-cent.us wrote:
>> I don't think I've seen this with CentOS 6 or 5, but I had to repair an
>> external RAID box this morning. The server, running CentOS 7, has an
>> LSI HBA card in it, and it's presented to the system as /dev/sdb.
>> I shut off the RAID controller, powered it off,
>
2015 Dec 21
4
Extending a CentOS disk without reboot
Hello,
Today a virtual server under my administration ran out of disk space, so
I had to get extra space to it.. while it was running, because it's an
important web server without any kind of HA/Load Balancing (don't do
that, kids).
So what happened, is that after another administrator had extended the
disk from VMWare, I tried to extend it from the OS side.
Everything went well,
2012 Jul 03
2
[PATCH v2] virtio-scsi: hotplug support for virtio-scsi
This patch implements the hotplug support for virtio-scsi.
When there is a device attached/detached, the virtio-scsi driver will be
signaled via event virtual queue and it will add/remove the scsi device
in question automatically.
v2: handle no_event event
Signed-off-by: Cong Meng <mc at linux.vnet.ibm.com>
Signed-off-by: Sen Wang <senwang at linux.vnet.ibm.com>
---