Displaying 20 results from an estimated 4000 matches similar to: "[Qemu-devel] virtio-console downgrade the virtio-pci-blk performance"
2019 Sep 03
2
[PATCH v3 00/13] virtio-fs: shared file system for virtual machines
On Tue, Sep 03, 2019 at 10:05:02AM +0200, Miklos Szeredi wrote:
> [Cc: virtualization at lists.linux-foundation.org, "Michael S. Tsirkin"
> <mst at redhat.com>, Jason Wang <jasowang at redhat.com>]
>
> It'd be nice to have an ACK for this from the virtio maintainers.
>
> Thanks,
> Miklos
Can the patches themselves be posted to the relevant list(s)
2019 Sep 03
2
[PATCH v3 00/13] virtio-fs: shared file system for virtual machines
On Tue, Sep 03, 2019 at 10:05:02AM +0200, Miklos Szeredi wrote:
> [Cc: virtualization at lists.linux-foundation.org, "Michael S. Tsirkin"
> <mst at redhat.com>, Jason Wang <jasowang at redhat.com>]
>
> It'd be nice to have an ACK for this from the virtio maintainers.
>
> Thanks,
> Miklos
Can the patches themselves be posted to the relevant list(s)
2014 Mar 23
0
for Chris Mason ( iowatcher graphs)
Hello. Sorry for writing to btrfs mailing list, but personal mail
reject my message.
Saying "
<chris.mason@fusionio.com>: host 10.101.1.19[10.101.1.19] said: 554 5.4.6 Hop
count exceeded - possible mail loop (in reply to end of DATA command)
Final-Recipient: rfc822; chris.mason@fusionio.com
Action: failed
Status: 5.0.0
Diagnostic-Code: X-Spam-&-Virus-Firewall; host
2019 Oct 14
0
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
On 10/14/19 3:12 AM, Li Feng wrote:
> Hi Cole & Michal,
>
> I'm sorry for my late response, I just end my journey today.
> Thank your response, your suggestion is very helpful to me.
>
> I have added Michal in this mail, Michal helps me review my initial patchset.
> (https://www.spinics.net/linux/fedora/libvir/msg191339.html)
>
Whoops I missed that posting, I
2019 Oct 09
0
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Sorry for the late reply, and thanks Jano for pointing out elsewhere
that this didn't receive a response.
On 8/12/19 5:56 AM, Li Feng wrote:
> Hi Guys,
>
> And I want to add the vhost-user-scsi-pci/vhost-user-blk-pci support
> for libvirt.
>
> The usage in qemu like this:
>
> Vhost-SCSI
> -chardev socket,id=char0,path=/var/tmp/vhost.0
> -device
2019 Jul 19
3
Samba async performance - bottleneck or bug?
Hi David,
Thanks for your reply.
> Hmm, so this "async" (sync=disabled?) ZFS tunable means that it
> completely ignores O_SYNC and O_DIRECT and runs the entire workload in
> RAM? I know nothing about ZFS, but that sounds like a mighty dangerous
> setting for production deployments.
Yes, you are correct - sync writes will flush to RAM, just like async, will stay in RAM for
2019 Oct 15
1
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Cole Robinson <crobinso@redhat.com> 于2019年10月15日周二 上午1:48写道:
>
> On 10/14/19 3:12 AM, Li Feng wrote:
> > Hi Cole & Michal,
> >
> > I'm sorry for my late response, I just end my journey today.
> > Thank your response, your suggestion is very helpful to me.
> >
> > I have added Michal in this mail, Michal helps me review my initial patchset.
2012 Mar 25
3
attempt to access beyond end of device and livelock
Hi Dongyang, Yan,
When testing BTRFS with RAID 0 metadata on linux-3.3, we see discard
ranges exceeding the end of the block device [1], potentially causing
dataloss; when this occurs, filesystem writeback becomes catatonic due
to continual resubmission.
The reproducer is quite simple [2]. Hope this proves useful...
Thanks,
Daniel
--- [1]
attempt to access beyond end of device
ram0: rw=129,
2019 Aug 12
2
Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Hi Guys,
And I want to add the vhost-user-scsi-pci/vhost-user-blk-pci support
for libvirt.
The usage in qemu like this:
Vhost-SCSI
-chardev socket,id=char0,path=/var/tmp/vhost.0
-device vhost-user-scsi-pci,id=scsi0,chardev=char0
Vhost-BLK
-chardev socket,id=char1,path=/var/tmp/vhost.1
-device vhost-user-blk-pci,id=blk0,chardev=char1
What type should I add for libvirt.
Type1:
<hostdev
2019 Jul 25
0
Samba async performance - bottleneck or bug?
Hi,
On Fri, 19 Jul 2019 23:26:55 +0000, douxevip wrote:
> So to summarize, this is the situation:
>
> 1) I run a fio benchmark requesting, small, random, async writes. Command is "fio --direct=1 --sync=0 --rw=randwrite --bs=4K --numjobs=1 --iodepth=1 --runtime=60 --group_reporting --name=sambatest --size=32k --time_based". I run this command on both the host, as on the Samba
2019 Aug 06
1
Samba async performance - bottleneck or bug?
Hi David,
> You're still using direct I/O with fio, which will likely disallow
> client side caching with oplocks/leases.
Is there a way to bypass this with settings in smb.conf at all and transform all writes to async?
> I'd recommend checking that your (cifs.ko?) client is using a relatively
> modern SMB2+ dialect and that leases are enabled on both sides.
Yes, I
2019 Jul 19
0
Samba async performance - bottleneck or bug?
Hi,
On Thu, 18 Jul 2019 19:04:47 +0000, douxevip via samba wrote:
> Hi,
>
> I have a ZFS dataset that has sync writes disabled (setting sync=disabled) which means that it will only do async writes, and sync requests get converted to async writes. The ZFS dataset is hosted on a single Samsung 840 Pro 512GB SATA SSD.
> I have this same dataset served as a Samba share, using Proxmox VE
2019 Oct 14
2
Re: [libvirt] Add support for vhost-user-scsi-pci/vhost-user-blk-pci
Hi Cole & Michal,
I'm sorry for my late response, I just end my journey today.
Thank your response, your suggestion is very helpful to me.
I have added Michal in this mail, Michal helps me review my initial patchset.
(https://www.spinics.net/linux/fedora/libvir/msg191339.html)
All concern about this feature is the XML design.
My original XML design exposes more details of Qemu.
2019 Jul 18
2
Samba async performance - bottleneck or bug?
Hi,
I have a ZFS dataset that has sync writes disabled (setting sync=disabled) which means that it will only do async writes, and sync requests get converted to async writes. The ZFS dataset is hosted on a single Samsung 840 Pro 512GB SATA SSD.
I have this same dataset served as a Samba share, using Proxmox VE 6. Samba version 4.9.5-Debian (Buster), protocol SMB3_11. Kernel version 5.0.15.
To
2018 Mar 20
0
Gluster very poor performance when copying small files (1x (2+1) = 3, SSD)
Excellent description, thank you.
With performance.write-behind-trickling-writes ON (default):
## 4k randwrite
# fio --randrepeat=1 --ioengine=libaio --gtod_reduce=1 --name=test --filename=test --bs=4k --iodepth=32 --size=256MB --readwrite=randwrite
test: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
Jobs: 1
2014 Jun 26
0
[PATCH v2 0/2] block: virtio-blk: support multi vq per virtio-blk
On 2014-06-25 20:08, Ming Lei wrote:
> Hi,
>
> These patches try to support multi virtual queues(multi-vq) in one
> virtio-blk device, and maps each virtual queue(vq) to blk-mq's
> hardware queue.
>
> With this approach, both scalability and performance on virtio-blk
> device can get improved.
>
> For verifying the improvement, I implements virtio-blk multi-vq
2018 Oct 30
0
Fw: Re: [SPDK] VM boot failed sometimes if using vhost-user-blk with spdk
Forwarded to centos mailing list
-----Original Messages-----
From: wuzhouhui <wuzhouhui14 at mails.ucas.ac.cn>
Sent Time: 2018-10-30 14:06:00 (Tuesday)
To: "storage performance development kit" <spdk at lists.01.org>
Cc: centos at centos.org, qemu-discuss at nongnu.org
Subject: Re: [SPDK] VM boot failed sometimes if using vhost-user-blk with spdk
I enable debug of
2014 Jun 26
0
[PATCH v3 0/2] block: virtio-blk: support multi vq per virtio-blk
On Thu, Jun 26, 2014 at 5:41 PM, Ming Lei <ming.lei at canonical.com> wrote:
> Hi,
>
> These patches try to support multi virtual queues(multi-vq) in one
> virtio-blk device, and maps each virtual queue(vq) to blk-mq's
> hardware queue.
>
> With this approach, both scalability and performance on virtio-blk
> device can get improved.
>
> For verifying the
2014 Jul 01
0
[PATCH v3 0/2] block: virtio-blk: support multi vq per virtio-blk
On 2014-06-30 19:36, Ming Lei wrote:
> Hi Jens and Rusty,
>
> On Thu, Jun 26, 2014 at 8:04 PM, Ming Lei <ming.lei at canonical.com> wrote:
>> On Thu, Jun 26, 2014 at 5:41 PM, Ming Lei <ming.lei at canonical.com> wrote:
>>> Hi,
>>>
>>> These patches try to support multi virtual queues(multi-vq) in one
>>> virtio-blk device, and maps
2014 Jul 01
2
[PATCH v3 0/2] block: virtio-blk: support multi vq per virtio-blk
Hi Jens and Rusty,
On Thu, Jun 26, 2014 at 8:04 PM, Ming Lei <ming.lei at canonical.com> wrote:
> On Thu, Jun 26, 2014 at 5:41 PM, Ming Lei <ming.lei at canonical.com> wrote:
>> Hi,
>>
>> These patches try to support multi virtual queues(multi-vq) in one
>> virtio-blk device, and maps each virtual queue(vq) to blk-mq's
>> hardware queue.
>>