similar to: [PATCH v5 0/4] virtio-fs: shared file system for virtual machines

Displaying 16 results from an estimated 16 matches similar to: "[PATCH v5 0/4] virtio-fs: shared file system for virtual machines"

2019 Sep 12
0
[PATCH v5 0/4] virtio-fs: shared file system for virtual machines
On Thu, Sep 12, 2019 at 10:14:11AM +0200, Miklos Szeredi wrote: > On Wed, Sep 11, 2019 at 5:54 PM Stefan Hajnoczi <stefanha at redhat.com> wrote: > > > > On Tue, Sep 10, 2019 at 05:12:02PM +0200, Miklos Szeredi wrote: > > > I've folded the series from Vivek and fixed a couple of TODO comments > > > myself. AFAICS two issues remain that need to be
2020 Sep 25
0
[RFC PATCH 01/19] tools/virtiofsd: add support for --socket-group
If you like running QEMU as a normal user (very common for TCG runs) but you have to run virtiofsd as a root user you run into connection problems. Adding support for an optional --socket-group allows the users to keep using the command line. Signed-off-by: Alex Benn?e <alex.bennee at linaro.org> Reviewed-by: Stefan Hajnoczi <stefanha at redhat.com> --- v1 - tweak documentation and
2023 Apr 04
1
Virtiofsd
Hi In QEMU 8 virtiofsd has been removed in favor of the rust version. Which includes options that are not longer supported, Do you have a view on what should be used going forwards to support virtiofsd in libvirt with qemu 8? The options are showing as depreciated, -o <compat-options>... Options in a format compatible with the legacy implementation [deprecated] Rust
2023 Apr 04
1
Virtiofsd
On Tue, Apr 04, 2023 at 17:12:15 +0100, Simon Fairweather wrote: > Hi > > In QEMU 8 virtiofsd has been removed in favor of the rust version. Which > includes options that are not longer supported, > > > Do you have a view on what should be used going forwards to support > virtiofsd in libvirt with qemu 8? > > The options are showing as depreciated, > -o
2008 Dec 09
1
File uploaded to webDAV server on GlusterFS AFR - ends up without xattr!
Hello list. I'm testing GlusterFS AFR mode as a solution for implementing a highly available webDAV file storage for our production environment. Whlie doing performance tests I've notticed a strange behavior: the files which are uploaded via a webDAV server, end up without extended attributes, which removes the ability to self-heal. The set up is a simple testing environment with 2
2023 Feb 08
2
virtio-fs: adding support for multi-queue
On 08/02/2023 11:43, Stefan Hajnoczi wrote: > On Wed, Feb 08, 2023 at 09:33:33AM +0100, Peter-Jan Gootzen wrote: >> >> >> On 07/02/2023 22:57, Vivek Goyal wrote: >>> On Tue, Feb 07, 2023 at 04:32:02PM -0500, Stefan Hajnoczi wrote: >>>> On Tue, Feb 07, 2023 at 02:53:58PM -0500, Vivek Goyal wrote: >>>>> On Tue, Feb 07, 2023 at 02:45:39PM
2016 Jan 06
0
[klibc:master] MIPS: Update archfcntl.h
Commit-ID: 3fefc6a404a970a911417d0345618a7e9abfef70 Gitweb: http://git.kernel.org/?p=libs/klibc/klibc.git;a=commit;h=3fefc6a404a970a911417d0345618a7e9abfef70 Author: Ben Hutchings <ben at decadent.org.uk> AuthorDate: Wed, 6 Jan 2016 00:43:25 +0000 Committer: H. Peter Anvin <hpa at linux.intel.com> CommitDate: Tue, 5 Jan 2016 17:45:36 -0800 [klibc] MIPS: Update archfcntl.h
2019 Jan 18
0
[klibc:master] fcntl: Fix file locking numbers for 64-bit architectures
Commit-ID: 3cb3ceea23ecbf2c804bf732e8234552d097b94f Gitweb: http://git.kernel.org/?p=libs/klibc/klibc.git;a=commit;h=3cb3ceea23ecbf2c804bf732e8234552d097b94f Author: Ben Hutchings <ben at decadent.org.uk> AuthorDate: Fri, 18 Jan 2019 18:39:21 +0000 Committer: Ben Hutchings <ben at decadent.org.uk> CommitDate: Fri, 18 Jan 2019 18:46:45 +0000 [klibc] fcntl: Fix file locking
2020 Sep 25
20
[RFC PATCH 00/19] vhost-user-rpmb (Replay Protected Memory Block)
Hi, This is an initial implementation of a vhost-user backend for the VirtIO RPMB device. The device is currently in the draft of the next VirtIO specification and describes block device which uses combination of a key, nonce, hashing and a persistent write counter to prevent replay attacks (hence Replay Protected Memory Block). It is implemented as a vhost-user device because we want to
2013 Jun 17
0
gluster client timeouts / found conflict
Hi list Recently I've experienced more and more input/output errors from my most write heavy gluster filesystem. The logfile on the gluster servers show nothing, but the client(s) that get the input/output errors (and timeouts) will as far as I can tell get errors such as : [2013-06-14 15:55:56] W [fuse-bridge.c:493:fuse_entry_cbk] glusterfs-fuse: LOOKUP(/369/60702093) inode (ptr=0x1efd440,
2010 Jan 07
2
Random directory/files gets unavailable after sometime
Hello, I am using glusterfs v3.0.0 and having some problems with random directory/files. They work fine for some time ( hours ) and them suddenly gets unavailable: # ls -lh ls: cannot access MyDir: No such file or directory total 107M d????????? ? ? ? ? ? MyDir ( long dir list, intentionally hidden ) At the logs i get a lot of messages like those ones: [2010-01-07
2008 Aug 01
1
file descriptor in bad state
I've just setup a simple gluster storage system on Centos 5.2 x64 w/ gluster 1.3.10 I have three storage bricks and one client Everytime i run iozone across this setup, i seem to get a bad file descriptor around the 4k mark. Any thoughts why? I'm sure more info is wanted, i'm just not sure what else to include at this point. thanks [root at green gluster]# cat
2008 Dec 10
3
AFR healing problem after returning one node.
I've got configuration which in simple includes combination of afrs and unify - servers exports n[1-3]-brick[12] and n[1-3]-ns and client got cluster configuration: volume afr-ns type cluster/afr subvolumes n1-ns n2-ns n3-ns option data-self-heal on option metadata-self-heal on option entry-self-heal on end-volume volume afr1 type cluster/afr subvolumes n1-brick2
2012 Dec 07
2
Performance problems while running doveadm purge
I have a rather large and active mdbox (28 GB, 3M mess, 1200 deliveries/day). I usually have no problems working with those mails, and there is some batch processing going on (via doveadm). Every few weeks I try my luck running doveadm purge, and this a) crunches about 5GB (to be expected), b) takes rather long (ok), c) leads to long stretches of blocked mdbox, which is the problem. I always
2008 Oct 07
2
1.1.4 and trouble over NFS
Hello, I have some trouble with the current setup (it's a testing environment): 2 server with Dovecot 1.1.4 from source (OS Debian testing 2.6.26) (name: "exim" and "exim2") 1 NFS server (OS Debian testing 2.6.26) I use NFS v.4, indexes shared over NFS. The relavant part of Dovecot configuration: dotlock_use_excl = yes mail_nfs_storage = yes mail_nfs_index = yes
2016 Jan 06
5
[PATCH klibc 0/5] klibc architecture fixes
Here's an assortment of build and run-time fixes for various architectures that we've applied in Debian. Ben. Aurelien Jarno (1): ppc64: fix struct stat Ben Hutchings (2): MIPS: Update archfcntl.h syscalls: Override detection of direct socket syscalls on i386, m68k, s390 Helge Deller (1): Add pread and pwrite 32bit syscall wrappers for parisc Mauricio Faria de Oliveira