João Baúto
2019-Apr-29 10:25 UTC
[Gluster-users] parallel-readdir prevents directories and files listing - Bug 1670382
Hi, I have an 8 brick distributed volume where Windows and Linux clients mount the volume via samba and headless compute servers using gluster native fuse. With parallel-readdir on, if a Windows client creates a new folder, the folder is indeed created but invisible to the Windows client. Accessing the same samba share in a Linux client, the folder is again visible and with normal behavior. The same folder is also visible when mounting via gluster native fuse. The Windows client can list existing directories and rename them while, for files, everything seems to be working fine. Gluster servers: CentOS 7.5 with Gluster 5.3 and Samba 4.8.3-4.el7.0.1 from @fasttrack Clients tested: Windows 10, Ubuntu 18.10, CentOS 7.5 https://bugzilla.redhat.com/show_bug.cgi?id=1670382 Volume Name: tank Type: Distribute Volume ID: 9582685f-07fa-41fd-b9fc-ebab3a6989cf Status: Started Snapshot Count: 0 Number of Bricks: 8 Transport-type: tcp Bricks: Brick1: swp-gluster-01:/tank/volume1/brick Brick2: swp-gluster-02:/tank/volume1/brick Brick3: swp-gluster-03:/tank/volume1/brick Brick4: swp-gluster-04:/tank/volume1/brick Brick5: swp-gluster-01:/tank/volume2/brick Brick6: swp-gluster-02:/tank/volume2/brick Brick7: swp-gluster-03:/tank/volume2/brick Brick8: swp-gluster-04:/tank/volume2/brick Options Reconfigured: performance.parallel-readdir: on performance.readdir-ahead: on performance.cache-invalidation: on performance.md-cache-timeout: 600 storage.batch-fsync-delay-usec: 0 performance.write-behind-window-size: 32MB performance.stat-prefetch: on performance.read-ahead: on performance.read-ahead-page-count: 16 performance.rda-request-size: 131072 performance.quick-read: on performance.open-behind: on performance.nl-cache-timeout: 600 performance.nl-cache: on performance.io-thread-count: 64 performance.io-cache: off performance.flush-behind: on performance.client-io-threads: off performance.write-behind: off performance.cache-samba-metadata: on network.inode-lru-limit: 0 features.cache-invalidation-timeout: 600 features.cache-invalidation: on cluster.readdir-optimize: on cluster.lookup-optimize: on client.event-threads: 4 server.event-threads: 16 features.quota-deem-statfs: on nfs.disable: on features.quota: on features.inode-quota: on cluster.enable-shared-storage: disable Cheers, Jo?o Ba?to -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190429/1f8c5db1/attachment.html>
Amar Tumballi Suryanarayan
2019-May-01 12:29 UTC
[Gluster-users] parallel-readdir prevents directories and files listing - Bug 1670382
On Mon, Apr 29, 2019 at 3:56 PM Jo?o Ba?to < joao.bauto at neuro.fchampalimaud.org> wrote:> Hi, > > I have an 8 brick distributed volume where Windows and Linux clients mount > the volume via samba and headless compute servers using gluster native > fuse. With parallel-readdir on, if a Windows client creates a new folder, > the folder is indeed created but invisible to the Windows client. Accessing > the same samba share in a Linux client, the folder is again visible and > with normal behavior. The same folder is also visible when mounting via > gluster native fuse. > > The Windows client can list existing directories and rename them while, > for files, everything seems to be working fine. > > Gluster servers: CentOS 7.5 with Gluster 5.3 and Samba 4.8.3-4.el7.0.1 > from @fasttrack > Clients tested: Windows 10, Ubuntu 18.10, CentOS 7.5 > > https://bugzilla.redhat.com/show_bug.cgi?id=1670382 >Thanks for the bug report. Will look into this, and get back. Last I knew, we recommended to avoid fuse and samba shares on same volume (Mainly as we couldn't spend a lot of effort on testing the configuration). Anyways, we would treat the behavior as bug for sure. One possible path looking at below volume info is to disable 'stat-prefetch' option and see if it helps. Next option I would try is to disable readdir-ahead. Regards, Amar> <https://bugzilla.redhat.com/show_bug.cgi?id=1670382> > > Volume Name: tank > Type: Distribute > Volume ID: 9582685f-07fa-41fd-b9fc-ebab3a6989cf > Status: Started > Snapshot Count: 0 > Number of Bricks: 8 > Transport-type: tcp > Bricks: > Brick1: swp-gluster-01:/tank/volume1/brick > Brick2: swp-gluster-02:/tank/volume1/brick > Brick3: swp-gluster-03:/tank/volume1/brick > Brick4: swp-gluster-04:/tank/volume1/brick > Brick5: swp-gluster-01:/tank/volume2/brick > Brick6: swp-gluster-02:/tank/volume2/brick > Brick7: swp-gluster-03:/tank/volume2/brick > Brick8: swp-gluster-04:/tank/volume2/brick > Options Reconfigured: > performance.parallel-readdir: on > performance.readdir-ahead: on > performance.cache-invalidation: on > performance.md-cache-timeout: 600 > storage.batch-fsync-delay-usec: 0 > performance.write-behind-window-size: 32MB > performance.stat-prefetch: on > performance.read-ahead: on > performance.read-ahead-page-count: 16 > performance.rda-request-size: 131072 > performance.quick-read: on > performance.open-behind: on > performance.nl-cache-timeout: 600 > performance.nl-cache: on > performance.io-thread-count: 64 > performance.io-cache: off > performance.flush-behind: on > performance.client-io-threads: off > performance.write-behind: off > performance.cache-samba-metadata: on > network.inode-lru-limit: 0 > features.cache-invalidation-timeout: 600 > features.cache-invalidation: on > cluster.readdir-optimize: on > cluster.lookup-optimize: on > client.event-threads: 4 > server.event-threads: 16 > features.quota-deem-statfs: on > nfs.disable: on > features.quota: on > features.inode-quota: on > cluster.enable-shared-storage: disable > > Cheers, > > Jo?o Ba?to > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users-- Amar Tumballi (amarts) -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190501/9c91d3c6/attachment.html>