Just checked: when the write is >= 340 bytes, everything works as
supposed. If the write is smaller, the error takes place. And when it
does, nothing is logged on the server. The client, however, logs the
following:
[2019-01-23 23:28:54.554664] W [MSGID: 103046]
[rdma.c:3502:gf_rdma_decode_header] 0-rpc-transport/rdma: received a msg
of type RDMA_ERROR
[2019-01-23 23:28:54.554728] W [MSGID: 103046]
[rdma.c:3939:gf_rdma_process_recv] 0-rpc-transport/rdma: peer
(172.24.1.6:49152), couldn't encode or decode the msg properly or write
chunks were not provided for replies that were bigger than
RDMA_INLINE_THRESHOLD (2048)
[2019-01-23 23:28:54.554765] W [MSGID: 114031]
[client-rpc-fops_v2.c:680:client4_0_writev_cbk] 0-gfs-client-5: remote
operation failed [Transport endpoint is not connected]
[2019-01-23 23:28:54.554850] W [fuse-bridge.c:1436:fuse_err_cbk]
0-glusterfs-fuse: 1723199: FLUSH() ERR => -1 (Transport endpoint is not
connected)
Lindolfo Meira, MSc
Diretor Geral, Centro Nacional de Supercomputa??o
Universidade Federal do Rio Grande do Sul
+55 (51) 3308-3139
On Wed, 23 Jan 2019, Lindolfo Meira wrote:
> Hi Jim. Thanks for taking the time.
>
> Sorry I didn't express myself properly. It's not a simple matter of
> permissions. Users can write to the volume alright. It's when vim and
nano
> are used, or when small file writes are performed (by cat or echo), that
> it doesn't work. The file is updated with the write in the server, but
it
> shows up as empty in the client.
>
> I guess it has something to do with the size of the write, because I ran a
> test writing to a file one byte at a time, and it never showed up as
> having any content in the client (although in the server it kept growing
> accordingly).
>
> I should point out that I'm using a sharded volume. But when I was
testing
> a striped volume, it also happened. Output of "gluster volume
info"
> follows bellow:
>
> Volume Name: gfs
> Type: Distribute
> Volume ID: b5ef065f-1ba2-481f-8108-e8f6d2d3f036
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6
> Transport-type: rdma
> Bricks:
> Brick1: pfs01-ib:/mnt/data
> Brick2: pfs02-ib:/mnt/data
> Brick3: pfs03-ib:/mnt/data
> Brick4: pfs04-ib:/mnt/data
> Brick5: pfs05-ib:/mnt/data
> Brick6: pfs06-ib:/mnt/data
> Options Reconfigured:
> nfs.disable: on
> features.shard: on
>
>
>
> Lindolfo Meira, MSc
> Diretor Geral, Centro Nacional de Supercomputa??o
> Universidade Federal do Rio Grande do Sul
> +55 (51) 3308-3139
>
> On Wed, 23 Jan 2019, Jim Kinney wrote:
>
> > Check permissions on the mount. I have multiple dozens of systems
> > mounting 18 "exports" using fuse and it works for multiple
user
> > read/write based on user access permissions to the mount point space.
> > /home is mounted for 150+ users plus another dozen+ lab storage
spaces.
> > I do manage user access with freeIPA across all systems to keep things
> > consistent.
> > On Wed, 2019-01-23 at 19:31 -0200, Lindolfo Meira wrote:
> > > Am I missing something here? A mere write operation, using vim or
> > > nano, cannot be performed on a gluster volume mounted over fuse!
What
> > > gives?
> > > Lindolfo Meira, MScDiretor Geral, Centro Nacional de
> > > Supercomputa??oUniversidade Federal do Rio Grande do Sul+55 (51)
> > >
3308-3139_______________________________________________Gluster-users
> > > mailing listGluster-users at gluster.org
> > > https://lists.gluster.org/mailman/listinfo/gluster-users
> > --
> > James P. Kinney III
> >
> > Every time you stop a school, you will have to build a jail. What you
> > gain at one end you lose at the other. It's like feeding a dog on
his
> > own tail. It won't fatten the dog.
> > - Speech 11/23/1900 Mark Twain
> >
> > http://heretothereideas.blogspot.com/
> >
> >