Hi,
Is it possible for you to repeat tests after patching glusterfs with attached
patch? Also please send us the entire server log files after you run into the
issue.
regards,
Raghavendra.
----- Original Message -----> From: "Beat Rubischon" <beat at 0x1b.ch>
> To: gluster-users at gluster.org
> Sent: Monday, February 21, 2011 10:51:23 AM
> Subject: [Gluster-users] Issue in RDMA transport
> Hello!
>
> I found some memory corruption in the RDMA transport layer.
>
> Setup is CentOS 5.5, Mellanox OFED 1.5.2 / OpenFabrics OFED 1.5.2,
> ConnectX-2 cards, GlusterFS 3.1.2 / Git Master Branch.
>
> Application is ANSYS CFX wit transient cases, running with strange
> corecounts like 6 or 12.
>
> Symptoms are failure during the write out of the case. Errors are
> recorded in the brick's and client's logs:
>
> node24:/var/log/glusterfs/home.log
> [2011-02-04 15:41:19.688110] W [fuse-bridge.c:1761:fuse_writev_cbk]
> glusterfs-fuse: 29810266: WRITE => -1 (Bad address)
>
> server2:/var/log/glusterfs/bricks/brick07.log
> [2011-02-04 15:41:19.687733] E [posix.c:2504:posix_writev] home-posix:
> write failed: offset 538534184, Bad address
>
> I was able to reproduce the error using a single brick and a single
> client. Running server and client on the same system didn't pop up the
> error, the data must pass a wire to trigger the bug. Switching to TCP
> over IPoIB was a successful workaround.
>
> It looks like a pointer in the iovec structure used by the writev is
> screwed up during the transport over RDMA. I can imagine that the
> debugging would be rather hard, hopefully you'll be able to find the
> root cause. Feel free to ask for additional logs or traces, I'll try
> to
> provide them.
>
> Beat
>
> --
> \|/ Beat Rubischon <beat at 0x1b.ch>
> ( 0-0 ) http://www.0x1b.ch/~beat/
> oOO--(_)--OOo---------------------------------------------------
> Meine Erlebnisse, Gedanken und Traeume: http://www.0x1b.ch/blog/
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users