Kaleb S. KEITHLEY
2018-Aug-10 13:39 UTC
[Gluster-users] ganesha.nfsd process dies when copying files
On 08/10/2018 09:23 AM, Karli Sj?berg wrote:> On Fri, 2018-08-10 at 21:23 +0800, Pui Edylie wrote: >> Hi Karli, >> >> Storhaug works with glusterfs 4.1.2 and latest nfs-ganesha. >> >> I just installed them last weekend ... they are working very well :) > > Okay, awesome! > > Is there any documentation on how to do that? >https://github.com/gluster/storhaug/wiki -- Kaleb -------------- next part -------------- A non-text attachment was scrubbed... Name: signature.asc Type: application/pgp-signature Size: 488 bytes Desc: OpenPGP digital signature URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180810/4366269a/attachment.sig>
Karli Sjöberg
2018-Aug-14 13:10 UTC
[Gluster-users] ganesha.nfsd process dies when copying files
On Fri, 2018-08-10 at 09:39 -0400, Kaleb S. KEITHLEY wrote:> On 08/10/2018 09:23 AM, Karli Sj?berg wrote: > > On Fri, 2018-08-10 at 21:23 +0800, Pui Edylie wrote: > > > Hi Karli, > > > > > > Storhaug works with glusterfs 4.1.2 and latest nfs-ganesha. > > > > > > I just installed them last weekend ... they are working very well > > > :) > > > > Okay, awesome! > > > > Is there any documentation on how to do that? > > > > https://github.com/gluster/storhaug/wiki >Thanks Kaleb and Edy! I have now redone the cluster using the latest and greatest following the above guide and repeated the same test I was doing before (the rsync while loop) with success. I let (forgot) it run for about a day and it was still chugging along nicely when I aborted it, so success there! On to the next test; the catastrophic failure test- where one of the servers dies, I'm having a more difficult time with. 1) I start with mounting the share over NFS 4.1 and then proceed with writing a 8 GiB large random data file with 'dd', while "hard-cutting" the power to the server I'm writing to, the transfer just stops indefinitely, until the server comes back again. Is that supposed to happen? Like this: # dd if=/dev/urandom of=/var/tmp/test.bin bs=1M count=8192 # mount -o vers=4.1 hv03v.localdomain:/data /mnt/ # dd if=/var/tmp/test.bin of=/mnt/test.bin bs=1M status=progress 2434793472 bytes (2,4 GB, 2,3 GiB) copied, 42 s, 57,9 MB/s (here I cut the power and let it be for almost two hours before turning it on again) dd: error writing '/mnt/test.bin': Remote I/O error 2325+0 records in 2324+0 records out 2436890624 bytes (2,4 GB, 2,3 GiB) copied, 6944,84 s, 351 kB/s # umount /mnt Here the unmount command hung and I had to hard reset the client. 2) Another question I have is why some files "change" as you copy them out to the Gluster storage? Is that the way it should be? This time, I deleted eveything in the destination directory to start over: # mount -o vers=4.1 hv03v.localdomain:/data /mnt/ # rm -f /mnt/test.bin # dd if=/var/tmp/test.bin of=/mnt/test.bin bs=1M status=progress 8557428736 bytes (8,6 GB, 8,0 GiB) copied, 122 s, 70,1 MB/s 8192+0 records in 8192+0 records out 8589934592 bytes (8,6 GB, 8,0 GiB) copied, 123,039 s, 69,8 MB/s # md5sum /var/tmp/test.bin 073867b68fa8eaa382ffe05adb90b583 /var/tmp/test.bin # md5sum /mnt/test.bin 634187d367f856f3f5fb31846f796397 /mnt/test.bin # umount /mnt Thanks in advance! /K