Hi Kelvin, We've filed a bug http://bugs.gluster.com/cgi-bin/bugzilla3/show_bug.cgi?id=921 for the crash you reported. The workaround you mentioned does not appear to be related to the crash reported. Can you send us the backtrace of the core file: $ gdb -c <core-file> <glusterfs-binary path> (gdb) thread apply all bt full Also attach the complete log files of the servers and the clients and the volfiles used for the configuration. Pavan ----- Original Message ----- From: "Kelvin Westlake" <Kelvin at netbasic.co.uk> To: gluster-users at gluster.org Sent: Monday, April 12, 2010 1:15:51 PM Subject: Re: [Gluster-users] Gluster crashing I've found a work around to this, on Centos 5.4, I completely removed all the fuse rpms (find them with rpm -qa | grep fuse), then downloaded the latest FUSE version and built it myself. -----Original Message----- From: Kelvin Westlake Sent: 09 April 2010 17:13 To: 'gluster-users at gluster.org' Subject: Gluster crashing Hi Guys I have 2 physical servers, which I'm trying to setup in a HA pair.. I'm sharing 2 volumes (home-gfs and shared-gfs) across these servers in RAID1 (replicated), then I'm mounting clients on each server as home/ and shared/, if I lose one of the Servers and then remount it, the client on the other server seems to crash out. The following is the long entries leading to the crash - [2010-04-09 16:50:44] E [socket.c:762:socket_connect_finish] 192.168.100.31-2: connection to 192.168.100.31:6996 failed (Connection refused) [2010-04-09 16:50:44] E [socket.c:762:socket_connect_finish] 192.168.100.31-2: connection to 192.168.100.31:6996 failed (Connection refused) [2010-04-09 16:51:31] N [client-protocol.c:6246:client_setvolume_cbk] 192.168.100.31-2: Connected to 192.168.100.31:6996, attached to remote volume 'brick2'. [2010-04-09 16:51:31] E [afr-self-heal-common.c:1237:sh_missing_entries_create] mirror-1: unknown file type: 01 pending frames: frame : type(1) op(READDIRP) patchset: git://git.sv.gnu.org/gluster.git signal received: 11 time of crash: 2010-04-09 16:51:31 configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.0.0git [0x371420] /usr/local/lib/glusterfs/3.0.0git/xlator/protocol/client.so(client_readd irp+0x1b4)[0x28d224] /usr/local/lib/glusterfs/3.0.0git/xlator/cluster/replicate.so(afr_do_rea ddir+0x4e2)[0x98a722] /usr/local/lib/glusterfs/3.0.0git/xlator/cluster/replicate.so(afr_readdi rp+0x48)[0x98a988] /usr/local/lib/glusterfs/3.0.0git/xlator/mount/fuse.so[0x113e98] /usr/local/lib/glusterfs/3.0.0git/xlator/mount/fuse.so[0x11b25a] /lib/libpthread.so.0[0xbeb73b] /lib/libc.so.6(clone+0x5e)[0xb69cfe] --------- This email with any attachments is for the exclusive and confidential use of the addressee(s) and may contain legally privileged information. Any other distribution, use or reproduction without the senders prior consent is unauthorised and strictly prohibited. If you receive this message in error please notify the sender by email and delete the message from your computer. Netbasic Limited registered office and business address is 9 Funtley Court, Funtley Hill, Fareham, Hampshire PO16 7UY. Company No. 04906681. Netbasic Limited is authorised and regulated by the Financial Services Authority in respect of regulated activities. Please note that many of our activities do not require FSA regulation. _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
Burnash, James
2010-May-12 14:29 UTC
[Gluster-users] Best Practices for Gluster Replication
Greetings List, I've searched through the Gluster wiki and a lot of threads to try to answer this question, but so far no real luck. Simply put - is it better to have replication handled by the clients, or by the bricks themselves? Volgen for a raid 1 solution creates a config file that does the mirroring on the client side - which I would take as an implicit endorsement from the Gluster team (great team, BTW). However, it seems to me that if the bricks replicated between themselves on our 10Gb storage network, it could save a lot of bandwidth for the clients and conceivably save them CPU cycles an I/O as well. Client machines have 1Gb connections to the storage network, and are running CentOS 5.2. Server machines have 10Gb connections to the storage network, and are running CentOS 5.4. Glusterfs.vol: ## file auto generated by /usr/bin/glusterfs-volgen (mount.vol) # Cmd line: # $ /usr/bin/glusterfs-volgen --name testfs --raid 1 jc1letgfs13-pfs1:/export/read-write jc1letgfs14-pfs1:/export/read-write jc1letgfs15-pfs1:/export/read-write jc1letgfs16-pfs1:/export/read-write jc1letgfs17-pfs1:/export/read-write jc1letgfs18-pfs1:/export/read-write # RAID 1 # TRANSPORT-TYPE tcp volume jc1letgfs17-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs17-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume jc1letgfs18-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs18-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume jc1letgfs13-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs13-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume jc1letgfs15-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs15-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume jc1letgfs16-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs16-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume jc1letgfs14-pfs1-1 type protocol/client option transport-type tcp option remote-host jc1letgfs14-pfs1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume brick1 end-volume volume mirror-0 type cluster/replicate subvolumes jc1letgfs13-pfs1-1 jc1letgfs14-pfs1-1 end-volume volume mirror-1 type cluster/replicate subvolumes jc1letgfs15-pfs1-1 jc1letgfs16-pfs1-1 end-volume volume mirror-2 type cluster/replicate subvolumes jc1letgfs17-pfs1-1 jc1letgfs18-pfs1-1 end-volume volume distribute type cluster/distribute subvolumes mirror-0 mirror-1 mirror-2 end-volume volume readahead type performance/read-ahead option page-count 4 subvolumes distribute end-volume volume iocache type performance/io-cache option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed 's/[^0-9]//g') / 5120 ))`MB option cache-timeout 1 subvolumes readahead end-volume volume quickread type performance/quick-read option cache-timeout 1 option max-file-size 64kB subvolumes iocache end-volume volume writebehind type performance/write-behind option cache-size 4MB subvolumes quickread end-volume volume statprefetch type performance/stat-prefetch subvolumes writebehind end-volume Glusterfsd.vol: ## file auto generated by /usr/bin/glusterfs-volgen (export.vol) # Cmd line: # $ /usr/bin/glusterfs-volgen --name testfs jc1letgfs13-pfs1:/export/read-write jc1letgfs14-pfs1:/export/read-write jc1letgfs15-pfs1:/export/read-write volume posix1 type storage/posix option directory /export/read-write end-volume volume locks1 type features/locks subvolumes posix1 end-volume volume brick1 type performance/io-threads option thread-count 8 subvolumes locks1 end-volume volume server-tcp type protocol/server option transport-type tcp option auth.addr.brick1.allow * option transport.socket.listen-port 6996 option transport.socket.nodelay on subvolumes brick1 end-volume James Burnash DISCLAIMER: This e-mail, and any attachments thereto, is intended only for use by the addressee(s) named herein and may contain legally privileged and/or confidential information. If you are not the intended recipient of this e-mail, you are hereby notified that any dissemination, distribution or copying of this e-mail, and any attachments thereto, is strictly prohibited. If you have received this in error, please immediately notify me and permanently delete the original and any copy of any e-mail and any printout thereof. E-mail transmission cannot be guaranteed to be secure or error-free. The sender therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of e-mail transmission. NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its discretion, monitor and review the content of all e-mail communications. http://www.knight.com