Hi Kelvin, We've filed a bug http://bugs.gluster.com/cgi-bin/bugzilla3/show_bug.cgi?id=921 for the crash you reported. The workaround you mentioned does not appear to be related to the crash reported. Can you send us the backtrace of the core file: $ gdb -c <core-file> <glusterfs-binary path> (gdb) thread apply all bt full Also attach the complete log files of the servers and the clients and the volfiles used for the configuration. Pavan ----- Original Message ----- From: "Kelvin Westlake" <Kelvin at netbasic.co.uk> To: gluster-users at gluster.org Sent: Monday, April 12, 2010 1:15:51 PM Subject: Re: [Gluster-users] Gluster crashing I've found a work around to this, on Centos 5.4, I completely removed all the fuse rpms (find them with rpm -qa | grep fuse), then downloaded the latest FUSE version and built it myself. -----Original Message----- From: Kelvin Westlake Sent: 09 April 2010 17:13 To: 'gluster-users at gluster.org' Subject: Gluster crashing Hi Guys I have 2 physical servers, which I'm trying to setup in a HA pair.. I'm sharing 2 volumes (home-gfs and shared-gfs) across these servers in RAID1 (replicated), then I'm mounting clients on each server as home/ and shared/, if I lose one of the Servers and then remount it, the client on the other server seems to crash out. The following is the long entries leading to the crash - [2010-04-09 16:50:44] E [socket.c:762:socket_connect_finish] 192.168.100.31-2: connection to 192.168.100.31:6996 failed (Connection refused) [2010-04-09 16:50:44] E [socket.c:762:socket_connect_finish] 192.168.100.31-2: connection to 192.168.100.31:6996 failed (Connection refused) [2010-04-09 16:51:31] N [client-protocol.c:6246:client_setvolume_cbk] 192.168.100.31-2: Connected to 192.168.100.31:6996, attached to remote volume 'brick2'. [2010-04-09 16:51:31] E [afr-self-heal-common.c:1237:sh_missing_entries_create] mirror-1: unknown file type: 01 pending frames: frame : type(1) op(READDIRP) patchset: git://git.sv.gnu.org/gluster.git signal received: 11 time of crash: 2010-04-09 16:51:31 configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.0.0git [0x371420] /usr/local/lib/glusterfs/3.0.0git/xlator/protocol/client.so(client_readd irp+0x1b4)[0x28d224] /usr/local/lib/glusterfs/3.0.0git/xlator/cluster/replicate.so(afr_do_rea ddir+0x4e2)[0x98a722] /usr/local/lib/glusterfs/3.0.0git/xlator/cluster/replicate.so(afr_readdi rp+0x48)[0x98a988] /usr/local/lib/glusterfs/3.0.0git/xlator/mount/fuse.so[0x113e98] /usr/local/lib/glusterfs/3.0.0git/xlator/mount/fuse.so[0x11b25a] /lib/libpthread.so.0[0xbeb73b] /lib/libc.so.6(clone+0x5e)[0xb69cfe] --------- This email with any attachments is for the exclusive and confidential use of the addressee(s) and may contain legally privileged information. Any other distribution, use or reproduction without the senders prior consent is unauthorised and strictly prohibited. If you receive this message in error please notify the sender by email and delete the message from your computer. Netbasic Limited registered office and business address is 9 Funtley Court, Funtley Hill, Fareham, Hampshire PO16 7UY. Company No. 04906681. Netbasic Limited is authorised and regulated by the Financial Services Authority in respect of regulated activities. Please note that many of our activities do not require FSA regulation. _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
Burnash, James
2010-May-12 14:29 UTC
[Gluster-users] Best Practices for Gluster Replication
Greetings List,
I've searched through the Gluster wiki and a lot of threads to try to answer
this question, but so far no real luck.
Simply put - is it better to have replication handled by the clients, or by the
bricks themselves?
Volgen for a raid 1 solution creates a config file that does the mirroring on
the client side - which I would take as an implicit endorsement from the Gluster
team (great team, BTW). However, it seems to me that if the bricks replicated
between themselves on our 10Gb storage network, it could save a lot of bandwidth
for the clients and conceivably save them CPU cycles an I/O as well.
Client machines have 1Gb connections to the storage network, and are running
CentOS 5.2.
Server machines have 10Gb connections to the storage network, and are running
CentOS 5.4.
Glusterfs.vol:
## file auto generated by /usr/bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /usr/bin/glusterfs-volgen --name testfs --raid 1
jc1letgfs13-pfs1:/export/read-write jc1letgfs14-pfs1:/export/read-write
jc1letgfs15-pfs1:/export/read-write jc1letgfs16-pfs1:/export/read-write
jc1letgfs17-pfs1:/export/read-write jc1letgfs18-pfs1:/export/read-write
# RAID 1
# TRANSPORT-TYPE tcp
volume jc1letgfs17-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs17-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume jc1letgfs18-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs18-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume jc1letgfs13-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs13-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume jc1letgfs15-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs15-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume jc1letgfs16-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs16-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume jc1letgfs14-pfs1-1
type protocol/client
option transport-type tcp
option remote-host jc1letgfs14-pfs1
option transport.socket.nodelay on
option transport.remote-port 6996
option remote-subvolume brick1
end-volume
volume mirror-0
type cluster/replicate
subvolumes jc1letgfs13-pfs1-1 jc1letgfs14-pfs1-1
end-volume
volume mirror-1
type cluster/replicate
subvolumes jc1letgfs15-pfs1-1 jc1letgfs16-pfs1-1
end-volume
volume mirror-2
type cluster/replicate
subvolumes jc1letgfs17-pfs1-1 jc1letgfs18-pfs1-1
end-volume
volume distribute
type cluster/distribute
subvolumes mirror-0 mirror-1 mirror-2
end-volume
volume readahead
type performance/read-ahead
option page-count 4
subvolumes distribute
end-volume
volume iocache
type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
option cache-timeout 1
subvolumes readahead
end-volume
volume quickread
type performance/quick-read
option cache-timeout 1
option max-file-size 64kB
subvolumes iocache
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes quickread
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes writebehind
end-volume
Glusterfsd.vol:
## file auto generated by /usr/bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /usr/bin/glusterfs-volgen --name testfs jc1letgfs13-pfs1:/export/read-write
jc1letgfs14-pfs1:/export/read-write jc1letgfs15-pfs1:/export/read-write
volume posix1
type storage/posix
option directory /export/read-write
end-volume
volume locks1
type features/locks
subvolumes posix1
end-volume
volume brick1
type performance/io-threads
option thread-count 8
subvolumes locks1
end-volume
volume server-tcp
type protocol/server
option transport-type tcp
option auth.addr.brick1.allow *
option transport.socket.listen-port 6996
option transport.socket.nodelay on
subvolumes brick1
end-volume
James Burnash
DISCLAIMER:
This e-mail, and any attachments thereto, is intended only for use by the
addressee(s) named herein and may contain legally privileged and/or confidential
information. If you are not the intended recipient of this e-mail, you are
hereby notified that any dissemination, distribution or copying of this e-mail,
and any attachments thereto, is strictly prohibited. If you have received this
in error, please immediately notify me and permanently delete the original and
any copy of any e-mail and any printout thereof. E-mail transmission cannot be
guaranteed to be secure or error-free. The sender therefore does not accept
liability for any errors or omissions in the contents of this message which
arise as a result of e-mail transmission.
NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its
discretion, monitor and review the content of all e-mail communications.
http://www.knight.com