ABHISHEK PALIWAL
2018-Sep-25 08:46 UTC
[Gluster-users] [Gluster-devel] Crash in glusterfs!!!
I don't have the step to reproduce, but its a race condition where it seems cleanup_and_exit() is accessing the data structure which are not yet initialised (as gluster is in starting phase), due to SIGTERM/SIGINT is sent in between. Regards, Abhishek On Mon, Sep 24, 2018 at 9:11 PM Pranith Kumar Karampuri <pkarampu at redhat.com> wrote:> > > On Mon, Sep 24, 2018 at 5:16 PM ABHISHEK PALIWAL <abhishpaliwal at gmail.com> > wrote: > >> Hi Pranith, >> >> As we know this problem is getting triggered at startup of the glusterd >> process when it received the SIGTERM. >> >> I think there is a problem in glusterfs code, if at startup someone sent >> the SIGTERM the exit handler should not be crash instead it should with >> some information. >> >> Could please let me know the possibility to fix it from glusterfs side? >> > > I am not as confident as you about the RC you provided. If you could give > the steps to re-create, I will be happy to confirm that the RC is correct > and then I will send out the fix. > > >> >> Regards, >> Abhishek >> >> On Mon, Sep 24, 2018 at 3:12 PM Pranith Kumar Karampuri < >> pkarampu at redhat.com> wrote: >> >>> >>> >>> On Mon, Sep 24, 2018 at 2:09 PM ABHISHEK PALIWAL < >>> abhishpaliwal at gmail.com> wrote: >>> >>>> Could you please let me know about the bug in libc which you are >>>> talking. >>>> >>> >>> No, I mean, if you give the steps to reproduce, we will be able to pin >>> point if the issue is with libc or glusterfs. >>> >>> >>>> >>>> On Mon, Sep 24, 2018 at 2:01 PM Pranith Kumar Karampuri < >>>> pkarampu at redhat.com> wrote: >>>> >>>>> >>>>> >>>>> On Mon, Sep 24, 2018 at 1:57 PM ABHISHEK PALIWAL < >>>>> abhishpaliwal at gmail.com> wrote: >>>>> >>>>>> If you see the source code in cleanup_and_exit() we are getting the >>>>>> SIGSEGV crash when 'exit(0)' is triggered. >>>>>> >>>>> >>>>> yes, that is what I was mentioning earlier. It is crashing in libc. So >>>>> either there is a bug in libc (glusterfs actually found 1 bug so far in >>>>> libc, so I wouldn't rule out that possibility) or there is something that >>>>> is happening in glusterfs which is leading to the problem. >>>>> Valgrind/address-sanitizer would help find where the problem could be in >>>>> some cases, so before reaching out libc developers, it is better to figure >>>>> out where the problem is. Do you have steps to recreate it? >>>>> >>>>> >>>>>> >>>>>> On Mon, Sep 24, 2018 at 1:41 PM Pranith Kumar Karampuri < >>>>>> pkarampu at redhat.com> wrote: >>>>>> >>>>>>> >>>>>>> >>>>>>> On Mon, Sep 24, 2018 at 1:36 PM ABHISHEK PALIWAL < >>>>>>> abhishpaliwal at gmail.com> wrote: >>>>>>> >>>>>>>> Hi Sanju, >>>>>>>> >>>>>>>> Do you have any update on this? >>>>>>>> >>>>>>> >>>>>>> This seems to happen while the process is dying, in libc. I am not >>>>>>> completely sure if there is anything glusterfs is contributing to it from >>>>>>> the bt at the moment. Do you have any steps to re-create this problem? It >>>>>>> is probably better to run the steps with valgrind/address-sanitizer and see >>>>>>> if it points to the problem in glusterfs. >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> Regards, >>>>>>>> Abhishek >>>>>>>> >>>>>>>> On Fri, Sep 21, 2018 at 4:07 PM ABHISHEK PALIWAL < >>>>>>>> abhishpaliwal at gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi Sanju, >>>>>>>>> >>>>>>>>> Output of 't a a bt full' >>>>>>>>> >>>>>>>>> (gdb) t a a bt full >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thread 7 (LWP 1743): >>>>>>>>> >>>>>>>>> #0 0x00003fffa3ea7e88 in __lll_lock_wait (futex=0x0, private=0) >>>>>>>>> at lowlevellock.c:43 >>>>>>>>> >>>>>>>>> r4 = 128 >>>>>>>>> >>>>>>>>> r7 = 0 >>>>>>>>> >>>>>>>>> arg2 = 128 >>>>>>>>> >>>>>>>>> r5 = 2 >>>>>>>>> >>>>>>>>> r8 = 1 >>>>>>>>> >>>>>>>>> r0 = 221 >>>>>>>>> >>>>>>>>> r3 = 0 >>>>>>>>> >>>>>>>>> r6 = 0 >>>>>>>>> >>>>>>>>> arg1 = 0 >>>>>>>>> >>>>>>>>> __err = 221 >>>>>>>>> >>>>>>>>> __ret = 0 >>>>>>>>> >>>>>>>>> #1 0x00003fffa3e9ef64 in __GI___pthread_mutex_lock >>>>>>>>> (mutex=0x100272a8) at ../nptl/pthread_mutex_lock.c:81 >>>>>>>>> >>>>>>>>> __futex = 0x100272a8 >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "__pthread_mutex_lock" >>>>>>>>> >>>>>>>>> type = <optimized out> >>>>>>>>> >>>>>>>>> id = <optimized out> >>>>>>>>> >>>>>>>>> #2 0x00003fffa3f6ce8c in _gf_msg (domain=0x3fff98006c90 >>>>>>>>> "c_glusterfs-client-0", file=0x3fff9fb34de0 "client.c", >>>>>>>>> function=0x3fff9fb34cd8 <__FUNCTION__.18849> "notify", >>>>>>>>> >>>>>>>>> line=<optimized out>, level=<optimized out>, errnum=<optimized >>>>>>>>> out>, trace=<optimized out>, msgid=114020, >>>>>>>>> >>>>>>>>> fmt=0x3fff9fb35350 "parent translators are ready, attempting >>>>>>>>> connect on transport") at logging.c:2058 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> msgstr = <optimized out> >>>>>>>>> >>>>>>>>> ap = <optimized out> >>>>>>>>> >>>>>>>>> this = 0x3fff980061f0 >>>>>>>>> >>>>>>>>> ctx = 0x10027010 >>>>>>>>> >>>>>>>>> callstr = '\000' <repeats 4095 times> >>>>>>>>> >>>>>>>>> passcallstr = 0 >>>>>>>>> >>>>>>>>> log_inited = 0 >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "_gf_msg" >>>>>>>>> >>>>>>>>> #3 0x00003fff9fb084ac in notify (this=0x3fff980061f0, >>>>>>>>> event=<optimized out>, data=0x3fff98008c50) at client.c:2116 >>>>>>>>> >>>>>>>>> conf = 0x3fff98056dd0 >>>>>>>>> >>>>>>>>> __FUNCTION__ = "notify" >>>>>>>>> >>>>>>>>> #4 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff980061f0, >>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>> >>>>>>>>> old_THIS = 0x3fff98008c50 >>>>>>>>> >>>>>>>>> ret = 0 >>>>>>>>> >>>>>>>>> #5 0x00003fffa3f87700 in default_notify (this=0x3fff98008c50, >>>>>>>>> event=<optimized out>, data=<optimized out>) at defaults.c:2302 >>>>>>>>> >>>>>>>>> list = 0x3fff9800a340 >>>>>>>>> >>>>>>>>> #6 0x00003fff9fac922c in afr_notify (this=0x3fff98008c50, >>>>>>>>> event=1, data=0x3fff9800a4c0, data2=<optimized out>) at afr-common.c:3967 >>>>>>>>> >>>>>>>>> priv = 0x3fff98010050 >>>>>>>>> >>>>>>>>> i = <optimized out> >>>>>>>>> >>>>>>>>> up_children = <optimized out> >>>>>>>>> >>>>>>>>> down_children = <optimized out> >>>>>>>>> >>>>>>>>> propagate = 1 >>>>>>>>> >>>>>>>>> had_heard_from_all = <optimized out> >>>>>>>>> >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> >>>>>>>>> have_heard_from_all = 0 >>>>>>>>> >>>>>>>>> idx = <optimized out> >>>>>>>>> >>>>>>>>> ret = 0 >>>>>>>>> >>>>>>>>> call_psh = <optimized out> >>>>>>>>> >>>>>>>>> input = 0x0 >>>>>>>>> >>>>>>>>> output = 0x0 >>>>>>>>> >>>>>>>>> had_quorum = <optimized out> >>>>>>>>> >>>>>>>>> has_quorum = <optimized out> >>>>>>>>> >>>>>>>>> __FUNCTION__ = "afr_notify" >>>>>>>>> >>>>>>>>> #7 0x00003fff9fad4994 in notify (this=<optimized out>, >>>>>>>>> event=<optimized out>, data=<optimized out>) at afr.c:38 >>>>>>>>> >>>>>>>>> ret = -1 >>>>>>>>> >>>>>>>>> ap = 0x3fffa034cc58 "" >>>>>>>>> >>>>>>>>> data2 = <optimized out> >>>>>>>>> >>>>>>>>> #8 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff98008c50, >>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>> >>>>>>>>> old_THIS = 0x3fff9800a4c0 >>>>>>>>> >>>>>>>>> ret = 0 >>>>>>>>> >>>>>>>>> #9 0x00003fffa3f87700 in default_notify (this=0x3fff9800a4c0, >>>>>>>>> event=<optimized out>, data=<optimized out>) at defaults.c:2302 >>>>>>>>> >>>>>>>>> list = 0x3fff9800b710 >>>>>>>>> >>>>>>>>> #10 0x00003fff9fa6b1e4 in notify (this=<optimized out>, >>>>>>>>> event=<optimized out>, data=<optimized out>) at io-stats.c:3064 >>>>>>>>> >>>>>>>>> ret = 0 >>>>>>>>> >>>>>>>>> args = {type = IOS_DUMP_TYPE_NONE, u = {logfp = 0x0, dict >>>>>>>>> = 0x0}} >>>>>>>>> >>>>>>>>> op = 0 >>>>>>>>> >>>>>>>>> list_cnt = 0 >>>>>>>>> >>>>>>>>> throughput = 0 >>>>>>>>> >>>>>>>>> time = 0 >>>>>>>>> >>>>>>>>> is_peek = _gf_false >>>>>>>>> >>>>>>>>> ap = 0x3fffa034ce68 "" >>>>>>>>> >>>>>>>>> __FUNCTION__ = "notify" >>>>>>>>> >>>>>>>>> #11 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff9800a4c0, >>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>> >>>>>>>>> old_THIS = 0x3fffa402d290 <global_xlator> >>>>>>>>> >>>>>>>>> ret = 0 >>>>>>>>> >>>>>>>>> #12 0x00003fffa3fbd560 in glusterfs_graph_parent_up >>>>>>>>> (graph=<optimized out>) at graph.c:440 >>>>>>>>> >>>>>>>>> trav = 0x3fff9800a4c0 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> #13 0x00003fffa3fbdb90 in glusterfs_graph_activate >>>>>>>>> (graph=0x3fff98000af0, ctx=0x10027010) at graph.c:688 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> __FUNCTION__ = "glusterfs_graph_activate" >>>>>>>>> >>>>>>>>> #14 0x000000001000a49c in glusterfs_process_volfp (ctx=0x10027010, >>>>>>>>> fp=0x3fff98001cd0) at glusterfsd.c:2221 >>>>>>>>> >>>>>>>>> graph = 0x3fff98000af0 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> trav = <optimized out> >>>>>>>>> >>>>>>>>> __FUNCTION__ = <error reading variable __FUNCTION__ >>>>>>>>> (Cannot access memory at address 0x10010ec0)> >>>>>>>>> >>>>>>>>> #15 0x000000001000fd08 in mgmt_getspec_cbk (req=<optimized out>, >>>>>>>>> iov=<optimized out>, count=<optimized out>, myframe=0x3fffa2bea06c) at >>>>>>>>> glusterfsd-mgmt.c:1561 >>>>>>>>> >>>>>>>>> rsp = {op_ret = 1059, op_errno = 0, >>>>>>>>> >>>>>>>>> spec = 0x3fff980018a0 "volume c_glusterfs-client-0\n >>>>>>>>> type protocol/client\n option password >>>>>>>>> 5fd8d83d-99f3-4630-97c6-965d7a8ead62\n option username >>>>>>>>> e65687aa-e135-445e-8778-48bb8fb19640\n option transport-type tcp\n >>>>>>>>> "..., xdata = {xdata_len = 0, xdata_val = 0x0}} >>>>>>>>> >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> >>>>>>>>> frame = 0x3fffa2bea06c >>>>>>>>> >>>>>>>>> ctx = 0x10027010 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> size = 1059 >>>>>>>>> >>>>>>>>> tmpfp = 0x3fff98001cd0 >>>>>>>>> >>>>>>>>> volfilebuf = 0x0 >>>>>>>>> >>>>>>>>> __FUNCTION__ = <error reading variable __FUNCTION__ (Cannot >>>>>>>>> access memory at address 0x10013570)> >>>>>>>>> >>>>>>>>> #16 0x00003fffa3f21ec4 in rpc_clnt_handle_reply (clnt=0x10089020, >>>>>>>>> pollin=0x3fff98001760) at rpc-clnt.c:775 >>>>>>>>> >>>>>>>>> conn = 0x10089050 >>>>>>>>> >>>>>>>>> saved_frame = <optimized out> >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> req = 0x1008931c >>>>>>>>> >>>>>>>>> xid = 1 >>>>>>>>> >>>>>>>>> __FUNCTION__ = "rpc_clnt_handle_reply" >>>>>>>>> >>>>>>>>> #17 0x00003fffa3f223d0 in rpc_clnt_notify (trans=<optimized out>, >>>>>>>>> mydata=0x10089050, event=<optimized out>, data=<optimized out>) at >>>>>>>>> rpc-clnt.c:933 >>>>>>>>> >>>>>>>>> conn = 0x10089050 >>>>>>>>> >>>>>>>>> clnt = <optimized out> >>>>>>>>> >>>>>>>>> ret = -1 >>>>>>>>> >>>>>>>>> req_info = 0x0 >>>>>>>>> >>>>>>>>> pollin = <optimized out> >>>>>>>>> >>>>>>>>> clnt_mydata = 0x0 >>>>>>>>> >>>>>>>>> old_THIS = 0x3fffa402d290 <global_xlator> >>>>>>>>> >>>>>>>>> __FUNCTION__ = "rpc_clnt_notify" >>>>>>>>> >>>>>>>>> #18 0x00003fffa3f1d4fc in rpc_transport_notify (this=<optimized >>>>>>>>> out>, event=<optimized out>, data=<optimized out>) at rpc-transport.c:546 >>>>>>>>> >>>>>>>>> ret = -1 >>>>>>>>> >>>>>>>>> __FUNCTION__ = "rpc_transport_notify" >>>>>>>>> >>>>>>>>> #19 0x00003fffa0401d44 in socket_event_poll_in (this=this at entry=0x1008ab80) >>>>>>>>> at socket.c:2236 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> pollin = 0x3fff98001760 >>>>>>>>> >>>>>>>>> priv = 0x1008b820 >>>>>>>>> >>>>>>>>> #20 0x00003fffa040489c in socket_event_handler (fd=<optimized >>>>>>>>> out>, idx=<optimized out>, data=0x1008ab80, poll_in=<optimized out>, >>>>>>>>> poll_out=<optimized out>, poll_err=<optimized out>) >>>>>>>>> >>>>>>>>> at socket.c:2349 >>>>>>>>> >>>>>>>>> this = 0x1008ab80 >>>>>>>>> >>>>>>>>> priv = 0x1008b820 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> __FUNCTION__ = "socket_event_handler" >>>>>>>>> >>>>>>>>> #21 0x00003fffa3fe2874 in event_dispatch_epoll_handler >>>>>>>>> (event=0x3fffa034d6a0, event_pool=0x10045bc0) at event-epoll.c:575 >>>>>>>>> >>>>>>>>> handler = @0x3fffa041f620: 0x3fffa04046f0 >>>>>>>>> <socket_event_handler> >>>>>>>>> >>>>>>>>> gen = 1 >>>>>>>>> >>>>>>>>> slot = 0x1007cd80 >>>>>>>>> >>>>>>>>> data = <optimized out> >>>>>>>>> >>>>>>>>> ret = -1 >>>>>>>>> >>>>>>>>> fd = 9 >>>>>>>>> >>>>>>>>> ev_data = 0x3fffa034d6a8 >>>>>>>>> >>>>>>>>> idx = 1 >>>>>>>>> >>>>>>>>> #22 event_dispatch_epoll_worker (data=0x1008bd50) at >>>>>>>>> event-epoll.c:678 >>>>>>>>> >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> >>>>>>>>> event = {events = 1, data = {ptr = 0x100000001, fd = 1, >>>>>>>>> u32 = 1, u64 = 4294967297}} >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> ev_data = 0x1008bd50 >>>>>>>>> >>>>>>>>> event_pool = 0x10045bc0 >>>>>>>>> >>>>>>>>> myindex = <optimized out> >>>>>>>>> >>>>>>>>> timetodie = 0 >>>>>>>>> >>>>>>>>> __FUNCTION__ = "event_dispatch_epoll_worker" >>>>>>>>> >>>>>>>>> #23 0x00003fffa3e9bb30 in start_thread (arg=0x3fffa034e160) at >>>>>>>>> pthread_create.c:462 >>>>>>>>> >>>>>>>>> pd = 0x3fffa034e160 >>>>>>>>> >>>>>>>>> now = <optimized out> >>>>>>>>> >>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>> {2380233324717718430, 70367199403008, 2380233324703897146, 0, 0, >>>>>>>>> 70367128645632, 70367137030688, 8388608, 70367199363104, 269008208, >>>>>>>>> >>>>>>>>> 70368094386592, 70367199388632, 70367200825640, 3, >>>>>>>>> 0, 70367199388648, 70368094386240, 70368094386296, 4001536, 70367199364120, >>>>>>>>> 70367137027904, -3187653596, >>>>>>>>> >>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>> canceltype = 0}}} >>>>>>>>> >>>>>>>>> not_first_call = <optimized out> >>>>>>>>> >>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>> >>>>>>>>> sp = <optimized out> >>>>>>>>> >>>>>>>>> freesize = <optimized out> >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>> >>>>>>>>> #24 0x00003fffa3de60fc in .__clone () at >>>>>>>>> ../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96 >>>>>>>>> >>>>>>>>> No locals. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thread 6 (LWP 1735): >>>>>>>>> >>>>>>>>> #0 0x00003fffa3ea3ccc in __pthread_cond_timedwait >>>>>>>>> (cond=0x10059a98, mutex=0x10059a70, abstime=0x3fffa141f670) at >>>>>>>>> pthread_cond_timedwait.c:198 >>>>>>>>> >>>>>>>>> r4 = 393 >>>>>>>>> >>>>>>>>> r7 = 0 >>>>>>>>> >>>>>>>>> arg5 = 0 >>>>>>>>> >>>>>>>>> arg2 = <optimized out> >>>>>>>>> >>>>>>>>> r5 = 2 >>>>>>>>> >>>>>>>>> r8 = 4294967295 >>>>>>>>> >>>>>>>>> arg6 = 4294967295 >>>>>>>>> >>>>>>>>> arg3 = 2 >>>>>>>>> >>>>>>>>> r0 = 221 >>>>>>>>> >>>>>>>>> r3 = 516 >>>>>>>>> >>>>>>>>> r6 = 70367154665072 >>>>>>>>> >>>>>>>>> arg4 = 70367154665072 >>>>>>>>> >>>>>>>>> arg1 = 268802716 >>>>>>>>> >>>>>>>>> __err = <optimized out> >>>>>>>>> >>>>>>>>> __ret = <optimized out> >>>>>>>>> >>>>>>>>> futex_val = 2 >>>>>>>>> >>>>>>>>> buffer = {__routine = @0x3fffa3ec0b50: 0x3fffa3ea3400 >>>>>>>>> <__condvar_cleanup>, __arg = 0x3fffa141f540, __canceltype = 0, __prev = 0x0} >>>>>>>>> >>>>>>>>> cbuffer = {oldtype = 0, cond = 0x10059a98, mutex >>>>>>>>> 0x10059a70, bc_seq = 0} >>>>>>>>> >>>>>>>>> result = 0 >>>>>>>>> >>>>>>>>> pshared = 0 >>>>>>>>> >>>>>>>>> pi_flag = 0 >>>>>>>>> >>>>>>>>> err = <optimized out> >>>>>>>>> >>>>>>>>> val = <optimized out> >>>>>>>>> >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> >>>>>>>>> seq = 0 >>>>>>>>> >>>>>>>>> #1 0x00003fffa3fc0e74 in syncenv_task (proc=0x10053eb0) at >>>>>>>>> syncop.c:607 >>>>>>>>> >>>>>>>>> env = 0x10053eb0 >>>>>>>>> >>>>>>>>> task = 0x0 >>>>>>>>> >>>>>>>>> sleep_till = {tv_sec = 1536845230, tv_nsec = 0} >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> #2 0x00003fffa3fc1cdc in syncenv_processor (thdata=0x10053eb0) at >>>>>>>>> syncop.c:699 >>>>>>>>> >>>>>>>>> env = 0x10053eb0 >>>>>>>>> >>>>>>>>> proc = 0x10053eb0 >>>>>>>>> >>>>>>>>> task = <optimized out> >>>>>>>>> >>>>>>>>> #3 0x00003fffa3e9bb30 in start_thread (arg=0x3fffa1420160) at >>>>>>>>> pthread_create.c:462 >>>>>>>>> >>>>>>>>> pd = 0x3fffa1420160 >>>>>>>>> >>>>>>>>> now = <optimized out> >>>>>>>>> >>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>> {2380233324731669406, 70367199403008, 2380233324703897146, 0, 0, >>>>>>>>> 70367146283008, 70367154668064, 8388608, 70367199363104, 268779184, >>>>>>>>> >>>>>>>>> 268779184, 70367199388632, 70367200820192, 3, 0, >>>>>>>>> 70367199388648, 70368094386080, 70368094386136, 4001536, 70367199364120, >>>>>>>>> 70367154665280, -3187653564, >>>>>>>>> >>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>> canceltype = 0}}} >>>>>>>>> >>>>>>>>> not_first_call = <optimized out> >>>>>>>>> >>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>> >>>>>>>>> sp = <optimized out> >>>>>>>>> >>>>>>>>> freesize = <optimized out> >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>> >>>>>>>>> #4 0x00003fffa3de60fc in .__clone () at >>>>>>>>> ../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96 >>>>>>>>> >>>>>>>>> No locals. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thread 5 (LWP 1746): >>>>>>>>> >>>>>>>>> #0 0x00003fffa3ea7e38 in __lll_lock_wait (futex=0x100272a8, >>>>>>>>> private=<optimized out>) at lowlevellock.c:46 >>>>>>>>> >>>>>>>>> r4 = 128 >>>>>>>>> >>>>>>>>> r7 = 0 >>>>>>>>> >>>>>>>>> r5 = 2 >>>>>>>>> >>>>>>>>> r8 = 1 >>>>>>>>> >>>>>>>>> arg3 = 2 >>>>>>>>> >>>>>>>>> r0 = 221 >>>>>>>>> >>>>>>>>> r3 = 512 >>>>>>>>> >>>>>>>>> r6 = 0 >>>>>>>>> >>>>>>>>> arg4 = 0 >>>>>>>>> >>>>>>>>> arg1 = 268595880 >>>>>>>>> >>>>>>>>> __err = <optimized out> >>>>>>>>> >>>>>>>>> __ret = <optimized out> >>>>>>>>> >>>>>>>>> #1 0x00003fffa3e9ef64 in __GI___pthread_mutex_lock >>>>>>>>> (mutex=0x100272a8) at ../nptl/pthread_mutex_lock.c:81 >>>>>>>>> >>>>>>>>> __futex = 0x100272a8 >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "__pthread_mutex_lock" >>>>>>>>> >>>>>>>>> type = <optimized out> >>>>>>>>> >>>>>>>>> id = <optimized out> >>>>>>>>> >>>>>>>>> #2 0x00003fffa3f6ce8c in _gf_msg (domain=0x3fffa4009e38 "epoll", >>>>>>>>> file=0x3fffa4009e28 "event-epoll.c", function=0x3fffa4009db8 >>>>>>>>> <__FUNCTION__.8510> "event_dispatch_epoll_worker", >>>>>>>>> >>>>>>>>> line=<optimized out>, level=<optimized out>, errnum=<optimized >>>>>>>>> out>, trace=<optimized out>, msgid=101190, fmt=0x3fffa4009f48 "Started >>>>>>>>> thread with index %d") at logging.c:2058 >>>>>>>>> >>>>>>>>> ret = <optimized out> >>>>>>>>> >>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>> >>>>>>>>> msgstr = <optimized out> >>>>>>>>> >>>>>>>>> ap = <optimized out> >>>>>>>>> >>>>>>>>> this = 0x3fffa402d290 <global_xlator> >>>>>>>>> >>>>>>>>> ctx = 0x10027010 >>>>>>>>> >>>>>>>>> callstr = '\000' <repeats 4095 times> >>>>>>>>> >>>>>>>>> passcallstr = 0 >>>>>>>>> >>>>>>>>> log_inited = 0 >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "_gf_msg" >>>>>>>>> >>>>>>>>> #3 0x00003fffa3fe265c in event_dispatch_epoll_worker >>>>>>>>> (data=0x3fff9802ffd0) at event-epoll.c:631 >>>>>>>>> >>>>>>>>> event = {events = 0, data = {ptr = 0x0, fd = 0, u32 = 0, >>>>>>>>> u64 = 0}} >>>>>>>>> >>>>>>>>> ret = -1 >>>>>>>>> >>>>>>>>> ev_data = 0x3fff9802ffd0 >>>>>>>>> >>>>>>>>> event_pool = 0x10045bc0 >>>>>>>>> >>>>>>>>> myindex = <optimized out> >>>>>>>>> >>>>>>>>> timetodie = 0 >>>>>>>>> >>>>>>>>> __FUNCTION__ = "event_dispatch_epoll_worker" >>>>>>>>> >>>>>>>>> #4 0x00003fffa3e9bb30 in start_thread (arg=0x3fff9f37d160) at >>>>>>>>> pthread_create.c:462 >>>>>>>>> >>>>>>>>> pd = 0x3fff9f37d160 >>>>>>>>> >>>>>>>>> now = <optimized out> >>>>>>>>> >>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>> {2380233325439331230, 70367199403008, 2380233324703897146, 0, 269006880, >>>>>>>>> 70367112060928, 70367120445984, 8388608, 70367199363104, >>>>>>>>> >>>>>>>>> 70366999543760, 70367137025008, 70367199388632, >>>>>>>>> 70367200825640, 3, 0, 70367199388648, 70367137024656, 70367137024712, >>>>>>>>> 4001536, 70367199364120, 70367120443200, -3183328188, >>>>>>>>> >>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>> canceltype = 0}}} >>>>>>>>> >>>>>>>>> not_first_call = <optimized out> >>>>>>>>> >>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>> >>>>>>>>> sp = <optimized out> >>>>>>>>> >>>>>>>>> freesize = <optimized out> >>>>>>>>> >>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>> >>>>>>>>> #5 0x00003fffa3de60fc in .__clone () at >>>>>>>>> ../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96 >>>>>>>>> >>>>>>>>> No locals. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thread 4 (LWP 1733): >>>>>>>>> >>>>>>>>> #0 0x00003fffa3ea9150 in .__nanosleep () at >>>>>>>>> ../sysdeps/unix/syscall-template.S:84 >>>>>>>>> >>>>>>>>> No locals. >>>>>>>>> >>>>>>>>> #1 0x00003fffa3f93f28 in gf_timer_proc (ctx=0x10027010) at >>>>>>>>> timer.c:205 >>>>>>>>> >>>>>>>>> now = 20528921491 >>>>>>>>> >>>>>>>>> now_ts = {tv_sec = 20, tv_nsec = 528921491} >>>>>>>>> >>>>>>>>> reg = 0x100533c0 >>>>>>>>> >>>>>>>>> sleepts = {tv_sec = 0, tv_nsec = 2000000} >>>>>>>>> >>>>>>>>> event = 0x1008d280 >>>>>>>>> >>>>>>>>> old_THIS = <optimized out> >>>>>>>>> >>>>>>>>> __FUNCTION__ = "gf_timer_proc" >>>>>>>>> >>>>>>>>> #2 0x00003fffa3e9bb30 in start_thread (arg=0x3fffa2420160) at >>>>>>>>> pthread_create.c:462 >>>>>>>>> >>>>>>>>> pd = 0x3fffa2420160 >>>>>>>>> >>>>>>>>> now = <optimized out> >>>>>>>>> >>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>> {2380233324681337758, 70367199403008, 2380233 >>>>>>>>> >>>>>>>>-- Regards Abhishek Paliwal -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180925/332781ab/attachment.html>
Pranith Kumar Karampuri
2018-Sep-25 08:57 UTC
[Gluster-users] [Gluster-devel] Crash in glusterfs!!!
On Tue, Sep 25, 2018 at 2:17 PM ABHISHEK PALIWAL <abhishpaliwal at gmail.com> wrote:> I don't have the step to reproduce, but its a race condition where it > seems cleanup_and_exit() is accessing the data structure which are not yet > initialised (as gluster is in starting phase), due to SIGTERM/SIGINT is > sent in between. >But the crash happened inside exit() code for which will be in libc which doesn't access any data structures in glusterfs.> > Regards, > Abhishek > > On Mon, Sep 24, 2018 at 9:11 PM Pranith Kumar Karampuri < > pkarampu at redhat.com> wrote: > >> >> >> On Mon, Sep 24, 2018 at 5:16 PM ABHISHEK PALIWAL <abhishpaliwal at gmail.com> >> wrote: >> >>> Hi Pranith, >>> >>> As we know this problem is getting triggered at startup of the glusterd >>> process when it received the SIGTERM. >>> >>> I think there is a problem in glusterfs code, if at startup someone sent >>> the SIGTERM the exit handler should not be crash instead it should with >>> some information. >>> >>> Could please let me know the possibility to fix it from glusterfs side? >>> >> >> I am not as confident as you about the RC you provided. If you could give >> the steps to re-create, I will be happy to confirm that the RC is correct >> and then I will send out the fix. >> >> >>> >>> Regards, >>> Abhishek >>> >>> On Mon, Sep 24, 2018 at 3:12 PM Pranith Kumar Karampuri < >>> pkarampu at redhat.com> wrote: >>> >>>> >>>> >>>> On Mon, Sep 24, 2018 at 2:09 PM ABHISHEK PALIWAL < >>>> abhishpaliwal at gmail.com> wrote: >>>> >>>>> Could you please let me know about the bug in libc which you are >>>>> talking. >>>>> >>>> >>>> No, I mean, if you give the steps to reproduce, we will be able to pin >>>> point if the issue is with libc or glusterfs. >>>> >>>> >>>>> >>>>> On Mon, Sep 24, 2018 at 2:01 PM Pranith Kumar Karampuri < >>>>> pkarampu at redhat.com> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Mon, Sep 24, 2018 at 1:57 PM ABHISHEK PALIWAL < >>>>>> abhishpaliwal at gmail.com> wrote: >>>>>> >>>>>>> If you see the source code in cleanup_and_exit() we are getting the >>>>>>> SIGSEGV crash when 'exit(0)' is triggered. >>>>>>> >>>>>> >>>>>> yes, that is what I was mentioning earlier. It is crashing in libc. >>>>>> So either there is a bug in libc (glusterfs actually found 1 bug so far in >>>>>> libc, so I wouldn't rule out that possibility) or there is something that >>>>>> is happening in glusterfs which is leading to the problem. >>>>>> Valgrind/address-sanitizer would help find where the problem could be in >>>>>> some cases, so before reaching out libc developers, it is better to figure >>>>>> out where the problem is. Do you have steps to recreate it? >>>>>> >>>>>> >>>>>>> >>>>>>> On Mon, Sep 24, 2018 at 1:41 PM Pranith Kumar Karampuri < >>>>>>> pkarampu at redhat.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Sep 24, 2018 at 1:36 PM ABHISHEK PALIWAL < >>>>>>>> abhishpaliwal at gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi Sanju, >>>>>>>>> >>>>>>>>> Do you have any update on this? >>>>>>>>> >>>>>>>> >>>>>>>> This seems to happen while the process is dying, in libc. I am not >>>>>>>> completely sure if there is anything glusterfs is contributing to it from >>>>>>>> the bt at the moment. Do you have any steps to re-create this problem? It >>>>>>>> is probably better to run the steps with valgrind/address-sanitizer and see >>>>>>>> if it points to the problem in glusterfs. >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> Regards, >>>>>>>>> Abhishek >>>>>>>>> >>>>>>>>> On Fri, Sep 21, 2018 at 4:07 PM ABHISHEK PALIWAL < >>>>>>>>> abhishpaliwal at gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Sanju, >>>>>>>>>> >>>>>>>>>> Output of 't a a bt full' >>>>>>>>>> >>>>>>>>>> (gdb) t a a bt full >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thread 7 (LWP 1743): >>>>>>>>>> >>>>>>>>>> #0 0x00003fffa3ea7e88 in __lll_lock_wait (futex=0x0, private=0) >>>>>>>>>> at lowlevellock.c:43 >>>>>>>>>> >>>>>>>>>> r4 = 128 >>>>>>>>>> >>>>>>>>>> r7 = 0 >>>>>>>>>> >>>>>>>>>> arg2 = 128 >>>>>>>>>> >>>>>>>>>> r5 = 2 >>>>>>>>>> >>>>>>>>>> r8 = 1 >>>>>>>>>> >>>>>>>>>> r0 = 221 >>>>>>>>>> >>>>>>>>>> r3 = 0 >>>>>>>>>> >>>>>>>>>> r6 = 0 >>>>>>>>>> >>>>>>>>>> arg1 = 0 >>>>>>>>>> >>>>>>>>>> __err = 221 >>>>>>>>>> >>>>>>>>>> __ret = 0 >>>>>>>>>> >>>>>>>>>> #1 0x00003fffa3e9ef64 in __GI___pthread_mutex_lock >>>>>>>>>> (mutex=0x100272a8) at ../nptl/pthread_mutex_lock.c:81 >>>>>>>>>> >>>>>>>>>> __futex = 0x100272a8 >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "__pthread_mutex_lock" >>>>>>>>>> >>>>>>>>>> type = <optimized out> >>>>>>>>>> >>>>>>>>>> id = <optimized out> >>>>>>>>>> >>>>>>>>>> #2 0x00003fffa3f6ce8c in _gf_msg (domain=0x3fff98006c90 >>>>>>>>>> "c_glusterfs-client-0", file=0x3fff9fb34de0 "client.c", >>>>>>>>>> function=0x3fff9fb34cd8 <__FUNCTION__.18849> "notify", >>>>>>>>>> >>>>>>>>>> line=<optimized out>, level=<optimized out>, >>>>>>>>>> errnum=<optimized out>, trace=<optimized out>, msgid=114020, >>>>>>>>>> >>>>>>>>>> fmt=0x3fff9fb35350 "parent translators are ready, attempting >>>>>>>>>> connect on transport") at logging.c:2058 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> msgstr = <optimized out> >>>>>>>>>> >>>>>>>>>> ap = <optimized out> >>>>>>>>>> >>>>>>>>>> this = 0x3fff980061f0 >>>>>>>>>> >>>>>>>>>> ctx = 0x10027010 >>>>>>>>>> >>>>>>>>>> callstr = '\000' <repeats 4095 times> >>>>>>>>>> >>>>>>>>>> passcallstr = 0 >>>>>>>>>> >>>>>>>>>> log_inited = 0 >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "_gf_msg" >>>>>>>>>> >>>>>>>>>> #3 0x00003fff9fb084ac in notify (this=0x3fff980061f0, >>>>>>>>>> event=<optimized out>, data=0x3fff98008c50) at client.c:2116 >>>>>>>>>> >>>>>>>>>> conf = 0x3fff98056dd0 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "notify" >>>>>>>>>> >>>>>>>>>> #4 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff980061f0, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>>> >>>>>>>>>> old_THIS = 0x3fff98008c50 >>>>>>>>>> >>>>>>>>>> ret = 0 >>>>>>>>>> >>>>>>>>>> #5 0x00003fffa3f87700 in default_notify (this=0x3fff98008c50, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at defaults.c:2302 >>>>>>>>>> >>>>>>>>>> list = 0x3fff9800a340 >>>>>>>>>> >>>>>>>>>> #6 0x00003fff9fac922c in afr_notify (this=0x3fff98008c50, >>>>>>>>>> event=1, data=0x3fff9800a4c0, data2=<optimized out>) at afr-common.c:3967 >>>>>>>>>> >>>>>>>>>> priv = 0x3fff98010050 >>>>>>>>>> >>>>>>>>>> i = <optimized out> >>>>>>>>>> >>>>>>>>>> up_children = <optimized out> >>>>>>>>>> >>>>>>>>>> down_children = <optimized out> >>>>>>>>>> >>>>>>>>>> propagate = 1 >>>>>>>>>> >>>>>>>>>> had_heard_from_all = <optimized out> >>>>>>>>>> >>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>> >>>>>>>>>> have_heard_from_all = 0 >>>>>>>>>> >>>>>>>>>> idx = <optimized out> >>>>>>>>>> >>>>>>>>>> ret = 0 >>>>>>>>>> >>>>>>>>>> call_psh = <optimized out> >>>>>>>>>> >>>>>>>>>> input = 0x0 >>>>>>>>>> >>>>>>>>>> output = 0x0 >>>>>>>>>> >>>>>>>>>> had_quorum = <optimized out> >>>>>>>>>> >>>>>>>>>> has_quorum = <optimized out> >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "afr_notify" >>>>>>>>>> >>>>>>>>>> #7 0x00003fff9fad4994 in notify (this=<optimized out>, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at afr.c:38 >>>>>>>>>> >>>>>>>>>> ret = -1 >>>>>>>>>> >>>>>>>>>> ap = 0x3fffa034cc58 "" >>>>>>>>>> >>>>>>>>>> data2 = <optimized out> >>>>>>>>>> >>>>>>>>>> #8 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff98008c50, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>>> >>>>>>>>>> old_THIS = 0x3fff9800a4c0 >>>>>>>>>> >>>>>>>>>> ret = 0 >>>>>>>>>> >>>>>>>>>> #9 0x00003fffa3f87700 in default_notify (this=0x3fff9800a4c0, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at defaults.c:2302 >>>>>>>>>> >>>>>>>>>> list = 0x3fff9800b710 >>>>>>>>>> >>>>>>>>>> #10 0x00003fff9fa6b1e4 in notify (this=<optimized out>, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at io-stats.c:3064 >>>>>>>>>> >>>>>>>>>> ret = 0 >>>>>>>>>> >>>>>>>>>> args = {type = IOS_DUMP_TYPE_NONE, u = {logfp = 0x0, dict >>>>>>>>>> = 0x0}} >>>>>>>>>> >>>>>>>>>> op = 0 >>>>>>>>>> >>>>>>>>>> list_cnt = 0 >>>>>>>>>> >>>>>>>>>> throughput = 0 >>>>>>>>>> >>>>>>>>>> time = 0 >>>>>>>>>> >>>>>>>>>> is_peek = _gf_false >>>>>>>>>> >>>>>>>>>> ap = 0x3fffa034ce68 "" >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "notify" >>>>>>>>>> >>>>>>>>>> #11 0x00003fffa3f68ca0 in xlator_notify (xl=0x3fff9800a4c0, >>>>>>>>>> event=<optimized out>, data=<optimized out>) at xlator.c:491 >>>>>>>>>> >>>>>>>>>> old_THIS = 0x3fffa402d290 <global_xlator> >>>>>>>>>> >>>>>>>>>> ret = 0 >>>>>>>>>> >>>>>>>>>> #12 0x00003fffa3fbd560 in glusterfs_graph_parent_up >>>>>>>>>> (graph=<optimized out>) at graph.c:440 >>>>>>>>>> >>>>>>>>>> trav = 0x3fff9800a4c0 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> #13 0x00003fffa3fbdb90 in glusterfs_graph_activate >>>>>>>>>> (graph=0x3fff98000af0, ctx=0x10027010) at graph.c:688 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "glusterfs_graph_activate" >>>>>>>>>> >>>>>>>>>> #14 0x000000001000a49c in glusterfs_process_volfp >>>>>>>>>> (ctx=0x10027010, fp=0x3fff98001cd0) at glusterfsd.c:2221 >>>>>>>>>> >>>>>>>>>> graph = 0x3fff98000af0 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> trav = <optimized out> >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = <error reading variable __FUNCTION__ >>>>>>>>>> (Cannot access memory at address 0x10010ec0)> >>>>>>>>>> >>>>>>>>>> #15 0x000000001000fd08 in mgmt_getspec_cbk (req=<optimized out>, >>>>>>>>>> iov=<optimized out>, count=<optimized out>, myframe=0x3fffa2bea06c) at >>>>>>>>>> glusterfsd-mgmt.c:1561 >>>>>>>>>> >>>>>>>>>> rsp = {op_ret = 1059, op_errno = 0, >>>>>>>>>> >>>>>>>>>> spec = 0x3fff980018a0 "volume c_glusterfs-client-0\n >>>>>>>>>> type protocol/client\n option password >>>>>>>>>> 5fd8d83d-99f3-4630-97c6-965d7a8ead62\n option username >>>>>>>>>> e65687aa-e135-445e-8778-48bb8fb19640\n option transport-type tcp\n >>>>>>>>>> "..., xdata = {xdata_len = 0, xdata_val = 0x0}} >>>>>>>>>> >>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>> >>>>>>>>>> frame = 0x3fffa2bea06c >>>>>>>>>> >>>>>>>>>> ctx = 0x10027010 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> size = 1059 >>>>>>>>>> >>>>>>>>>> tmpfp = 0x3fff98001cd0 >>>>>>>>>> >>>>>>>>>> volfilebuf = 0x0 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = <error reading variable __FUNCTION__ >>>>>>>>>> (Cannot access memory at address 0x10013570)> >>>>>>>>>> >>>>>>>>>> #16 0x00003fffa3f21ec4 in rpc_clnt_handle_reply (clnt=0x10089020, >>>>>>>>>> pollin=0x3fff98001760) at rpc-clnt.c:775 >>>>>>>>>> >>>>>>>>>> conn = 0x10089050 >>>>>>>>>> >>>>>>>>>> saved_frame = <optimized out> >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> req = 0x1008931c >>>>>>>>>> >>>>>>>>>> xid = 1 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "rpc_clnt_handle_reply" >>>>>>>>>> >>>>>>>>>> #17 0x00003fffa3f223d0 in rpc_clnt_notify (trans=<optimized out>, >>>>>>>>>> mydata=0x10089050, event=<optimized out>, data=<optimized out>) at >>>>>>>>>> rpc-clnt.c:933 >>>>>>>>>> >>>>>>>>>> conn = 0x10089050 >>>>>>>>>> >>>>>>>>>> clnt = <optimized out> >>>>>>>>>> >>>>>>>>>> ret = -1 >>>>>>>>>> >>>>>>>>>> req_info = 0x0 >>>>>>>>>> >>>>>>>>>> pollin = <optimized out> >>>>>>>>>> >>>>>>>>>> clnt_mydata = 0x0 >>>>>>>>>> >>>>>>>>>> old_THIS = 0x3fffa402d290 <global_xlator> >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "rpc_clnt_notify" >>>>>>>>>> >>>>>>>>>> #18 0x00003fffa3f1d4fc in rpc_transport_notify (this=<optimized >>>>>>>>>> out>, event=<optimized out>, data=<optimized out>) at rpc-transport.c:546 >>>>>>>>>> >>>>>>>>>> ret = -1 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "rpc_transport_notify" >>>>>>>>>> >>>>>>>>>> #19 0x00003fffa0401d44 in socket_event_poll_in (this=this at entry=0x1008ab80) >>>>>>>>>> at socket.c:2236 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> pollin = 0x3fff98001760 >>>>>>>>>> >>>>>>>>>> priv = 0x1008b820 >>>>>>>>>> >>>>>>>>>> #20 0x00003fffa040489c in socket_event_handler (fd=<optimized >>>>>>>>>> out>, idx=<optimized out>, data=0x1008ab80, poll_in=<optimized out>, >>>>>>>>>> poll_out=<optimized out>, poll_err=<optimized out>) >>>>>>>>>> >>>>>>>>>> at socket.c:2349 >>>>>>>>>> >>>>>>>>>> this = 0x1008ab80 >>>>>>>>>> >>>>>>>>>> priv = 0x1008b820 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "socket_event_handler" >>>>>>>>>> >>>>>>>>>> #21 0x00003fffa3fe2874 in event_dispatch_epoll_handler >>>>>>>>>> (event=0x3fffa034d6a0, event_pool=0x10045bc0) at event-epoll.c:575 >>>>>>>>>> >>>>>>>>>> handler = @0x3fffa041f620: 0x3fffa04046f0 >>>>>>>>>> <socket_event_handler> >>>>>>>>>> >>>>>>>>>> gen = 1 >>>>>>>>>> >>>>>>>>>> slot = 0x1007cd80 >>>>>>>>>> >>>>>>>>>> data = <optimized out> >>>>>>>>>> >>>>>>>>>> ret = -1 >>>>>>>>>> >>>>>>>>>> fd = 9 >>>>>>>>>> >>>>>>>>>> ev_data = 0x3fffa034d6a8 >>>>>>>>>> >>>>>>>>>> idx = 1 >>>>>>>>>> >>>>>>>>>> #22 event_dispatch_epoll_worker (data=0x1008bd50) at >>>>>>>>>> event-epoll.c:678 >>>>>>>>>> >>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>> >>>>>>>>>> event = {events = 1, data = {ptr = 0x100000001, fd = 1, >>>>>>>>>> u32 = 1, u64 = 4294967297}} >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> ev_data = 0x1008bd50 >>>>>>>>>> >>>>>>>>>> event_pool = 0x10045bc0 >>>>>>>>>> >>>>>>>>>> myindex = <optimized out> >>>>>>>>>> >>>>>>>>>> timetodie = 0 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "event_dispatch_epoll_worker" >>>>>>>>>> >>>>>>>>>> #23 0x00003fffa3e9bb30 in start_thread (arg=0x3fffa034e160) at >>>>>>>>>> pthread_create.c:462 >>>>>>>>>> >>>>>>>>>> pd = 0x3fffa034e160 >>>>>>>>>> >>>>>>>>>> now = <optimized out> >>>>>>>>>> >>>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>>> {2380233324717718430, 70367199403008, 2380233324703897146, 0, 0, >>>>>>>>>> 70367128645632, 70367137030688, 8388608, 70367199363104, 269008208, >>>>>>>>>> >>>>>>>>>> 70368094386592, 70367199388632, 70367200825640, >>>>>>>>>> 3, 0, 70367199388648, 70368094386240, 70368094386296, 4001536, >>>>>>>>>> 70367199364120, 70367137027904, -3187653596, >>>>>>>>>> >>>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>>> canceltype = 0}}} >>>>>>>>>> >>>>>>>>>> not_first_call = <optimized out> >>>>>>>>>> >>>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>>> >>>>>>>>>> sp = <optimized out> >>>>>>>>>> >>>>>>>>>> freesize = <optimized out> >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>>> >>>>>>>>>> #24 0x00003fffa3de60fc in .__clone () at >>>>>>>>>> ../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96 >>>>>>>>>> >>>>>>>>>> No locals. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thread 6 (LWP 1735): >>>>>>>>>> >>>>>>>>>> #0 0x00003fffa3ea3ccc in __pthread_cond_timedwait >>>>>>>>>> (cond=0x10059a98, mutex=0x10059a70, abstime=0x3fffa141f670) at >>>>>>>>>> pthread_cond_timedwait.c:198 >>>>>>>>>> >>>>>>>>>> r4 = 393 >>>>>>>>>> >>>>>>>>>> r7 = 0 >>>>>>>>>> >>>>>>>>>> arg5 = 0 >>>>>>>>>> >>>>>>>>>> arg2 = <optimized out> >>>>>>>>>> >>>>>>>>>> r5 = 2 >>>>>>>>>> >>>>>>>>>> r8 = 4294967295 >>>>>>>>>> >>>>>>>>>> arg6 = 4294967295 >>>>>>>>>> >>>>>>>>>> arg3 = 2 >>>>>>>>>> >>>>>>>>>> r0 = 221 >>>>>>>>>> >>>>>>>>>> r3 = 516 >>>>>>>>>> >>>>>>>>>> r6 = 70367154665072 >>>>>>>>>> >>>>>>>>>> arg4 = 70367154665072 >>>>>>>>>> >>>>>>>>>> arg1 = 268802716 >>>>>>>>>> >>>>>>>>>> __err = <optimized out> >>>>>>>>>> >>>>>>>>>> __ret = <optimized out> >>>>>>>>>> >>>>>>>>>> futex_val = 2 >>>>>>>>>> >>>>>>>>>> buffer = {__routine = @0x3fffa3ec0b50: 0x3fffa3ea3400 >>>>>>>>>> <__condvar_cleanup>, __arg = 0x3fffa141f540, __canceltype = 0, __prev = 0x0} >>>>>>>>>> >>>>>>>>>> cbuffer = {oldtype = 0, cond = 0x10059a98, mutex >>>>>>>>>> 0x10059a70, bc_seq = 0} >>>>>>>>>> >>>>>>>>>> result = 0 >>>>>>>>>> >>>>>>>>>> pshared = 0 >>>>>>>>>> >>>>>>>>>> pi_flag = 0 >>>>>>>>>> >>>>>>>>>> err = <optimized out> >>>>>>>>>> >>>>>>>>>> val = <optimized out> >>>>>>>>>> >>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>> >>>>>>>>>> seq = 0 >>>>>>>>>> >>>>>>>>>> #1 0x00003fffa3fc0e74 in syncenv_task (proc=0x10053eb0) at >>>>>>>>>> syncop.c:607 >>>>>>>>>> >>>>>>>>>> env = 0x10053eb0 >>>>>>>>>> >>>>>>>>>> task = 0x0 >>>>>>>>>> >>>>>>>>>> sleep_till = {tv_sec = 1536845230, tv_nsec = 0} >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> #2 0x00003fffa3fc1cdc in syncenv_processor (thdata=0x10053eb0) >>>>>>>>>> at syncop.c:699 >>>>>>>>>> >>>>>>>>>> env = 0x10053eb0 >>>>>>>>>> >>>>>>>>>> proc = 0x10053eb0 >>>>>>>>>> >>>>>>>>>> task = <optimized out> >>>>>>>>>> >>>>>>>>>> #3 0x00003fffa3e9bb30 in start_thread (arg=0x3fffa1420160) at >>>>>>>>>> pthread_create.c:462 >>>>>>>>>> >>>>>>>>>> pd = 0x3fffa1420160 >>>>>>>>>> >>>>>>>>>> now = <optimized out> >>>>>>>>>> >>>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>>> {2380233324731669406, 70367199403008, 2380233324703897146, 0, 0, >>>>>>>>>> 70367146283008, 70367154668064, 8388608, 70367199363104, 268779184, >>>>>>>>>> >>>>>>>>>> 268779184, 70367199388632, 70367200820192, 3, 0, >>>>>>>>>> 70367199388648, 70368094386080, 70368094386136, 4001536, 70367199364120, >>>>>>>>>> 70367154665280, -3187653564, >>>>>>>>>> >>>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>>> canceltype = 0}}} >>>>>>>>>> >>>>>>>>>> not_first_call = <optimized out> >>>>>>>>>> >>>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>>> >>>>>>>>>> sp = <optimized out> >>>>>>>>>> >>>>>>>>>> freesize = <optimized out> >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>>> >>>>>>>>>> #4 0x00003fffa3de60fc in .__clone () at >>>>>>>>>> ../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96 >>>>>>>>>> >>>>>>>>>> No locals. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thread 5 (LWP 1746): >>>>>>>>>> >>>>>>>>>> #0 0x00003fffa3ea7e38 in __lll_lock_wait (futex=0x100272a8, >>>>>>>>>> private=<optimized out>) at lowlevellock.c:46 >>>>>>>>>> >>>>>>>>>> r4 = 128 >>>>>>>>>> >>>>>>>>>> r7 = 0 >>>>>>>>>> >>>>>>>>>> r5 = 2 >>>>>>>>>> >>>>>>>>>> r8 = 1 >>>>>>>>>> >>>>>>>>>> arg3 = 2 >>>>>>>>>> >>>>>>>>>> r0 = 221 >>>>>>>>>> >>>>>>>>>> r3 = 512 >>>>>>>>>> >>>>>>>>>> r6 = 0 >>>>>>>>>> >>>>>>>>>> arg4 = 0 >>>>>>>>>> >>>>>>>>>> arg1 = 268595880 >>>>>>>>>> >>>>>>>>>> __err = <optimized out> >>>>>>>>>> >>>>>>>>>> __ret = <optimized out> >>>>>>>>>> >>>>>>>>>> #1 0x00003fffa3e9ef64 in __GI___pthread_mutex_lock >>>>>>>>>> (mutex=0x100272a8) at ../nptl/pthread_mutex_lock.c:81 >>>>>>>>>> >>>>>>>>>> __futex = 0x100272a8 >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "__pthread_mutex_lock" >>>>>>>>>> >>>>>>>>>> type = <optimized out> >>>>>>>>>> >>>>>>>>>> id = <optimized out> >>>>>>>>>> >>>>>>>>>> #2 0x00003fffa3f6ce8c in _gf_msg (domain=0x3fffa4009e38 "epoll", >>>>>>>>>> file=0x3fffa4009e28 "event-epoll.c", function=0x3fffa4009db8 >>>>>>>>>> <__FUNCTION__.8510> "event_dispatch_epoll_worker", >>>>>>>>>> >>>>>>>>>> line=<optimized out>, level=<optimized out>, >>>>>>>>>> errnum=<optimized out>, trace=<optimized out>, msgid=101190, >>>>>>>>>> fmt=0x3fffa4009f48 "Started thread with index %d") at logging.c:2058 >>>>>>>>>> >>>>>>>>>> ret = <optimized out> >>>>>>>>>> >>>>>>>>>> ---Type <return> to continue, or q <return> to quit--- >>>>>>>>>> >>>>>>>>>> msgstr = <optimized out> >>>>>>>>>> >>>>>>>>>> ap = <optimized out> >>>>>>>>>> >>>>>>>>>> this = 0x3fffa402d290 <global_xlator> >>>>>>>>>> >>>>>>>>>> ctx = 0x10027010 >>>>>>>>>> >>>>>>>>>> callstr = '\000' <repeats 4095 times> >>>>>>>>>> >>>>>>>>>> passcallstr = 0 >>>>>>>>>> >>>>>>>>>> log_inited = 0 >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "_gf_msg" >>>>>>>>>> >>>>>>>>>> #3 0x00003fffa3fe265c in event_dispatch_epoll_worker >>>>>>>>>> (data=0x3fff9802ffd0) at event-epoll.c:631 >>>>>>>>>> >>>>>>>>>> event = {events = 0, data = {ptr = 0x0, fd = 0, u32 = 0, >>>>>>>>>> u64 = 0}} >>>>>>>>>> >>>>>>>>>> ret = -1 >>>>>>>>>> >>>>>>>>>> ev_data = 0x3fff9802ffd0 >>>>>>>>>> >>>>>>>>>> event_pool = 0x10045bc0 >>>>>>>>>> >>>>>>>>>> myindex = <optimized out> >>>>>>>>>> >>>>>>>>>> timetodie = 0 >>>>>>>>>> >>>>>>>>>> __FUNCTION__ = "event_dispatch_epoll_worker" >>>>>>>>>> >>>>>>>>>> #4 0x00003fffa3e9bb30 in start_thread (arg=0x3fff9f37d160) at >>>>>>>>>> pthread_create.c:462 >>>>>>>>>> >>>>>>>>>> pd = 0x3fff9f37d160 >>>>>>>>>> >>>>>>>>>> now = <optimized out> >>>>>>>>>> >>>>>>>>>> unwind_buf = {cancel_jmp_buf = {{jmp_buf >>>>>>>>>> {2380233325439331230, 70367199403008, 2380233324703897146, 0, 269006880, >>>>>>>>>> 70367112060928, 70367120445984, 8388608, 70367199363104, >>>>>>>>>> >>>>>>>>>> 70366999543760, 70367137025008, 70367199388632, >>>>>>>>>> 70367200825640, 3, 0, 70367199388648, 70367137024656, 70367137024712, >>>>>>>>>> 4001536, 70367199364120, 70367120443200, -3183328188, >>>>>>>>>> >>>>>>>>>> 0 <repeats 42 times>}, mask_was_saved = 0}}, priv >>>>>>>>>> = {pad = {0x0, 0x0, 0x0, 0x0}, data = {prev = 0x0, cleanup = 0x0, >>>>>>>>>> canceltype = 0}}} >>>>>>>>>> >>>>>>>>>> not_first_call = <optimized out> >>>>>>>>>> >>>>>>>>>> pagesize_m1 = <optimized out> >>>>>>>>>> >>>>>>>>>> sp = <optimized out> >>>>>>>>>> >>>>>>>>>> freesize = <optimized out> >>>>>>>>>> >>>>>>>>>> __PRETTY_FUNCTION__ = "start_thread" >>>>>>>>>> <p >>>>>>>>>> class="m_3924698905562214407m_-746409795816116485m_5332120683958399093m_691617268217482248 >>>>>>>>>> >>>>>>>>>-- Pranith -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180925/ea56ac88/attachment.html>