[Gluster-users] [Gluster-devel] Crash in glusterd!!!
    ABHISHEK PALIWAL 
    abhishpaliwal at gmail.com
       
    Wed Dec  6 09:56:24 UTC 2017
    
    
  
Hi Atin,
Please find the backtrace and logs files attached here.
Also, below are the BT from core.
(gdb) bt
#0  0x00003fff8834b898 in __GI_raise (sig=<optimized out>) at
../sysdeps/unix/sysv/linux/raise.c:55
#1  0x00003fff88350fd0 in __GI_abort () at abort.c:89
[**ALERT: The abort() might not be exactly invoked from the following
function line.
                If the trail function contains multiple abort() calls, then
you should cross check by other means to get correct abort() call location.
                This is due to the optimized compilation which hides the
debug info for multiple abort() calls in a given function.
                Refer TR HU16995 for more information]
#2  0x00003fff8838be04 in __libc_message (do_abort=<optimized out>,
fmt=<optimized out>) at ../sysdeps/posix/libc_fatal.c:175
#3  0x00003fff8839aba8 in malloc_printerr (action=<optimized out>,
str=0x3fff8847e498 "double free or corruption (!prev)", ptr=<optimized
out>, ar_ptr=<optimized out>) at malloc.c:5007
#4  0x00003fff8839ba40 in _int_free (av=0x3fff6c000020, p=<optimized out>,
have_lock=<optimized out>) at malloc.c:3868
#5  0x00003fff885e0814 in __gf_free (free_ptr=0x3fff6c045da0) at
mem-pool.c:336
#6  0x00003fff849093c4 in glusterd_friend_sm () at glusterd-sm.c:1295
#7  0x00003fff84901a58 in __glusterd_handle_incoming_unfriend_req
(req=0x3fff8481c06c) at glusterd-handler.c:2606
#8  0x00003fff848fb870 in glusterd_big_locked_handler (req=0x3fff8481c06c,
actor_fn=@0x3fff84a43e70: 0x3fff84901830
<__glusterd_handle_incoming_unfriend_req>) at glusterd-handler.c:83
#9  0x00003fff848fbd08 in glusterd_handle_incoming_unfriend_req
(req=<optimized out>) at glusterd-handler.c:2615
#10 0x00003fff8854e87c in rpcsvc_handle_rpc_call (svc=0x10062fd0
<_GLOBAL__sub_I__ZN27UehChSwitchFachToDchC_ActorC2EP12RTControllerP10RTActorRef()+1148>,
trans=<optimized out>, msg=0x3fff6c000920) at rpcsvc.c:705
#11 0x00003fff8854eb7c in rpcsvc_notify (trans=0x3fff74002210,
mydata=<optimized out>, event=<optimized out>, data=<optimized out>) at
rpcsvc.c:799
#12 0x00003fff885514fc in rpc_transport_notify (this=<optimized out>,
event=<optimized out>, data=<optimized out>) at rpc-transport.c:546
#13 0x00003fff847fcd44 in socket_event_poll_in (this=this at entry=0x3fff74002210)
at socket.c:2236
#14 0x00003fff847ff89c in socket_event_handler (fd=<optimized out>,
idx=<optimized out>, data=0x3fff74002210, poll_in=<optimized out>,
poll_out=<optimized out>, poll_err=<optimized out>) at socket.c:2349
#15 0x00003fff88616874 in event_dispatch_epoll_handler
(event=0x3fff83d9d6a0, event_pool=0x10045bc0
<_GLOBAL__sub_I__ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>)
at event-epoll.c:575
#16 event_dispatch_epoll_worker (data=0x100bb4a0
<main_thread_func__()+1756>) at event-epoll.c:678
#17 0x00003fff884cfb10 in start_thread (arg=0x3fff83d9e160) at
pthread_create.c:339
#18 0x00003fff88419c0c in .__clone () at
../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96
(gdb) bt full
#0  0x00003fff8834b898 in __GI_raise (sig=<optimized out>) at
../sysdeps/unix/sysv/linux/raise.c:55
        r4 = 1560
        r7 = 16
        arg2 = 1560
        r5 = 6
        r8 = 0
        arg3 = 6
        r0 = 250
        r3 = 0
        r6 = 8
        arg1 = 0
        sc_err = <optimized out>
        sc_ret = <optimized out>
        pd = 0x3fff83d9e160
        pid = 0
---Type <return> to continue, or q <return> to quit---
        selftid = 1560
#1  0x00003fff88350fd0 in __GI_abort () at abort.c:89
        save_stage = 2
        act = {__sigaction_handler = {sa_handler = 0x0, sa_sigaction =
0x0}, sa_mask = {__val = {0 <repeats 16 times>}}, sa_flags = 0, sa_restorer
= 0x0}
        sigs = {__val = {32, 0 <repeats 15 times>}}
[**ALERT: The abort() might not be exactly invoked from the following
function line.
                If the trail function contains multiple abort() calls, then
you should cross check by other means to get correct abort() call location.
                This is due to the optimized compilation which hides the
debug info for multiple abort() calls in a given function.
                Refer TR HU16995 for more information]
#2  0x00003fff8838be04 in __libc_message (do_abort=<optimized out>,
fmt=<optimized out>) at ../sysdeps/posix/libc_fatal.c:175
        ap = <optimized out>
        fd = <optimized out>
        on_2 = <optimized out>
        list = <optimized out>
        nlist = <optimized out>
        cp = <optimized out>
        written = <optimized out>
#3  0x00003fff8839aba8 in malloc_printerr (action=<optimized out>,
str=0x3fff8847e498 "*double free or corruption (!prev)*", ptr=<optimized
out>, ar_ptr=<optimized out>) at malloc.c:5007
        buf = "00003fff6c045d60"
        cp = <optimized out>
        ar_ptr = <optimized out>
        ptr = <optimized out>
        str = 0x3fff8847e498 "double free or corruption (!prev)"
        action = 3
#4  0x00003fff8839ba40 in _int_free (av=0x3fff6c000020, p=<optimized out>,
have_lock=<optimized out>) at malloc.c:3868
        size = <optimized out>
        fb = <optimized out>
        nextchunk = <optimized out>
        nextsize = <optimized out>
        nextinuse = <optimized out>
        prevsize = <optimized out>
        bck = <optimized out>
        fwd = <optimized out>
        errstr = <optimized out>
        locked = <optimized out>
        __func__ = "_int_free"
#5  0x00003fff885e0814 in __gf_free (free_ptr=0x3fff6c045da0) at
mem-pool.c:336
        ptr = 0x3fff6c045d60
        mem_acct = <optimized out>
        header = 0x3fff6c045d60
        free_ptr = 0x3fff6c045da0
#6  0x00003fff849093c4 in glusterd_friend_sm () at glusterd-sm.c:1295
        event = 0x3fff6c045da0
        tmp = 0x3fff6c045da0
        ret = <optimized out>
---Type <return> to continue, or q <return> to quit---
        handler = @0x3fff84a44038: 0x3fff84906750
<glusterd_ac_friend_remove>
        state = 0x3fff84a390c0 <glusterd_state_befriended>
        peerinfo = <optimized out>
        event_type = GD_FRIEND_EVENT_REMOVE_FRIEND
        is_await_conn = <optimized out>
        quorum_action = <optimized out>
        old_state = GD_FRIEND_STATE_BEFRIENDED
        this = <optimized out>
        priv = 0x3fff84748050
        __FUNCTION__ = "glusterd_friend_sm"
#7  0x00003fff84901a58 in __glusterd_handle_incoming_unfriend_req
(req=0x3fff8481c06c) at glusterd-handler.c:2606
        ret = 0
        friend_req = {uuid = "\231\214R¦\177\223I\216\236Õ\214dÎöy¡",
hostname = 0x3fff6c028ef0 "", port = 0, vols = {vols_len = 0, vols_val =
0x0}}
        remote_hostname = "10.32.0.48", '\000' <repeats 98 times>
        __FUNCTION__ = "__glusterd_handle_incoming_unfriend_req"
#8  0x00003fff848fb870 in glusterd_big_locked_handler (req=0x3fff8481c06c,
actor_fn=@0x3fff84a43e70: 0x3fff84901830
<__glusterd_handle_incoming_unfriend_req>) at glusterd-handler.c:83
        priv = 0x3fff84748050
        ret = -1
#9  0x00003fff848fbd08 in glusterd_handle_incoming_unfriend_req
(req=<optimized out>) at glusterd-handler.c:2615
No locals.
#10 0x00003fff8854e87c in rpcsvc_handle_rpc_call (svc=0x10062fd0
<_GLOBAL__sub_I__ZN27UehChSwitchFachToDchC_ActorC2EP12RTControllerP10RTActorRef()+1148>,
trans=<optimized out>, msg=0x3fff6c000920) at rpcsvc.c:705
        actor = 0x3fff84a38860 <gd_svc_peer_actors+192>
        actor_fn = @0x3fff84a43ab0: 0x3fff848fbcf0
<glusterd_handle_incoming_unfriend_req>
        req = 0x3fff8481c06c
        ret = -1
        port = <optimized out>
        unprivileged = <optimized out>
        reply = <optimized out>
        drc = <optimized out>
        __FUNCTION__ = "rpcsvc_handle_rpc_call"
#11 0x00003fff8854eb7c in rpcsvc_notify (trans=0x3fff74002210,
mydata=<optimized out>, event=<optimized out>, data=<optimized out>) at
rpcsvc.c:799
        ret = -1
        msg = <optimized out>
        new_trans = 0x0
        svc = <optimized out>
        listener = 0x0
        __FUNCTION__ = "rpcsvc_notify"
#12 0x00003fff885514fc in rpc_transport_notify (this=<optimized out>,
event=<optimized out>, data=<optimized out>) at rpc-transport.c:546
        ret = -1
        __FUNCTION__ = "rpc_transport_notify"
#13 0x00003fff847fcd44 in socket_event_poll_in (this=this at entry=0x3fff74002210)
at socket.c:2236
        ret = <optimized out>
        pollin = 0x3fff6c000920
        priv = 0x3fff74002d50
#14 0x00003fff847ff89c in socket_event_handler (fd=<optimized out>,
idx=<optimized out>, data=0x3fff74002210, poll_in=<optimized out>,
poll_out=<optimized out>, poll_err=<optimized out>) at socket.c:2349
---Type <return> to continue, or q <return> to quit---
        this = 0x3fff74002210
        priv = 0x3fff74002d50
        ret = <optimized out>
        __FUNCTION__ = "socket_event_handler"
#15 0x00003fff88616874 in event_dispatch_epoll_handler
(event=0x3fff83d9d6a0, event_pool=0x10045bc0
<_GLOBAL__sub_I__ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>)
at event-epoll.c:575
        handler = @0x3fff8481a620: 0x3fff847ff6f0 <socket_event_handler>
        gen = 1
        slot = 0x100803f0
<_GLOBAL__sub_I__ZN24RoamIfFroRrcRoExtAttribDC2Ev()+232>
        data = <optimized out>
        ret = -1
        fd = 8
        ev_data = 0x3fff83d9d6a8
        idx = 7
#16 event_dispatch_epoll_worker (data=0x100bb4a0
<main_thread_func__()+1756>) at event-epoll.c:678
        event = {events = 1, data = {ptr = 0x700000001, fd = 7, u32 = 7,
u64 = 30064771073}}
        ret = <optimized out>
        ev_data = 0x100bb4a0 <main_thread_func__()+1756>
        event_pool = 0x10045bc0
<_GLOBAL__sub_I__ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>
        myindex = <optimized out>
        timetodie = 0
        __FUNCTION__ = "event_dispatch_epoll_worker"
#17 0x00003fff884cfb10 in start_thread (arg=0x3fff83d9e160) at
pthread_create.c:339
        pd = 0x3fff83d9e160
        now = <optimized out>
        unwind_buf = {cancel_jmp_buf = {{jmp_buf = {-6868946778599096053,
70366736145408, -6868946778421678961, 0, 0, 70366652919808, 70366661304864,
8388608, 70366736105504, 269202592, 70367897957648, 70366736131032,
70366737568040, 3, 0, 70366736131048, 70367897957296, 70367897957352,
4001536, 70366736106520, 70366661302080, -3187654076, 0 <repeats 42
times>}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x0, 0x0}, data =
{prev = 0x0, cleanup = 0x0, canceltype = 0}}}
        not_first_call = <optimized out>
        pagesize_m1 = <optimized out>
        sp = <optimized out>
        freesize = <optimized out>
        __PRETTY_FUNCTION__ = "start_thread"
#18 0x00003fff88419c0c in .__clone () at
../sysdeps/unix/sysv/linux/powerpc/powerpc64/clone.S:96
No locals.
Regards,
Abhishek
On Wed, Dec 6, 2017 at 3:21 PM, Atin Mukherjee <amukherj at redhat.com> wrote:
> Without the glusterd log file and the core file or the backtrace I can't
> comment anything.
>
> On Wed, Dec 6, 2017 at 3:09 PM, ABHISHEK PALIWAL <abhishpaliwal at gmail.com>
> wrote:
>
>> Any suggestion....
>>
>> On Dec 6, 2017 11:51, "ABHISHEK PALIWAL" <abhishpaliwal at gmail.com> wrote:
>>
>>> Hi Team,
>>>
>>> We are getting the crash in glusterd after start of it. When I tried to
>>> debug in brick logs we are getting below errors:
>>>
>>> [2017-12-01 14:10:14.684122] E [MSGID: 100018]
>>> [glusterfsd.c:1960:glusterfs_pidfile_update] 0-glusterfsd: pidfile
>>> /system/glusterd/vols/c_glusterfs/run/10.32.1.144-opt-lvmdir-c2-brick.pid
>>> lock failed [Resource temporarily unavailable]
>>> :
>>> :
>>> :
>>> [2017-12-01 14:10:16.862903] E [MSGID: 113001]
>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=18:
>>> key:trusted.bit-rot.version [No space left on device]
>>> [2017-12-01 14:10:16.862985] I [MSGID: 115063]
>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server: 92:
>>> FTRUNCATE 1 (934f08b7-e3b5-4690-84fc-742a4b1fb78b)==> (No space left on
>>> device) [No space left on device]
>>> [2017-12-01 14:10:16.907037] E [MSGID: 113001]
>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>> key:trusted.bit-rot.version [No space left on device]
>>> [2017-12-01 14:10:16.907108] I [MSGID: 115063]
>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server: 35:
>>> FTRUNCATE 0 (109d6537-a1ec-4556-8ce1-04c365c451eb)==> (No space left on
>>> device) [No space left on device]
>>> [2017-12-01 14:10:16.947541] E [MSGID: 113001]
>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>> key:trusted.bit-rot.version [No space left on device]
>>> [2017-12-01 14:10:16.947623] I [MSGID: 115063]
>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server: 70:
>>> FTRUNCATE 0 (8f9c8054-b0d7-4b93-a95b-cd3ab249c56d)==> (No space left on
>>> device) [No space left on device]
>>> [2017-12-01 14:10:16.968515] E [MSGID: 113001]
>>> [posix.c:4616:_posix_remove_xattr] 0-c_glusterfs-posix: removexattr
>>> failed on /opt/lvmdir/c2/brick/.glusterfs/00/00/00000000-0000-0000-0000-000000000001/configuration
>>> (for trusted.glusterfs.dht) [No space left on device]
>>> [2017-12-01 14:10:16.968589] I [MSGID: 115058]
>>> [server-rpc-fops.c:740:server_removexattr_cbk] 0-c_glusterfs-server:
>>> 90: REMOVEXATTR <gfid:a240d2fd-869c-408d-9b95-62ee1bff074e>
>>> (a240d2fd-869c-408d-9b95-62ee1bff074e) of key  ==> (No space left on
>>> device) [No space left on device]
>>> [2017-12-01 14:10:17.039815] E [MSGID: 113001]
>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>> key:trusted.bit-rot.version [No space left on device]
>>> [2017-12-01 14:10:17.039900] I [MSGID: 115063]
>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>> 152: FTRUNCATE 0 (d67bcfcd-ff19-4b58-9823-46d6cce9ace3)==> (No space
>>> left on device) [No space left on device]
>>> [2017-12-01 14:10:17.048767] E [MSGID: 113001]
>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>> key:trusted.bit-rot.version [No space left on device]
>>> [2017-12-01 14:10:17.048874] I [MSGID: 115063]
>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>> 163: FTRUNCATE 0 (0e3ee6ad-408b-4fcf-a1a7-4262ec113316)==> (No space
>>> left on device) [No space left on device]
>>> [2017-12-01 14:10:17.075007] E [MSGID: 113001]
>>> [posix.c:4616:_posix_remove_xattr] 0-c_glusterfs-posix: removexattr
>>> failed on /opt/lvmdir/c2/brick/.glusterfs/00/00/00000000-0000-0000-0000-000000000001/java
>>> (for trusted.glusterfs.dht) [No space left on device]
>>>
>>> Also, we are having the lack disk space.
>>>
>>> Could any one please explain me what glusterd is doing in brick so that
>>> it is causing of its crash.
>>>
>>> Please find the brick logs in attachment.
>>>
>>> Thanks in advance!!!
>>> --
>>> Regards
>>> Abhishek Paliwal
>>>
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-devel
>>
>
>
-- 
Regards
Abhishek Paliwal
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20171206/15a9aed6/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: glusterfs.7z
Type: application/x-7z-compressed
Size: 148198 bytes
Desc: not available
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20171206/15a9aed6/attachment-0001.bin>
    
    
More information about the Gluster-users
mailing list