[Gluster-users] [Gluster-devel] Crash in glusterd!!!
ABHISHEK PALIWAL
abhishpaliwal at gmail.com
Wed Dec 6 12:09:12 UTC 2017
I hope these logs were sufficient... please let me know if you require more
logs.
On Wed, Dec 6, 2017 at 3:26 PM, ABHISHEK PALIWAL <abhishpaliwal at gmail.com>
wrote:
> Hi Atin,
>
> Please find the backtrace and logs files attached here.
>
> Also, below are the BT from core.
>
> (gdb) bt
>
> #0 0x00003fff8834b898 in __GI_raise (sig=<optimized out>) at
> ../sysdeps/unix/sysv/linux/raise.c:55
>
> #1 0x00003fff88350fd0 in __GI_abort () at abort.c:89
>
>
>
> [**ALERT: The abort() might not be exactly invoked from the following
> function line.
>
> If the trail function contains multiple abort() calls,
> then you should cross check by other means to get correct abort() call
> location.
>
> This is due to the optimized compilation which hides the
> debug info for multiple abort() calls in a given function.
>
> Refer TR HU16995 for more information]
>
>
>
> #2 0x00003fff8838be04 in __libc_message (do_abort=<optimized out>,
> fmt=<optimized out>) at ../sysdeps/posix/libc_fatal.c:175
>
> #3 0x00003fff8839aba8 in malloc_printerr (action=<optimized out>,
> str=0x3fff8847e498 "double free or corruption (!prev)", ptr=<optimized
> out>, ar_ptr=<optimized out>) at malloc.c:5007
>
> #4 0x00003fff8839ba40 in _int_free (av=0x3fff6c000020, p=<optimized out>,
> have_lock=<optimized out>) at malloc.c:3868
>
> #5 0x00003fff885e0814 in __gf_free (free_ptr=0x3fff6c045da0) at
> mem-pool.c:336
>
> #6 0x00003fff849093c4 in glusterd_friend_sm () at glusterd-sm.c:1295
>
> #7 0x00003fff84901a58 in __glusterd_handle_incoming_unfriend_req
> (req=0x3fff8481c06c) at glusterd-handler.c:2606
>
> #8 0x00003fff848fb870 in glusterd_big_locked_handler (req=0x3fff8481c06c,
> actor_fn=@0x3fff84a43e70: 0x3fff84901830 <__glusterd_handle_incoming_unfriend_req>)
> at glusterd-handler.c:83
>
> #9 0x00003fff848fbd08 in glusterd_handle_incoming_unfriend_req
> (req=<optimized out>) at glusterd-handler.c:2615
>
> #10 0x00003fff8854e87c in rpcsvc_handle_rpc_call (svc=0x10062fd0
> <_GLOBAL__sub_I__ZN27UehChSwitchFachToDchC_ActorC2EP12RTControllerP10RTActorRef()+1148>,
> trans=<optimized out>, msg=0x3fff6c000920) at rpcsvc.c:705
>
> #11 0x00003fff8854eb7c in rpcsvc_notify (trans=0x3fff74002210,
> mydata=<optimized out>, event=<optimized out>, data=<optimized out>) at
> rpcsvc.c:799
>
> #12 0x00003fff885514fc in rpc_transport_notify (this=<optimized out>,
> event=<optimized out>, data=<optimized out>) at rpc-transport.c:546
>
> #13 0x00003fff847fcd44 in socket_event_poll_in (this=this at entry=0x3fff74002210)
> at socket.c:2236
>
> #14 0x00003fff847ff89c in socket_event_handler (fd=<optimized out>,
> idx=<optimized out>, data=0x3fff74002210, poll_in=<optimized out>,
> poll_out=<optimized out>, poll_err=<optimized out>) at socket.c:2349
>
> #15 0x00003fff88616874 in event_dispatch_epoll_handler
> (event=0x3fff83d9d6a0, event_pool=0x10045bc0 <_GLOBAL__sub_I__
> ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>)
> at event-epoll.c:575
>
> #16 event_dispatch_epoll_worker (data=0x100bb4a0
> <main_thread_func__()+1756>) at event-epoll.c:678
>
> #17 0x00003fff884cfb10 in start_thread (arg=0x3fff83d9e160) at
> pthread_create.c:339
>
> #18 0x00003fff88419c0c in .__clone () at ../sysdeps/unix/sysv/linux/
> powerpc/powerpc64/clone.S:96
>
>
>
> (gdb) bt full
>
> #0 0x00003fff8834b898 in __GI_raise (sig=<optimized out>) at
> ../sysdeps/unix/sysv/linux/raise.c:55
>
> r4 = 1560
>
> r7 = 16
>
> arg2 = 1560
>
> r5 = 6
>
> r8 = 0
>
> arg3 = 6
>
> r0 = 250
>
> r3 = 0
>
> r6 = 8
>
> arg1 = 0
>
> sc_err = <optimized out>
>
> sc_ret = <optimized out>
>
> pd = 0x3fff83d9e160
>
> pid = 0
>
> ---Type <return> to continue, or q <return> to quit---
>
> selftid = 1560
>
> #1 0x00003fff88350fd0 in __GI_abort () at abort.c:89
>
> save_stage = 2
>
> act = {__sigaction_handler = {sa_handler = 0x0, sa_sigaction =
> 0x0}, sa_mask = {__val = {0 <repeats 16 times>}}, sa_flags = 0, sa_restorer
> = 0x0}
>
> sigs = {__val = {32, 0 <repeats 15 times>}}
>
>
>
> [**ALERT: The abort() might not be exactly invoked from the following
> function line.
>
> If the trail function contains multiple abort() calls,
> then you should cross check by other means to get correct abort() call
> location.
>
> This is due to the optimized compilation which hides the
> debug info for multiple abort() calls in a given function.
>
> Refer TR HU16995 for more information]
>
>
>
> #2 0x00003fff8838be04 in __libc_message (do_abort=<optimized out>,
> fmt=<optimized out>) at ../sysdeps/posix/libc_fatal.c:175
>
> ap = <optimized out>
>
> fd = <optimized out>
>
> on_2 = <optimized out>
>
> list = <optimized out>
>
> nlist = <optimized out>
>
> cp = <optimized out>
>
> written = <optimized out>
>
> #3 0x00003fff8839aba8 in malloc_printerr (action=<optimized out>,
> str=0x3fff8847e498 "*double free or corruption (!prev)*", ptr=<optimized
> out>, ar_ptr=<optimized out>) at malloc.c:5007
>
> buf = "00003fff6c045d60"
>
> cp = <optimized out>
>
> ar_ptr = <optimized out>
>
> ptr = <optimized out>
>
> str = 0x3fff8847e498 "double free or corruption (!prev)"
>
> action = 3
>
> #4 0x00003fff8839ba40 in _int_free (av=0x3fff6c000020, p=<optimized out>,
> have_lock=<optimized out>) at malloc.c:3868
>
> size = <optimized out>
>
> fb = <optimized out>
>
> nextchunk = <optimized out>
>
> nextsize = <optimized out>
>
> nextinuse = <optimized out>
>
> prevsize = <optimized out>
>
> bck = <optimized out>
>
> fwd = <optimized out>
>
> errstr = <optimized out>
>
> locked = <optimized out>
>
> __func__ = "_int_free"
>
> #5 0x00003fff885e0814 in __gf_free (free_ptr=0x3fff6c045da0) at
> mem-pool.c:336
>
> ptr = 0x3fff6c045d60
>
> mem_acct = <optimized out>
>
> header = 0x3fff6c045d60
>
> free_ptr = 0x3fff6c045da0
>
> #6 0x00003fff849093c4 in glusterd_friend_sm () at glusterd-sm.c:1295
>
> event = 0x3fff6c045da0
>
> tmp = 0x3fff6c045da0
>
> ret = <optimized out>
>
> ---Type <return> to continue, or q <return> to quit---
>
> handler = @0x3fff84a44038: 0x3fff84906750
> <glusterd_ac_friend_remove>
>
> state = 0x3fff84a390c0 <glusterd_state_befriended>
>
> peerinfo = <optimized out>
>
> event_type = GD_FRIEND_EVENT_REMOVE_FRIEND
>
> is_await_conn = <optimized out>
>
> quorum_action = <optimized out>
>
> old_state = GD_FRIEND_STATE_BEFRIENDED
>
> this = <optimized out>
>
> priv = 0x3fff84748050
>
> __FUNCTION__ = "glusterd_friend_sm"
>
> #7 0x00003fff84901a58 in __glusterd_handle_incoming_unfriend_req
> (req=0x3fff8481c06c) at glusterd-handler.c:2606
>
> ret = 0
>
> friend_req = {uuid = "\231\214R¦\177\223I\216\236Õ\214dÎöy¡",
> hostname = 0x3fff6c028ef0 "", port = 0, vols = {vols_len = 0, vols_val =
> 0x0}}
>
> remote_hostname = "10.32.0.48", '\000' <repeats 98 times>
>
> __FUNCTION__ = "__glusterd_handle_incoming_unfriend_req"
>
> #8 0x00003fff848fb870 in glusterd_big_locked_handler (req=0x3fff8481c06c,
> actor_fn=@0x3fff84a43e70: 0x3fff84901830 <__glusterd_handle_incoming_unfriend_req>)
> at glusterd-handler.c:83
>
> priv = 0x3fff84748050
>
> ret = -1
>
> #9 0x00003fff848fbd08 in glusterd_handle_incoming_unfriend_req
> (req=<optimized out>) at glusterd-handler.c:2615
>
> No locals.
>
> #10 0x00003fff8854e87c in rpcsvc_handle_rpc_call (svc=0x10062fd0
> <_GLOBAL__sub_I__ZN27UehChSwitchFachToDchC_ActorC2EP12RTControllerP10RTActorRef()+1148>,
> trans=<optimized out>, msg=0x3fff6c000920) at rpcsvc.c:705
>
> actor = 0x3fff84a38860 <gd_svc_peer_actors+192>
>
> actor_fn = @0x3fff84a43ab0: 0x3fff848fbcf0
> <glusterd_handle_incoming_unfriend_req>
>
> req = 0x3fff8481c06c
>
> ret = -1
>
> port = <optimized out>
>
> unprivileged = <optimized out>
>
> reply = <optimized out>
>
> drc = <optimized out>
>
> __FUNCTION__ = "rpcsvc_handle_rpc_call"
>
> #11 0x00003fff8854eb7c in rpcsvc_notify (trans=0x3fff74002210,
> mydata=<optimized out>, event=<optimized out>, data=<optimized out>) at
> rpcsvc.c:799
>
> ret = -1
>
> msg = <optimized out>
>
> new_trans = 0x0
>
> svc = <optimized out>
>
> listener = 0x0
>
> __FUNCTION__ = "rpcsvc_notify"
>
> #12 0x00003fff885514fc in rpc_transport_notify (this=<optimized out>,
> event=<optimized out>, data=<optimized out>) at rpc-transport.c:546
>
> ret = -1
>
> __FUNCTION__ = "rpc_transport_notify"
>
> #13 0x00003fff847fcd44 in socket_event_poll_in (this=this at entry=0x3fff74002210)
> at socket.c:2236
>
> ret = <optimized out>
>
> pollin = 0x3fff6c000920
>
> priv = 0x3fff74002d50
>
> #14 0x00003fff847ff89c in socket_event_handler (fd=<optimized out>,
> idx=<optimized out>, data=0x3fff74002210, poll_in=<optimized out>,
> poll_out=<optimized out>, poll_err=<optimized out>) at socket.c:2349
>
> ---Type <return> to continue, or q <return> to quit---
>
> this = 0x3fff74002210
>
> priv = 0x3fff74002d50
>
> ret = <optimized out>
>
> __FUNCTION__ = "socket_event_handler"
>
> #15 0x00003fff88616874 in event_dispatch_epoll_handler
> (event=0x3fff83d9d6a0, event_pool=0x10045bc0 <_GLOBAL__sub_I__
> ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>)
> at event-epoll.c:575
>
> handler = @0x3fff8481a620: 0x3fff847ff6f0 <socket_event_handler>
>
> gen = 1
>
> slot = 0x100803f0 <_GLOBAL__sub_I__ZN24RoamIfFroRrcRoExtAttribDC2
> Ev()+232>
>
> data = <optimized out>
>
> ret = -1
>
> fd = 8
>
> ev_data = 0x3fff83d9d6a8
>
> idx = 7
>
> #16 event_dispatch_epoll_worker (data=0x100bb4a0
> <main_thread_func__()+1756>) at event-epoll.c:678
>
> event = {events = 1, data = {ptr = 0x700000001, fd = 7, u32 = 7,
> u64 = 30064771073}}
>
> ret = <optimized out>
>
> ev_data = 0x100bb4a0 <main_thread_func__()+1756>
>
> event_pool = 0x10045bc0 <_GLOBAL__sub_I__
> ZN29DrhIfRhControlPdrProxyC_ActorC2EP12RTControllerP10RTActorRef()+116>
>
> myindex = <optimized out>
>
> timetodie = 0
>
> __FUNCTION__ = "event_dispatch_epoll_worker"
>
> #17 0x00003fff884cfb10 in start_thread (arg=0x3fff83d9e160) at
> pthread_create.c:339
>
> pd = 0x3fff83d9e160
>
> now = <optimized out>
>
> unwind_buf = {cancel_jmp_buf = {{jmp_buf = {-6868946778599096053,
> 70366736145408, -6868946778421678961, 0, 0, 70366652919808, 70366661304864,
> 8388608, 70366736105504, 269202592, 70367897957648, 70366736131032,
> 70366737568040, 3, 0, 70366736131048, 70367897957296, 70367897957352,
> 4001536, 70366736106520, 70366661302080, -3187654076, 0 <repeats 42
> times>}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x0, 0x0}, data =
> {prev = 0x0, cleanup = 0x0, canceltype = 0}}}
>
> not_first_call = <optimized out>
>
> pagesize_m1 = <optimized out>
>
> sp = <optimized out>
>
> freesize = <optimized out>
>
> __PRETTY_FUNCTION__ = "start_thread"
>
> #18 0x00003fff88419c0c in .__clone () at ../sysdeps/unix/sysv/linux/
> powerpc/powerpc64/clone.S:96
>
> No locals.
>
> Regards,
> Abhishek
>
> On Wed, Dec 6, 2017 at 3:21 PM, Atin Mukherjee <amukherj at redhat.com>
> wrote:
>
>> Without the glusterd log file and the core file or the backtrace I can't
>> comment anything.
>>
>> On Wed, Dec 6, 2017 at 3:09 PM, ABHISHEK PALIWAL <abhishpaliwal at gmail.com
>> > wrote:
>>
>>> Any suggestion....
>>>
>>> On Dec 6, 2017 11:51, "ABHISHEK PALIWAL" <abhishpaliwal at gmail.com>
>>> wrote:
>>>
>>>> Hi Team,
>>>>
>>>> We are getting the crash in glusterd after start of it. When I tried to
>>>> debug in brick logs we are getting below errors:
>>>>
>>>> [2017-12-01 14:10:14.684122] E [MSGID: 100018]
>>>> [glusterfsd.c:1960:glusterfs_pidfile_update] 0-glusterfsd: pidfile
>>>> /system/glusterd/vols/c_glusterfs/run/10.32.1.144-opt-lvmdir-c2-brick.pid
>>>> lock failed [Resource temporarily unavailable]
>>>> :
>>>> :
>>>> :
>>>> [2017-12-01 14:10:16.862903] E [MSGID: 113001]
>>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=18:
>>>> key:trusted.bit-rot.version [No space left on device]
>>>> [2017-12-01 14:10:16.862985] I [MSGID: 115063]
>>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>>> 92: FTRUNCATE 1 (934f08b7-e3b5-4690-84fc-742a4b1fb78b)==> (No space
>>>> left on device) [No space left on device]
>>>> [2017-12-01 14:10:16.907037] E [MSGID: 113001]
>>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>>> key:trusted.bit-rot.version [No space left on device]
>>>> [2017-12-01 14:10:16.907108] I [MSGID: 115063]
>>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>>> 35: FTRUNCATE 0 (109d6537-a1ec-4556-8ce1-04c365c451eb)==> (No space
>>>> left on device) [No space left on device]
>>>> [2017-12-01 14:10:16.947541] E [MSGID: 113001]
>>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>>> key:trusted.bit-rot.version [No space left on device]
>>>> [2017-12-01 14:10:16.947623] I [MSGID: 115063]
>>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>>> 70: FTRUNCATE 0 (8f9c8054-b0d7-4b93-a95b-cd3ab249c56d)==> (No space
>>>> left on device) [No space left on device]
>>>> [2017-12-01 14:10:16.968515] E [MSGID: 113001]
>>>> [posix.c:4616:_posix_remove_xattr] 0-c_glusterfs-posix: removexattr
>>>> failed on /opt/lvmdir/c2/brick/.glusterfs/00/00/00000000-0000-0000-0000-000000000001/configuration
>>>> (for trusted.glusterfs.dht) [No space left on device]
>>>> [2017-12-01 14:10:16.968589] I [MSGID: 115058]
>>>> [server-rpc-fops.c:740:server_removexattr_cbk] 0-c_glusterfs-server:
>>>> 90: REMOVEXATTR <gfid:a240d2fd-869c-408d-9b95-62ee1bff074e>
>>>> (a240d2fd-869c-408d-9b95-62ee1bff074e) of key ==> (No space left on
>>>> device) [No space left on device]
>>>> [2017-12-01 14:10:17.039815] E [MSGID: 113001]
>>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>>> key:trusted.bit-rot.version [No space left on device]
>>>> [2017-12-01 14:10:17.039900] I [MSGID: 115063]
>>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>>> 152: FTRUNCATE 0 (d67bcfcd-ff19-4b58-9823-46d6cce9ace3)==> (No space
>>>> left on device) [No space left on device]
>>>> [2017-12-01 14:10:17.048767] E [MSGID: 113001]
>>>> [posix-helpers.c:1228:posix_fhandle_pair] 0-c_glusterfs-posix: fd=17:
>>>> key:trusted.bit-rot.version [No space left on device]
>>>> [2017-12-01 14:10:17.048874] I [MSGID: 115063]
>>>> [server-rpc-fops.c:1317:server_ftruncate_cbk] 0-c_glusterfs-server:
>>>> 163: FTRUNCATE 0 (0e3ee6ad-408b-4fcf-a1a7-4262ec113316)==> (No space
>>>> left on device) [No space left on device]
>>>> [2017-12-01 14:10:17.075007] E [MSGID: 113001]
>>>> [posix.c:4616:_posix_remove_xattr] 0-c_glusterfs-posix: removexattr
>>>> failed on /opt/lvmdir/c2/brick/.glusterfs/00/00/00000000-0000-0000-0000-000000000001/java
>>>> (for trusted.glusterfs.dht) [No space left on device]
>>>>
>>>> Also, we are having the lack disk space.
>>>>
>>>> Could any one please explain me what glusterd is doing in brick so that
>>>> it is causing of its crash.
>>>>
>>>> Please find the brick logs in attachment.
>>>>
>>>> Thanks in advance!!!
>>>> --
>>>> Regards
>>>> Abhishek Paliwal
>>>>
>>>
>>> _______________________________________________
>>> Gluster-devel mailing list
>>> Gluster-devel at gluster.org
>>> http://lists.gluster.org/mailman/listinfo/gluster-devel
>>>
>>
>>
>
>
> --
>
>
>
>
> Regards
> Abhishek Paliwal
>
--
Regards
Abhishek Paliwal
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20171206/1b58442b/attachment.html>
More information about the Gluster-users
mailing list