[Gluster-users] glustershd coredump generated while reboot all 3 sn nodes

Ravishankar N ravishankar at redhat.com
Tue Oct 16 04:23:23 UTC 2018


Hi,

- Is this stock glusterfs-3.12.3? Or do you have any patches applied on 
top of it?

- If it is stock, could you create a BZ and attach the core file and the 
/var/log/glusterfs/ logs from 3 nodes at the time of crash?

Thanks,
Ravi


On 10/16/2018 08:45 AM, Zhou, Cynthia (NSB - CN/Hangzhou) wrote:
>
> Hi,
>
> This issue happened twice recently, when glustershd do heal, it 
> generate coredump occassinally,
>
> I do some debug and find that sometimes 
> afr_selfheal_unlocked_discover_on do lookup and saved the frame in 
> function rpc_clnt_submit, when reply comes, it find the saved frame , 
> but the address is different from the saved frame address, I think 
> this is wrong, but I can not find a clue how this happened?
>
> [root at mn-0:/home/robot]
>
> [Thread debugging using libthread_db enabled]
>
> Using host libthread_db library "/lib64/libthread_db.so.1".
>
> Core was generated by `/usr/sbin/glusterfs -s sn-0.local --volfile-id 
> gluster/glustershd -p /var/run/g'.
>
> Program terminated with signal SIGSEGV, Segmentation fault.
>
> #0  0x00007fb1a6fd9d24 in client3_3_lookup_cbk (req=0x7fb188010fb0, 
> iov=0x7fb188010ff0, count=1, myframe=*0x7fb188215740*) at 
> client-rpc-fops.c:2802
>
> 2802 client-rpc-fops.c: No such file or directory.
>
> [Current thread is 1 (Thread 0x7fb1a7a0e700 (LWP 8151))]
>
> Missing separate debuginfos, use: dnf debuginfo-install 
> rcp-pack-glusterfs-1.2.0-RCP2.wf29.x86_64
>
> (gdb) bt
>
> #0  0x00007fb1a6fd9d24 in client3_3_lookup_cbk (req=0x7fb188010fb0, 
> iov=0x7fb188010ff0, count=1, myframe=0x7fb188215740) at 
> client-rpc-fops.c:2802
>
> #1  0x00007fb1acf55d47 in rpc_clnt_handle_reply (clnt=0x7fb1a008fff0, 
> pollin=0x7fb1a0843910) at rpc-clnt.c:778
>
> #2  0x00007fb1acf562e5 in rpc_clnt_notify (trans=0x7fb1a00901c0, 
> mydata=0x7fb1a0090020, event=RPC_TRANSPORT_MSG_RECEIVED, 
> data=0x7fb1a0843910) at rpc-clnt.c:971
>
> #3  0x00007fb1acf52319 in rpc_transport_notify (this=0x7fb1a00901c0, 
> event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7fb1a0843910) at 
> rpc-transport.c:538
>
> #4  0x00007fb1a7e9934d in socket_event_poll_in (this=0x7fb1a00901c0, 
> notify_handled=_gf_true) at socket.c:2315
>
> #5  0x00007fb1a7e99992 in socket_event_handler (fd=20, idx=14, 
> gen=103, data=0x7fb1a00901c0, poll_in=1, poll_out=0, poll_err=0) at 
> socket.c:2471
>
> #6  0x00007fb1ad2005ac in event_dispatch_epoll_handler 
> (event_pool=0x175fb00, event=0x7fb1a7a0de84) at event-epoll.c:583
>
> #7  0x00007fb1ad200883 in event_dispatch_epoll_worker (data=0x17a73d0) 
> at event-epoll.c:659
>
> #8  0x00007fb1abf4c5da in start_thread () from /lib64/libpthread.so.0
>
> #9  0x00007fb1ab822cbf in clone () from /lib64/libc.so.6
>
> (gdb) info thread
>
>   Id   Target Id Frame
>
> * 1    Thread 0x7fb1a7a0e700 (LWP 8151) 0x00007fb1a6fd9d24 in 
> client3_3_lookup_cbk (req=0x7fb188010fb0, iov=0x7fb188010ff0, count=1, 
> myframe=0x7fb188215740) at client-rpc-fops.c:2802
>
>   2    Thread 0x7fb1aa0af700 (LWP 8147) 0x00007fb1ab761cbc in 
> sigtimedwait () from /lib64/libc.so.6
>
>   3    Thread 0x7fb1a98ae700 (LWP 8148) 0x00007fb1ab7f04b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   4    Thread 0x7fb1957fa700 (LWP 8266) 0x00007fb1abf528ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   5    Thread 0x7fb1a88ac700 (LWP 8150) 0x00007fb1abf528ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   6    Thread 0x7fb17f7fe700 (LWP 8269) 0x00007fb1abf5250c in 
> pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   7    Thread 0x7fb1aa8b0700 (LWP 8146) 0x00007fb1abf56300 in 
> nanosleep () from /lib64/libpthread.so.0
>
>   8    Thread 0x7fb1ad685780 (LWP 8145) 0x00007fb1abf4da3d in 
> __pthread_timedjoin_ex () from /lib64/libpthread.so.0
>
>   9    Thread 0x7fb1a542d700 (LWP 8251) 0x00007fb1ab7f04b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   10   Thread 0x7fb1a4c2c700 (LWP 8260) 0x00007fb1abf528ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   11   Thread 0x7fb196ffd700 (LWP 8263) 0x00007fb1abf528ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   12   Thread 0x7fb1a60d7700 (LWP 8247) 0x00007fb1ab822fe7 in 
> epoll_wait () from /lib64/libc.so.6
>
>   13   Thread 0x7fb1a90ad700 (LWP 8149) 0x00007fb1abf528ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
> (gdb) print (call_frame_t*)myframe
>
> $1 = (call_frame_t *) 0x7fb188215740
>
> (gdb) print *(call_frame_t*)myframe
>
> $2 = {root = 0x7fb1a0085090, parent = 0xcd4642c4a3efd678, frames = 
> {next = 0x151e2a92a5ae1bb, prev = 0x0}, *local = 0x0, this = 0x0, ret 
> = 0x0*, ref_count = 0, lock = {spinlock = 0, mutex = {__data = {
>
>         __lock = 0, __count = 0, __owner = 0, __nusers = 4, __kind = 
> 0, __spins = 0, __elision = 0, __list = {__prev = 0x7fb188215798, 
> __next = 0x7fb188215798}},
>
>       __size = '\000' <repeats 12 times>, "\004", '\000' <repeats 11 
> times>, "\230W!\210\261\177\000\000\230W!\210\261\177\000", __align = 
> 0}}, cookie = 0x7fb1882157a8, complete = (unknown: 2283886504),
>
>   op = 32689, begin = {tv_sec = 140400469825464, tv_usec = 
> 140400469825464}, end = {tv_sec = 140400878737576, tv_usec = 
> 140400132101048}, wind_from = 0x7fb18801cdc0 "", wind_to = 0x0, 
> unwind_from = 0x0,
>
>   unwind_to = 0x0}
>
> (gdb) thread 6
>
> [Switching to thread 6 (Thread 0x7fb17f7fe700 (LWP 8269))]
>
> #0  0x00007fb1abf5250c in pthread_cond_wait@@GLIBC_2.3.2 () from 
> /lib64/libpthread.so.0
>
> (gdb) bt
>
> #0  0x00007fb1abf5250c in pthread_cond_wait@@GLIBC_2.3.2 () from 
> /lib64/libpthread.so.0
>
> #1  0x00007fb1ad1dc993 in __syncbarrier_wait (barrier=0x7fb188014790, 
> waitfor=3) at syncop.c:1138
>
> #2  0x00007fb1ad1dc9e4 in syncbarrier_wait (barrier=0x7fb188014790, 
> waitfor=3) at syncop.c:1155
>
> #3  0x00007fb1a6d59cde in afr_selfheal_unlocked_discover_on 
> (*frame=0x7fb1882162d0*, inode=0x7fb188215740, gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177",
>
>     replies=0x7fb17f7fcf40, discover_on=0x7fb1a0084cb0 "\001\001\001", 
> <incomplete sequence \360\255\272>) at afr-self-heal-common.c:1809
>
> #4  0x00007fb1a6d59d80 in afr_selfheal_unlocked_discover 
> (*frame=0x7fb1882162d0*, inode=0x7fb188215740, gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177",
>
>     replies=0x7fb17f7fcf40) at afr-self-heal-common.c:1828
>
> #5  0x00007fb1a6d5e51f in afr_selfheal_unlocked_inspect 
> (frame=0x7fb1882162d0, this=0x7fb1a001db40, gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177",
>
>     link_inode=0x7fb17f7fd9c8, data_selfheal=0x7fb17f7fd9c4, 
> metadata_selfheal=0x7fb17f7fd9c0, entry_selfheal=0x7fb17f7fd9bc) at 
> afr-self-heal-common.c:2241
>
> #6  0x00007fb1a6d5f19b in afr_selfheal_do (frame=0x7fb1882162d0, 
> this=0x7fb1a001db40, gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177") at 
> afr-self-heal-common.c:2483
>
> #7  0x00007fb1a6d5f346 in afr_selfheal (this=0x7fb1a001db40, 
> gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177") at 
> afr-self-heal-common.c:2543
>
> #8  0x00007fb1a6d6ac5c in afr_shd_selfheal (healer=0x7fb1a0085640, 
> child=0, gfid=0x7fb17f7fdb00 
> "x\326\357\243\304BFͻ\341Z*\251\342Q\001\060\333\177\177\261\177") at 
> afr-self-heald.c:343
>
> #9  0x00007fb1a6d6b00b in afr_shd_index_heal (subvol=0x7fb1a00171e0, 
> entry=0x7fb1a0714180, parent=0x7fb17f7fddc0, data=0x7fb1a0085640) at 
> afr-self-heald.c:440
>
> #10 0x00007fb1ad201ed3 in syncop_mt_dir_scan (frame=0x7fb1a07a0e90, 
> subvol=0x7fb1a00171e0, loc=0x7fb17f7fddc0, pid=-6, 
> data=0x7fb1a0085640, fn=0x7fb1a6d6aebc <afr_shd_index_heal>, 
> xdata=0x7fb1a07b4ed0,
>
>     max_jobs=1, max_qlen=1024) at syncop-utils.c:407
>
> #11 0x00007fb1a6d6b2b5 in afr_shd_index_sweep (healer=0x7fb1a0085640, 
> vgfid=0x7fb1a6d93610 "glusterfs.xattrop_index_gfid") at 
> afr-self-heald.c:494
>
> #12 0x00007fb1a6d6b394 in afr_shd_index_sweep_all 
> (healer=0x7fb1a0085640) at afr-self-heald.c:517
>
> #13 0x00007fb1a6d6b697 in afr_shd_index_healer (data=0x7fb1a0085640) 
> at afr-self-heald.c:597
>
> #14 0x00007fb1abf4c5da in start_thread () from /lib64/libpthread.so.0
>
> #15 0x00007fb1ab822cbf in clone () from /lib64/libc.so.6
>
> *From:*Zhou, Cynthia (NSB - CN/Hangzhou)
> *Sent:* Thursday, October 11, 2018 3:36 PM
> *To:* Ravishankar N <ravishankar at redhat.com>
> *Cc:* gluster-users <gluster-users at gluster.org>
> *Subject:* glustershd coredump generated while reboot all 3 sn nodes
>
> Hi,
>
> I find that when restart sn node sometimes, the glustershd will exit 
> and generate coredump. It has happened twice in my env, I would like 
> to know your opinion on this issue, thanks!
>
> The glusterfs version I use is glusterfs3.12.3
>
> [root at sn-1:/root]
>
> # gluster v info log
>
> Volume Name: log
>
> Type: Replicate
>
> Volume ID: 87bcbaf8-5fa4-4060-9149-23f832befe92
>
> Status: Started
>
> Snapshot Count: 0
>
> Number of Bricks: 1 x 3 = 3
>
> Transport-type: tcp
>
> Bricks:
>
> Brick1: sn-0.local:/mnt/bricks/log/brick
>
> Brick2: sn-1.local:/mnt/bricks/log/brick
>
> Brick3: sn-2.local:/mnt/bricks/log/brick
>
> Options Reconfigured:
>
> server.allow-insecure: on
>
> cluster.quorum-type: auto
>
> network.ping-timeout: 42
>
> cluster.consistent-metadata: on
>
> cluster.favorite-child-policy: mtime
>
> cluster.quorum-reads: no
>
> cluster.server-quorum-type: none
>
> transport.address-family: inet
>
> nfs.disable: on
>
> performance.client-io-threads: off
>
> cluster.server-quorum-ratio: 51%
>
> [root at sn-1:/root]
>
> ///////////////////////////////////////////////glustershd 
> coredump////////////////////////////////////////////////////////////////
>
> # lz4 -d 
> core.glusterfs.0.c5f0c5547fbd4e5aa8f350b748e5675e.1812.1537967075000000.lz4
>
> Decoding file 
> core.glusterfs.0.c5f0c5547fbd4e5aa8f350b748e5675e.1812.1537967075000000
>
> core.glusterfs.0.c5f : decoded 263188480 bytes
>
> [root at sn-0:/mnt/export]
>
> # gdb /usr/sbin/glusterfs 
> core.glusterfs.0.c5f0c5547fbd4e5aa8f350b748e5675e.1812.1537967075000000
>
> GNU gdb (GDB) Fedora 8.1-14.wf29
>
> Copyright (C) 2018 Free Software Foundation, Inc.
>
> License GPLv3+: GNU GPL version 3 or later 
> <http://gnu.org/licenses/gpl.html>
>
> This is free software: you are free to change and redistribute it.
>
> There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
>
> and "show warranty" for details.
>
> This GDB was configured as "x86_64-redhat-linux-gnu".
>
> Type "show configuration" for configuration details.
>
> For bug reporting instructions, please see:
>
> <http://www.gnu.org/software/gdb/bugs/>.
>
> Find the GDB manual and other documentation resources online at:
>
> <http://www.gnu.org/software/gdb/documentation/>.
>
> For help, type "help".
>
> Type "apropos word" to search for commands related to "word"...
>
> Reading symbols from /usr/sbin/glusterfs...(no debugging symbols 
> found)...done.
>
> warning: core file may not match specified executable file.
>
> [New LWP 1818]
>
> [New LWP 1812]
>
> [New LWP 1813]
>
> [New LWP 1817]
>
> [New LWP 1966]
>
> [New LWP 1968]
>
> [New LWP 1970]
>
> [New LWP 1974]
>
> [New LWP 1976]
>
> [New LWP 1814]
>
> [New LWP 1815]
>
> [New LWP 1816]
>
> [New LWP 1828]
>
> [Thread debugging using libthread_db enabled]
>
> Using host libthread_db library "/lib64/libthread_db.so.1".
>
> Core was generated by `/usr/sbin/glusterfs -s sn-0.local --volfile-id 
> gluster/glustershd -p /var/run/g'.
>
> Program terminated with signal SIGSEGV, Segmentation fault.
>
> #0  0x00007f1b5e5d7d24 in client3_3_lookup_cbk (req=0x7f1b44002300, 
> iov=0x7f1b44002340, count=1, myframe=0x7f1b4401c850) at 
> client-rpc-fops.c:2802
>
> 2802 client-rpc-fops.c: No such file or directory.
>
> [Current thread is 1 (Thread 0x7f1b5f00c700 (LWP 1818))]
>
> Missing separate debuginfos, use: dnf debuginfo-install 
> rcp-pack-glusterfs-1.2.0_1_g54e6196-RCP2.wf29.x86_64
>
> (gdb) bt
>
> #0  0x00007f1b5e5d7d24 in client3_3_lookup_cbk (req=0x7f1b44002300, 
> iov=0x7f1b44002340, count=1, myframe=0x7f1b4401c850) at 
> client-rpc-fops.c:2802
>
> #1  0x00007f1b64553d47 in rpc_clnt_handle_reply (clnt=0x7f1b5808bbb0, 
> pollin=0x7f1b580c6620) at rpc-clnt.c:778
>
> #2  0x00007f1b645542e5 in rpc_clnt_notify (trans=0x7f1b5808bde0, 
> mydata=0x7f1b5808bbe0, event=RPC_TRANSPORT_MSG_RECEIVED, 
> data=0x7f1b580c6620) at rpc-clnt.c:971
>
> #3  0x00007f1b64550319 in rpc_transport_notify (this=0x7f1b5808bde0, 
> event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f1b580c6620) at 
> rpc-transport.c:538
>
> #4  0x00007f1b5f49734d in socket_event_poll_in (this=0x7f1b5808bde0, 
> notify_handled=_gf_true) at socket.c:2315
>
> #5  0x00007f1b5f497992 in socket_event_handler (fd=25, idx=15, gen=7, 
> data=0x7f1b5808bde0, poll_in=1, poll_out=0, poll_err=0) at socket.c:2471
>
> #6  0x00007f1b647fe5ac in event_dispatch_epoll_handler 
> (event_pool=0x230cb00, event=0x7f1b5f00be84) at event-epoll.c:583
>
> #7  0x00007f1b647fe883 in event_dispatch_epoll_worker (data=0x23543d0) 
> at event-epoll.c:659
>
> #8  0x00007f1b6354a5da in start_thread () from /lib64/libpthread.so.0
>
> #9  0x00007f1b62e20cbf in clone () from /lib64/libc.so.6
>
> *(gdb) print *(call_frame_t*)myframe*
>
> *$1 = {root = 0x100000000, parent = 0x100000005, frames = {next = 
> 0x7f1b4401c8a8, prev = 0x7f1b44010190}, local = 0x0, this = 0x0, ret = 
> 0x0, ref_count = 0, lock = {spinlock = 0, mutex = {__data = {*
>
> *__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, 
> __spins = 0, __elision = 0, __list = {__prev = 0x7f1b44010190, __next 
> = 0x0}}, *
>
> *      __size = '\000' <repeats 24 times>, 
> "\220\001\001D\033\177\000\000\000\000\000\000\000\000\000", __align = 
> 0}}, cookie = 0x7f1b4401ccf0, complete = _gf_false, op = GF_FOP_NULL, 
> begin = {*
>
> *tv_sec = 139755081730912, tv_usec = 139755081785872}, end = {tv_sec = 
> 448811404, tv_usec = 21474836481}, wind_from = 0x0, wind_to = 0x0, 
> unwind_from = 0x0, unwind_to = 0x0}*
>
> (gdb) info thread
>
>   Id   Target Id Frame
>
> * 1    Thread 0x7f1b5f00c700 (LWP 1818) 0x00007f1b5e5d7d24 in 
> client3_3_lookup_cbk (req=0x7f1b44002300, iov=0x7f1b44002340, count=1, 
> myframe=0x7f1b4401c850) at client-rpc-fops.c:2802
>
>   2    Thread 0x7f1b64c83780 (LWP 1812) 0x00007f1b6354ba3d in 
> __pthread_timedjoin_ex () from /lib64/libpthread.so.0
>
>   3    Thread 0x7f1b61eae700 (LWP 1813) 0x00007f1b63554300 in 
> nanosleep () from /lib64/libpthread.so.0
>
>   4    Thread 0x7f1b5feaa700 (LWP 1817) 0x00007f1b635508ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   5    Thread 0x7f1b5ca2b700 (LWP 1966) 0x00007f1b62dee4b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   6    Thread 0x7f1b4f7fe700 (LWP 1968) 0x00007f1b6355050c in 
> pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   7    Thread 0x7f1b4e7fc700 (LWP 1970) 0x00007f1b6355050c in 
> pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   8    Thread 0x7f1b4d7fa700 (LWP 1974) 0x00007f1b62dee4b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   9    Thread 0x7f1b33fff700 (LWP 1976) 0x00007f1b62dee4b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   10   Thread 0x7f1b616ad700 (LWP 1814) 0x00007f1b62d5fcbc in 
> sigtimedwait () from /lib64/libc.so.6
>
>   11   Thread 0x7f1b60eac700 (LWP 1815) 0x00007f1b62dee4b0 in 
> nanosleep () from /lib64/libc.so.6
>
>   12   Thread 0x7f1b606ab700 (LWP 1816) 0x00007f1b635508ca in 
> pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
>
>   13   Thread 0x7f1b5d6d5700 (LWP 1828) 0x00007f1b62e20fe7 in 
> epoll_wait () from /lib64/libc.so.6
>
> (gdb) quit
>
> The source code is like this, so from gdb it coredump because 
> frame->local is *NULL*!!
>
> From sn-0 journal log,
>
> Sep 26 16:04:40.034577 sn-0 systemd-coredump[2612]: Process 1812 
> (glusterfs) of user 0 dumped core.
>
>                                                     Stack trace of 
> thread 1818:
>
> #0  0x00007f1b5e5d7d24 client3_3_lookup_cbk (client.so)
>
> #1  0x00007f1b64553d47 rpc_clnt_handle_reply (libgfrpc.so.0)
>
> #2  0x00007f1b645542e5 rpc_clnt_notify (libgfrpc.so.0)
>
> #3  0x00007f1b64550319 rpc_transport_notify (libgfrpc.so.0)
>
> #4  0x00007f1b5f49734d socket_event_poll_in (socket.so)
>
>                                  #5  0x00007f1b5f497992 
> socket_event_handler (socket.so)
>
> #6  0x00007f1b647fe5ac event_dispatch_epoll_handler (libglusterfs.so.0)
>
>          #7  0x00007f1b647fe883 event_dispatch_epoll_worker 
> (libglusterfs.so.0)
>
> #8  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #9  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1812:
>
> #0  0x00007f1b6354ba3d __GI___pthread_timedjoin_ex (libpthread.so.0)
>
> #1  0x00007f1b647feae1 event_dispatch_epoll (libglusterfs.so.0)
>
>                                  #2  0x00007f1b647c2703 event_dispatch 
> (libglusterfs.so.0)
>
> #3  0x000000000040ab95 main (glusterfsd)
>
> #4  0x00007f1b62d4baf7 __libc_start_main (libc.so.6)
>
> #5  0x000000000040543a _start (glusterfsd)
>
>                                                     Stack trace of 
> thread 1813:
>
> #0  0x00007f1b63554300 __nanosleep (libpthread.so.0)
>
> #1  0x00007f1b647a04e5 gf_timer_proc (libglusterfs.so.0)
>
>                #2  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #3  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1817:
>
> #0  0x00007f1b635508ca pthread_cond_timedwait@@GLIBC_2.3.2 
> (libpthread.so.0)
>
> #1  0x00007f1b647d98e3 syncenv_task (libglusterfs.so.0)
>
> #2  0x00007f1b647d9b7e syncenv_processor (libglusterfs.so.0)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
>                                      #4  0x00007f1b62e20cbf __clone 
> (libc.so.6)
>
>                                                     Stack trace of 
> thread 1966:
>
>          #0  0x00007f1b62dee4b0 __nanosleep (libc.so.6)
>
> #1  0x00007f1b62dee38a sleep (libc.so.6)
>
> #2  0x00007f1b5e36970c afr_shd_index_healer (replicate.so)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1968:
>
> #0  0x00007f1b6355050c pthread_cond_wait@@GLIBC_2.3.2 (libpthread.so.0)
>
> #1  0x00007f1b647da993 __syncbarrier_wait (libglusterfs.so.0)
>
> #2  0x00007f1b647da9e4 syncbarrier_wait (libglusterfs.so.0)
>
>                       #3  0x00007f1b5e357cde 
> afr_selfheal_unlocked_discover_on (replicate.so)
>
> #4  0x00007f1b5e357d80 afr_selfheal_unlocked_discover (replicate.so)
>
>            #5  0x00007f1b5e363bf8 __afr_selfheal_entry_prepare 
> (replicate.so)
>
> #6  0x00007f1b5e3641c0 afr_selfheal_entry_dirent (replicate.so)
>
> #7  0x00007f1b5e36488a afr_selfheal_entry_do_subvol (replicate.so)
>
> #8  0x00007f1b5e365077 afr_selfheal_entry_do (replicate.so)
>
> #9  0x00007f1b5e3656b6 __afr_selfheal_entry (replicate.so)
>
> #10 0x00007f1b5e365bba afr_selfheal_entry (replicate.so)
>
> #11 0x00007f1b5e35d250 afr_selfheal_do (replicate.so)
>
>                                    #12 0x00007f1b5e35d346 afr_selfheal 
> (replicate.so)
>
> #13 0x00007f1b5e368c5c afr_shd_selfheal (replicate.so)
>
> #14 0x00007f1b5e36900b afr_shd_index_heal (replicate.so)
>
> #15 0x00007f1b647ffed3 syncop_mt_dir_scan (libglusterfs.so.0)
>
> #16 0x00007f1b5e3692b5 afr_shd_index_sweep (replicate.so)
>
> #17 0x00007f1b5e369394 afr_shd_index_sweep_all (replicate.so)
>
> #18 0x00007f1b5e369697 afr_shd_index_healer (replicate.so)
>
>                                    #19 0x00007f1b6354a5da start_thread 
> (libpthread.so.0)
>
> #20 0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1970:
>
> #0  0x00007f1b6355050c pthread_cond_wait@@GLIBC_2.3.2 (libpthread.so.0)
>
> #1  0x00007f1b647da993 __syncbarrier_wait (libglusterfs.so.0)
>
> #2  0x00007f1b647da9e4 syncbarrier_wait (libglusterfs.so.0)
>
> #3  0x00007f1b5e357742 afr_selfheal_unlocked_lookup_on (replicate.so)
>
> #4  0x00007f1b5e364204 afr_selfheal_entry_dirent (replicate.so)
>
> #5  0x00007f1b5e36488a afr_selfheal_entry_do_subvol (replicate.so)
>
> #6  0x00007f1b5e365077 afr_selfheal_entry_do (replicate.so)
>
> #7  0x00007f1b5e3656b6 __afr_selfheal_entry (replicate.so)
>
>                    #8  0x00007f1b5e365bba afr_selfheal_entry 
> (replicate.so)
>
> #9  0x00007f1b5e35d250 afr_selfheal_do (replicate.so)
>
> #10 0x00007f1b5e35d346 afr_selfheal (replicate.so)
>
> #11 0x00007f1b5e368c5c afr_shd_selfheal (replicate.so)
>
> #12 0x00007f1b5e36900b afr_shd_index_heal (replicate.so)
>
>                                          #13 0x00007f1b647ffed3 
> syncop_mt_dir_scan (libglusterfs.so.0)
>
> #14 0x00007f1b5e3692b5 afr_shd_index_sweep (replicate.so)
>
>          #15 0x00007f1b5e369394 afr_shd_index_sweep_all (replicate.so)
>
> #16 0x00007f1b5e369697 afr_shd_index_healer (replicate.so)
>
> #17 0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #18 0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1974:
>
> #0  0x00007f1b62dee4b0 __nanosleep (libc.so.6)
>
> #1  0x00007f1b62dee38a sleep (libc.so.6)
>
> #2  0x00007f1b5e36970c afr_shd_index_healer (replicate.so)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1976:
>
> #0  0x00007f1b62dee4b0 __nanosleep (libc.so.6)
>
>                         #1  0x00007f1b62dee38a sleep (libc.so.6)
>
> #2  0x00007f1b5e36970c afr_shd_index_healer (replicate.so)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1814:
>
> #0  0x00007f1b62d5fcbc __sigtimedwait (libc.so.6)
>
> #1  0x00007f1b63554afc sigwait (libpthread.so.0)
>
> #2  0x0000000000409ed7 glusterfs_sigwaiter (glusterfsd)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1815:
>
> #0  0x00007f1b62dee4b0 __nanosleep (libc.so.6)
>
>                             #1  0x00007f1b62dee38a sleep (libc.so.6)
>
> #2  0x00007f1b647c3f5c pool_sweeper (libglusterfs.so.0)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1816:
>
>                                                    #0 
> 0x00007f1b635508ca pthread_cond_timedwait@@GLIBC_2.3.2 (libpthread.so.0)
>
> #1  0x00007f1b647d98e3 syncenv_task (libglusterfs.so.0)
>
>                                 #2  0x00007f1b647d9b7e 
> syncenv_processor (libglusterfs.so.0)
>
> #3  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
> #4  0x00007f1b62e20cbf __clone (libc.so.6)
>
>                                                     Stack trace of 
> thread 1828:
>
> #0  0x00007f1b62e20fe7 epoll_wait (libc.so.6)
>
> #1  0x00007f1b647fe855 event_dispatch_epoll_worker (libglusterfs.so.0)
>
> #2  0x00007f1b6354a5da start_thread (libpthread.so.0)
>
>                                       #3  0x00007f1b62e20cbf __clone 
> (libc.so.6)
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20181016/99c98a2b/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image002.jpg
Type: image/jpeg
Size: 59028 bytes
Desc: not available
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20181016/99c98a2b/attachment.jpg>


More information about the Gluster-users mailing list