[Gluster-users] df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump

Mauro Tridici mauro.tridici at cmcc.it
Tue Sep 26 07:51:42 UTC 2017


Hi Ashish,

thank you for your answer.
Do you need complete client log file only or something else in particular?
Unfortunately, I never used “bt” command. Could you please provide me an usage example string?

I will provide all logs you need.
Thank you again,
Mauro

> Il giorno 26 set 2017, alle ore 09:30, Ashish Pandey <aspandey at redhat.com> ha scritto:
> 
> Hi Mauro,
> 
> We would require complete log file to debug this issue.
> Also, could you please provide some more information of the core after attaching to gdb and using command "bt".
> 
> ---
> Ashish
> 
> 
> From: "Mauro Tridici" <mauro.tridici at cmcc.it>
> To: "Gluster Users" <gluster-users at gluster.org>
> Sent: Monday, September 25, 2017 5:59:10 PM
> Subject: [Gluster-users] df command shows transport endpoint mount error on        gluster client v.3.10.5 + core dump
> 
> Dear Gluster Users,
> 
> I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
> 
> [root at s01 tier2]# gluster volume info
>  
> Volume Name: tier2
> Type: Distributed-Disperse
> Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 6 x (4 + 2) = 36
> Transport-type: tcp
> Bricks:
> Brick1: s01-stg:/gluster/mnt1/brick
> Brick2: s02-stg:/gluster/mnt1/brick
> Brick3: s03-stg:/gluster/mnt1/brick
> Brick4: s01-stg:/gluster/mnt2/brick
> Brick5: s02-stg:/gluster/mnt2/brick
> Brick6: s03-stg:/gluster/mnt2/brick
> Brick7: s01-stg:/gluster/mnt3/brick
> Brick8: s02-stg:/gluster/mnt3/brick
> Brick9: s03-stg:/gluster/mnt3/brick
> Brick10: s01-stg:/gluster/mnt4/brick
> Brick11: s02-stg:/gluster/mnt4/brick
> Brick12: s03-stg:/gluster/mnt4/brick
> Brick13: s01-stg:/gluster/mnt5/brick
> Brick14: s02-stg:/gluster/mnt5/brick
> Brick15: s03-stg:/gluster/mnt5/brick
> Brick16: s01-stg:/gluster/mnt6/brick
> Brick17: s02-stg:/gluster/mnt6/brick
> Brick18: s03-stg:/gluster/mnt6/brick
> Brick19: s01-stg:/gluster/mnt7/brick
> Brick20: s02-stg:/gluster/mnt7/brick
> Brick21: s03-stg:/gluster/mnt7/brick
> Brick22: s01-stg:/gluster/mnt8/brick
> Brick23: s02-stg:/gluster/mnt8/brick
> Brick24: s03-stg:/gluster/mnt8/brick
> Brick25: s01-stg:/gluster/mnt9/brick
> Brick26: s02-stg:/gluster/mnt9/brick
> Brick27: s03-stg:/gluster/mnt9/brick
> Brick28: s01-stg:/gluster/mnt10/brick
> Brick29: s02-stg:/gluster/mnt10/brick
> Brick30: s03-stg:/gluster/mnt10/brick
> Brick31: s01-stg:/gluster/mnt11/brick
> Brick32: s02-stg:/gluster/mnt11/brick
> Brick33: s03-stg:/gluster/mnt11/brick
> Brick34: s01-stg:/gluster/mnt12/brick
> Brick35: s02-stg:/gluster/mnt12/brick
> Brick36: s03-stg:/gluster/mnt12/brick
> Options Reconfigured:
> features.scrub: Active
> features.bitrot: on
> features.inode-quota: on
> features.quota: on
> performance.client-io-threads: on
> cluster.min-free-disk: 10
> cluster.quorum-type: auto
> transport.address-family: inet
> nfs.disable: on
> server.event-threads: 4
> client.event-threads: 4
> cluster.lookup-optimize: on
> performance.readdir-ahead: on
> performance.parallel-readdir: off
> cluster.readdir-optimize: on
> features.cache-invalidation: on
> features.cache-invalidation-timeout: 600
> performance.stat-prefetch: on
> performance.cache-invalidation: on
> performance.md-cache-timeout: 600
> network.inode-lru-limit: 50000
> performance.io-cache: off
> disperse.cpu-extensions: auto
> performance.io-thread-count: 16
> features.quota-deem-statfs: on
> features.default-soft-limit: 90
> cluster.server-quorum-type: server
> cluster.brick-multiplex: on
> cluster.server-quorum-ratio: 51%
> 
> I also started a long write test (about 69 TB to be written) from different gluster clients.
> One of this client returned the error "Transport Endpoint is not connected" during the rsync copy process.
> A core dump file has been generated by the issue; this is the core dump content:
> 
> GNU gdb (GDB) Red Hat Enterprise Linux (7.2-50.el6)
> Copyright (C) 2010 Free Software Foundation, Inc.
> License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
> This is free software: you are free to change and redistribute it.
> There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
> and "show warranty" for details.
> This GDB was configured as "x86_64-redhat-linux-gnu".
> For bug reporting instructions, please see:
> <http://www.gnu.org/software/gdb/bugs/>...
> Missing separate debuginfo for the main executable file
> Try: yum --disablerepo='*' --enablerepo='*-debuginfo' install /usr/lib/debug/.build-id/fb/4d988b681faa09bb74becacd7a24f4186e8185
> [New Thread 15802]
> [New Thread 15804]
> [New Thread 15805]
> [New Thread 6856]
> [New Thread 30432]
> [New Thread 30486]
> [New Thread 1619]
> [New Thread 15806]
> [New Thread 15810]
> [New Thread 30412]
> [New Thread 15809]
> [New Thread 15799]
> [New Thread 30487]
> [New Thread 15795]
> [New Thread 15797]
> [New Thread 15798]
> [New Thread 15800]
> [New Thread 15801]
> Core was generated by `/usr/sbin/glusterfs --volfile-server=s01-stg --volfile-id=/tier2 /tier2'.
> Program terminated with signal 6, Aborted.
> #0  0x00000032a74328a5 in ?? ()
> "/core.15795" is a core file.
> Please specify an executable to debug.
> 
> No problem detected on the gluster servers, no brick down and so on...
> 
> Anyone of us experienced the same problem? 
> If yes, how did you resolve it?
> 
> You can find the client "/var/log/messages" and "/var/log/glusterfs" log files content.
> Thank you in advance.
> Mauro Tridici
> 
> --------------------------------------
> 
> In /var/log/syslog-ng/messages on the client (OS=centos 6.2, 16 cores, 64GB RAM, gluster client v.3.10.5)
> 
> Sep 23 10:42:43 login2 tier2[15795]: pending frames:
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
> Sep 23 10:42:43 login2 tier2[15795]: patchset: git://git.gluster.org/glusterfs.git
> Sep 23 10:42:43 login2 tier2[15795]: signal received: 6
> Sep 23 10:42:43 login2 tier2[15795]: time of crash: 
> Sep 23 10:42:43 login2 tier2[15795]: 2017-09-23 08:42:43
> Sep 23 10:42:43 login2 tier2[15795]: configuration details:
> Sep 23 10:42:43 login2 tier2[15795]: argp 1
> Sep 23 10:42:43 login2 tier2[15795]: backtrace 1
> Sep 23 10:42:43 login2 tier2[15795]: dlfcn 1
> Sep 23 10:42:43 login2 tier2[15795]: libpthread 1
> Sep 23 10:42:43 login2 tier2[15795]: llistxattr 1
> Sep 23 10:42:43 login2 tier2[15795]: setfsid 1
> Sep 23 10:42:43 login2 tier2[15795]: spinlock 1
> Sep 23 10:42:43 login2 tier2[15795]: epoll.h 1
> Sep 23 10:42:43 login2 tier2[15795]: xattr.h 1
> Sep 23 10:42:43 login2 tier2[15795]: st_atim.tv_nsec 1
> Sep 23 10:42:43 login2 tier2[15795]: package-string: glusterfs 3.10.5
> Sep 23 10:42:43 login2 tier2[15795]: ---------
> Sep 23 10:42:43 login2 abrt[8605]: file /usr/sbin/glusterfsd seems to be deleted
> 
> In /var/log/glusterfs file:
> 
> pending frames:
> frame : type(1) op(WRITE)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(1) op(WRITE)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(1) op(WRITE)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(1) op(WRITE)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> frame : type(0) op(0)
> patchset: git://git.gluster.org/glusterfs.git
> signal received: 6
> time of crash: 
> 2017-09-23 08:42:43
> configuration details:
> argp 1
> backtrace 1
> dlfcn 1
> libpthread 1
> llistxattr 1
> setfsid 1
> spinlock 1
> epoll.h 1
> xattr.h 1
> st_atim.tv_nsec 1
> package-string: glusterfs 3.10.5
> /usr/lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0x9c)[0x7fd8d8b8af3c]
> /usr/lib64/libglusterfs.so.0(gf_print_trace+0x338)[0x7fd8d8b96538]
> /lib64/libc.so.6[0x32a7432920]
> /lib64/libc.so.6(gsignal+0x35)[0x32a74328a5]
> /lib64/libc.so.6(abort+0x175)[0x32a7434085]
> /lib64/libc.so.6[0x32a746ffe7]
> /lib64/libc.so.6[0x32a7475916]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x3181f)[0x7fd8cd51681f]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x31dad)[0x7fd8cd516dad]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x12027)[0x7fd8cd4f7027]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x11dc5)[0x7fd8cd4f6dc5]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x1475f)[0x7fd8cd4f975f]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x15178)[0x7fd8cd4fa178]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x31dc8)[0x7fd8cd516dc8]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x12027)[0x7fd8cd4f7027]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x11dc5)[0x7fd8cd4f6dc5]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x13f24)[0x7fd8cd4f8f24]
> /usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x33e12)[0x7fd8cd518e12]
> /usr/lib64/glusterfs/3.10.5/xlator/protocol/client.so(+0x2c420)[0x7fd8cd793420]
> /usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)[0x7fd8d8955ad5]
> /usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x195)[0x7fd8d8956c85]
> /usr/lib64/libgfrpc.so.0(rpc_transport_notify+0x28)[0x7fd8d8951d68]
> /usr/lib64/glusterfs/3.10.5/rpc-transport/socket.so(+0x9ccd)[0x7fd8ce5f1ccd]
> /usr/lib64/glusterfs/3.10.5/rpc-transport/socket.so(+0xaffe)[0x7fd8ce5f2ffe]
> /usr/lib64/libglusterfs.so.0(+0x87806)[0x7fd8d8be8806]
> /lib64/libpthread.so.0[0x32a8007851]
> /lib64/libc.so.6(clone+0x6d)[0x32a74e811d]
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
> 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170926/042eabdf/attachment.html>


More information about the Gluster-users mailing list