[Gluster-users] df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
Ashish Pandey
aspandey at redhat.com
Tue Sep 26 08:11:44 UTC 2017
Hi,
Following are the command to get the debug info for gluster -
gdb /usr/local/sbin/glusterfs <core file path>
Then on gdb prompt you need to type bt or backtrace
(gdb) bt
You can also provide the output of "thread apply all bt".
(gdb) thread apply all bt
Above commands should be executed on client node on which you have mounted the gluster volume and it crashed.
However, I am not sure if you have enabled the generation of core dump on your system and have setup path to coredump.
https://stackoverflow.com/questions/8305866/how-to-analyze-a-programs-core-dump-file-with-gdb
https://unix.stackexchange.com/questions/192716/how-to-set-the-core-dump-file-location-and-name
https://stackoverflow.com/questions/2065912/core-dumped-but-core-file-is-not-in-current-directory
--
Ashish
----- Original Message -----
From: "Mauro Tridici" <mauro.tridici at cmcc.it>
To: "Ashish Pandey" <aspandey at redhat.com>
Cc: "Gluster Users" <gluster-users at gluster.org>
Sent: Tuesday, September 26, 2017 1:21:42 PM
Subject: Re: [Gluster-users] df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
Hi Ashish,
thank you for your answer.
Do you need complete client log file only or something else in particular?
Unfortunately, I never used “bt” command. Could you please provide me an usage example string?
I will provide all logs you need.
Thank you again,
Mauro
Il giorno 26 set 2017, alle ore 09:30, Ashish Pandey < aspandey at redhat.com > ha scritto:
Hi Mauro,
We would require complete log file to debug this issue.
Also, could you please provide some more information of the core after attaching to gdb and using command "bt".
---
Ashish
----- Original Message -----
From: "Mauro Tridici" < mauro.tridici at cmcc.it >
To: "Gluster Users" < gluster-users at gluster.org >
Sent: Monday, September 25, 2017 5:59:10 PM
Subject: [Gluster-users] df command shows transport endpoint mount error on gluster client v.3.10.5 + core dump
Dear Gluster Users,
I implemented a distributed disperse 6x(4+2) gluster (v.3.10.5) volume with the following options:
[root at s01 tier2]# gluster volume info
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/brick
Brick6: s03-stg:/gluster/mnt2/brick
Brick7: s01-stg:/gluster/mnt3/brick
Brick8: s02-stg:/gluster/mnt3/brick
Brick9: s03-stg:/gluster/mnt3/brick
Brick10: s01-stg:/gluster/mnt4/brick
Brick11: s02-stg:/gluster/mnt4/brick
Brick12: s03-stg:/gluster/mnt4/brick
Brick13: s01-stg:/gluster/mnt5/brick
Brick14: s02-stg:/gluster/mnt5/brick
Brick15: s03-stg:/gluster/mnt5/brick
Brick16: s01-stg:/gluster/mnt6/brick
Brick17: s02-stg:/gluster/mnt6/brick
Brick18: s03-stg:/gluster/mnt6/brick
Brick19: s01-stg:/gluster/mnt7/brick
Brick20: s02-stg:/gluster/mnt7/brick
Brick21: s03-stg:/gluster/mnt7/brick
Brick22: s01-stg:/gluster/mnt8/brick
Brick23: s02-stg:/gluster/mnt8/brick
Brick24: s03-stg:/gluster/mnt8/brick
Brick25: s01-stg:/gluster/mnt9/brick
Brick26: s02-stg:/gluster/mnt9/brick
Brick27: s03-stg:/gluster/mnt9/brick
Brick28: s01-stg:/gluster/mnt10/brick
Brick29: s02-stg:/gluster/mnt10/brick
Brick30: s03-stg:/gluster/mnt10/brick
Brick31: s01-stg:/gluster/mnt11/brick
Brick32: s02-stg:/gluster/mnt11/brick
Brick33: s03-stg:/gluster/mnt11/brick
Brick34: s01-stg:/gluster/mnt12/brick
Brick35: s02-stg:/gluster/mnt12/brick
Brick36: s03-stg:/gluster/mnt12/brick
Options Reconfigured:
features.scrub: Active
features.bitrot: on
features.inode-quota: on
features.quota: on
performance.client-io-threads: on
cluster.min-free-disk: 10
cluster.quorum-type: auto
transport.address-family: inet
nfs.disable: on
server.event-threads: 4
client.event-threads: 4
cluster.lookup-optimize: on
performance.readdir-ahead: on
performance.parallel-readdir: off
cluster.readdir-optimize: on
features.cache-invalidation: on
features.cache-invalidation-timeout: 600
performance.stat-prefetch: on
performance.cache-invalidation: on
performance.md-cache-timeout: 600
network.inode-lru-limit: 50000
performance.io -cache: off
disperse.cpu-extensions: auto
performance.io -thread-count: 16
features.quota-deem-statfs: on
features.default-soft-limit: 90
cluster.server-quorum-type: server
cluster.brick-multiplex: on
cluster.server-quorum-ratio: 51%
I also started a long write test (about 69 TB to be written) from different gluster clients.
One of this client returned the error "Transport Endpoint is not connected" during the rsync copy process.
A core dump file has been generated by the issue; this is the core dump content:
GNU gdb (GDB) Red Hat Enterprise Linux (7.2-50.el6)
Copyright (C) 2010 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later < http://gnu.org/licenses/gpl.html >
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-redhat-linux-gnu".
For bug reporting instructions, please see:
< http://www.gnu.org/software/gdb/bugs/ >...
Missing separate debuginfo for the main executable file
Try: yum --disablerepo='*' --enablerepo='*-debuginfo' install /usr/lib/debug/.build-id/fb/4d988b681faa09bb74becacd7a24f4186e8185
[New Thread 15802]
[New Thread 15804]
[New Thread 15805]
[New Thread 6856]
[New Thread 30432]
[New Thread 30486]
[New Thread 1619]
[New Thread 15806]
[New Thread 15810]
[New Thread 30412]
[New Thread 15809]
[New Thread 15799]
[New Thread 30487]
[New Thread 15795]
[New Thread 15797]
[New Thread 15798]
[New Thread 15800]
[New Thread 15801]
Core was generated by `/usr/sbin/glusterfs --volfile-server=s01-stg --volfile-id=/tier2 /tier2'.
Program terminated with signal 6, Aborted.
#0 0x00000032a74328a5 in ?? ()
"/core.15795" is a core file.
Please specify an executable to debug.
No problem detected on the gluster servers, no brick down and so on...
Anyone of us experienced the same problem?
If yes, how did you resolve it?
You can find the client "/var/log/messages" and "/var/log/glusterfs" log files content.
Thank you in advance.
Mauro Tridici
--------------------------------------
In /var/log/syslog-ng/messages on the client (OS=centos 6.2, 16 cores, 64GB RAM, gluster client v.3.10.5)
Sep 23 10:42:43 login2 tier2[15795]: pending frames:
Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(1) op(WRITE)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: frame : type(0) op(0)
Sep 23 10:42:43 login2 tier2[15795]: patchset: git://git.gluster.org/glusterfs.git
Sep 23 10:42:43 login2 tier2[15795]: signal received: 6
Sep 23 10:42:43 login2 tier2[15795]: time of crash:
Sep 23 10:42:43 login2 tier2[15795]: 2017-09-23 08:42:43
Sep 23 10:42:43 login2 tier2[15795]: configuration details:
Sep 23 10:42:43 login2 tier2[15795]: argp 1
Sep 23 10:42:43 login2 tier2[15795]: backtrace 1
Sep 23 10:42:43 login2 tier2[15795]: dlfcn 1
Sep 23 10:42:43 login2 tier2[15795]: libpthread 1
Sep 23 10:42:43 login2 tier2[15795]: llistxattr 1
Sep 23 10:42:43 login2 tier2[15795]: setfsid 1
Sep 23 10:42:43 login2 tier2[15795]: spinlock 1
Sep 23 10:42:43 login2 tier2[15795]: epoll.h 1
Sep 23 10:42:43 login2 tier2[15795]: xattr.h 1
Sep 23 10:42:43 login2 tier2[15795]: st_atim.tv_nsec 1
Sep 23 10:42:43 login2 tier2[15795]: package-string: glusterfs 3.10.5
Sep 23 10:42:43 login2 tier2[15795]: ---------
Sep 23 10:42:43 login2 abrt[8605]: file /usr/sbin/glusterfsd seems to be deleted
In /var/log/glusterfs file:
pending frames:
frame : type(1) op(WRITE)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(1) op(WRITE)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(1) op(WRITE)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(1) op(WRITE)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
frame : type(0) op(0)
patchset: git://git.gluster.org/glusterfs.git
signal received: 6
time of crash:
2017-09-23 08:42:43
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.10.5
/usr/lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0x9c)[0x7fd8d8b8af3c]
/usr/lib64/libglusterfs.so.0(gf_print_trace+0x338)[0x7fd8d8b96538]
/lib64/libc.so.6[0x32a7432920]
/lib64/libc.so.6(gsignal+0x35)[0x32a74328a5]
/lib64/libc.so.6(abort+0x175)[0x32a7434085]
/lib64/libc.so.6[0x32a746ffe7]
/lib64/libc.so.6[0x32a7475916]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x3181f)[0x7fd8cd51681f]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x31dad)[0x7fd8cd516dad]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x12027)[0x7fd8cd4f7027]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x11dc5)[0x7fd8cd4f6dc5]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x1475f)[0x7fd8cd4f975f]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x15178)[0x7fd8cd4fa178]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x31dc8)[0x7fd8cd516dc8]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x12027)[0x7fd8cd4f7027]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x11dc5)[0x7fd8cd4f6dc5]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x13f24)[0x7fd8cd4f8f24]
/usr/lib64/glusterfs/3.10.5/xlator/cluster/disperse.so(+0x33e12)[0x7fd8cd518e12]
/usr/lib64/glusterfs/3.10.5/xlator/protocol/client.so(+0x2c420)[0x7fd8cd793420]
/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5)[0x7fd8d8955ad5]
/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x195)[0x7fd8d8956c85]
/usr/lib64/libgfrpc.so.0(rpc_transport_notify+0x28)[0x7fd8d8951d68]
/usr/lib64/glusterfs/3.10.5/rpc-transport/socket.so(+0x9ccd)[0x7fd8ce5f1ccd]
/usr/lib64/glusterfs/3.10.5/rpc-transport/socket.so(+0xaffe)[0x7fd8ce5f2ffe]
/usr/lib64/libglusterfs.so.0(+0x87806)[0x7fd8d8be8806]
/lib64/libpthread.so.0[0x32a8007851]
/lib64/libc.so.6(clone+0x6d)[0x32a74e811d]
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170926/5e62c8d7/attachment.html>
More information about the Gluster-users
mailing list