[Gluster-users] I/O error on replicated volume

Joe Julian joe at julianfamily.org
Thu Mar 26 20:20:02 UTC 2015


Every 3 seconds implies, to me, that it's trying to reconnect to a server.

On 03/26/2015 01:12 PM, Jonathan Heese wrote:
>
> Joe,
>
>
> Hmmm.... But every 3 seconds for all eternity? Seems a bit much for a 
> "warning", doesn't it?
>
>
> Did you see my last reply? My nfs-server.vol file seems to indicate 
> that RDMA is still in use in some capacity... Is this normal? If not, 
> how can I reconcile this?
>
>
> Thanks.
>
>
> Regards,
>
> Jon Heese
>
>
> ------------------------------------------------------------------------
> *From:* gluster-users-bounces at gluster.org 
> <gluster-users-bounces at gluster.org> on behalf of Joe Julian 
> <joe at julianfamily.org>
> *Sent:* Thursday, March 26, 2015 4:08 PM
> *To:* gluster-users at gluster.org
> *Subject:* Re: [Gluster-users] I/O error on replicated volume
> The RDMA warnings are not relevant if you don't use RDMA. It's simply 
> pointing out that it tried to register and it couldn't, which would be 
> expected if your system doesn't support it.
>
> On 03/23/2015 12:29 AM, Mohammed Rafi K C wrote:
>>
>> On 03/23/2015 11:28 AM, Jonathan Heese wrote:
>>> On Mar 23, 2015, at 1:20 AM, "Mohammed Rafi K C" 
>>> <rkavunga at redhat.com <mailto:rkavunga at redhat.com>> wrote:
>>>
>>>>
>>>> On 03/21/2015 07:49 PM, Jonathan Heese wrote:
>>>>>
>>>>> Mohamed,
>>>>>
>>>>>
>>>>> I have completed the steps you suggested (unmount all, stop the 
>>>>> volume, set the config.transport to tcp, start the volume, mount, 
>>>>> etc.), and the behavior has indeed changed.
>>>>>
>>>>>
>>>>> [root at duke ~]# gluster volume info
>>>>>
>>>>> Volume Name: gluster_disk
>>>>> Type: Replicate
>>>>> Volume ID: 2307a5a8-641e-44f4-8eaf-7cc2b704aafd
>>>>> Status: Started
>>>>> Number of Bricks: 1 x 2 = 2
>>>>> Transport-type: tcp
>>>>> Bricks:
>>>>> Brick1: duke-ib:/bricks/brick1
>>>>> Brick2: duchess-ib:/bricks/brick1
>>>>> Options Reconfigured:
>>>>> config.transport: tcp
>>>>>
>>>>>
>>>>> [root at duke ~]# gluster volume status
>>>>> Status of volume: gluster_disk
>>>>> Gluster process Port    Online  Pid
>>>>> ------------------------------------------------------------------------------
>>>>> Brick duke-ib:/bricks/brick1 49152   Y       16362
>>>>> Brick duchess-ib:/bricks/brick1 49152   Y       14155
>>>>> NFS Server on localhost 2049    Y       16374
>>>>> Self-heal Daemon on localhost                           N/A 
>>>>> Y       16381
>>>>> NFS Server on duchess-ib 2049    Y       14167
>>>>> Self-heal Daemon on duchess-ib                          N/A 
>>>>> Y       14174
>>>>>
>>>>> Task Status of Volume gluster_disk
>>>>> ------------------------------------------------------------------------------
>>>>> There are no active volume tasks
>>>>>
>>>>> I am no longer seeing the I/O errors during prolonged periods of 
>>>>> write I/O that I was seeing when the transport was set to rdma. 
>>>>> However, I am seeing this message on both nodes every 3 seconds 
>>>>> (almost exactly):
>>>>>
>>>>>
>>>>> ==> /var/log/glusterfs/nfs.log <==
>>>>> [2015-03-21 14:17:40.379719] W 
>>>>> [rdma.c:1076:gf_rdma_cm_event_handler] 0-gluster_disk-client-1: 
>>>>> cma event RDMA_CM_EVENT_REJECTED, error 8 (me:10.10.10.1:1023 
>>>>> peer:10.10.10.2:49152)
>>>>>
>>>>>
>>>>> Is this something to worry about?
>>>>>
>>>> If you are not using nfs to export the volumes, there is nothing to 
>>>> worry.
>>>
>>> I'm using the native glusterfs FUSE component to mount the volume 
>>> locally on both servers -- I assume that you're referring to the 
>>> standard NFS protocol stuff, which I'm not using here.
>>>
>>> Incidentally, I would like to keep my logs from filling up with junk 
>>> if possible.  Is there something I can do to get rid of these 
>>> (useless?) error messages?
>>
>> If i understand correctly, you are getting this enormous log message 
>> from nfs log only, all other logs and everything are fine now, right 
>> ? If that is the case, and you are not at all using nfs for exporting 
>> the volume, as  a workaround you can disable nfs for your volume or 
>> cluster. (gluster v set nfs.disable on). This will turnoff your 
>> gluster nfs server, and you will no longer get those log messages.
>>
>>
>>>>> Any idea why there are rdma pieces in play when I've set my 
>>>>> transport to tcp?
>>>>>
>>>>
>>>> there should not be any piece of rdma,if possible, can you paste 
>>>> the volfile for nfs server. You can find the volfile in 
>>>> /var/lib/glusterd/nfs/nfs-server.vol or 
>>>> /usr/local/var/lib/glusterd/nfs/nfs-server.vol
>>>
>>> I will get this for you when I can.  Thanks.
>>
>> If you can make it, that will be great help to understand the problem.
>>
>>
>> Rafi KC
>>
>>>
>>> Regards,
>>> Jon Heese
>>>
>>>> Rafi KC
>>>>>
>>>>> The actual I/O appears to be handled properly and I've seen no 
>>>>> further errors in the testing I've done so far.
>>>>>
>>>>>
>>>>> Thanks.
>>>>>
>>>>>
>>>>> Regards,
>>>>>
>>>>> Jon Heese
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------
>>>>> *From:* gluster-users-bounces at gluster.org 
>>>>> <gluster-users-bounces at gluster.org> on behalf of Jonathan Heese 
>>>>> <jheese at inetu.net>
>>>>> *Sent:* Friday, March 20, 2015 7:04 AM
>>>>> *To:* Mohammed Rafi K C
>>>>> *Cc:* gluster-users
>>>>> *Subject:* Re: [Gluster-users] I/O error on replicated volume
>>>>> Mohammed,
>>>>>
>>>>> Thanks very much for the reply.  I will try that and report back.
>>>>>
>>>>> Regards,
>>>>> Jon Heese
>>>>>
>>>>> On Mar 20, 2015, at 3:26 AM, "Mohammed Rafi K C" 
>>>>> <rkavunga at redhat.com <mailto:rkavunga at redhat.com>> wrote:
>>>>>
>>>>>>
>>>>>> On 03/19/2015 10:16 PM, Jonathan Heese wrote:
>>>>>>>
>>>>>>> Hello all,
>>>>>>>
>>>>>>> Does anyone else have any further suggestions for 
>>>>>>> troubleshooting this?
>>>>>>>
>>>>>>> To sum up: I have a 2 node 2 brick replicated volume, which 
>>>>>>> holds a handful of iSCSI image files which are mounted and 
>>>>>>> served up by tgtd (CentOS 6) to a handful of devices on a 
>>>>>>> dedicated iSCSI network. The most important iSCSI clients 
>>>>>>> (initiators) are four VMware ESXi 5.5 hosts that use the iSCSI 
>>>>>>> volumes as backing for their datastores for virtual machine storage.
>>>>>>>
>>>>>>> After a few minutes of sustained writing to the volume, I am 
>>>>>>> seeing a massive flood (over 1500 per second at times) of this 
>>>>>>> error in /var/log/glusterfs/mnt-gluster-disk.log:
>>>>>>>
>>>>>>> [2015-03-16 02:24:07.582801] W 
>>>>>>> [fuse-bridge.c:2242:fuse_writev_cbk] 0-glusterfs-fuse: 635358: 
>>>>>>> WRITE => -1 (Input/output error)
>>>>>>>
>>>>>>> When this happens, the ESXi box fails its write operation and 
>>>>>>> returns an error to the effect of “Unable to write data to 
>>>>>>> datastore”. I don’t see anything else in the supporting logs to 
>>>>>>> explain the root cause of the i/o errors.
>>>>>>>
>>>>>>> Any and all suggestions are appreciated.  Thanks.
>>>>>>>
>>>>>>
>>>>>> From the mount logs, i assume that your volume transport type is 
>>>>>> rdma. There are some known issues for rdma in 3.5.3, and the 
>>>>>> patch for to address those issues are already send to upstream 
>>>>>> [1]. From the logs, I'm not sure and it is hard to tell you 
>>>>>> whether this problem is something related to rdma transport or 
>>>>>> not. To make sure that the tcp transport is works well in this 
>>>>>> scenario, if possible can you try to reproduce the same using tcp 
>>>>>> type volumes. You can change the transport type of volume by 
>>>>>> doing the following step ( not recommended in normal use case).
>>>>>>
>>>>>> 1) unmount every client
>>>>>> 2) stop the volume
>>>>>> 3) run gluster volume set volname config.transport tcp
>>>>>> 4) start the volume again
>>>>>> 5) mount the clients
>>>>>>
>>>>>> [1] : http://goo.gl/2PTL61
>>>>>>
>>>>>> Regards
>>>>>> Rafi KC
>>>>>>
>>>>>>> /Jon Heese/
>>>>>>> /Systems Engineer/
>>>>>>> *INetU Managed Hosting*
>>>>>>> P: 610.266.7441 x 261
>>>>>>> F: 610.266.7434
>>>>>>> www.inetu.net <https://www.inetu.net/>
>>>>>>>
>>>>>>> /** This message contains confidential information, which also 
>>>>>>> may be privileged, and is intended only for the person(s) 
>>>>>>> addressed above. Any unauthorized use, distribution, copying or 
>>>>>>> disclosure of confidential and/or privileged information is 
>>>>>>> strictly prohibited. If you have received this communication in 
>>>>>>> error, please erase all copies of the message and its 
>>>>>>> attachments and notify the sender immediately via reply e-mail. **/
>>>>>>>
>>>>>>> *From:*Jonathan Heese
>>>>>>> *Sent:* Tuesday, March 17, 2015 12:36 PM
>>>>>>> *To:* 'Ravishankar N'; gluster-users at gluster.org
>>>>>>> *Subject:* RE: [Gluster-users] I/O error on replicated volume
>>>>>>>
>>>>>>> Ravi,
>>>>>>>
>>>>>>> The last lines in the mount log before the massive vomit of I/O 
>>>>>>> errors are from 22 minutes prior, and seem innocuous to me:
>>>>>>>
>>>>>>> [2015-03-16 01:37:07.126340] E 
>>>>>>> [client-handshake.c:1760:client_query_portmap_cbk] 
>>>>>>> 0-gluster_disk-client-0: failed to get the port number for 
>>>>>>> remote subvolume. Please run 'gluster volume status' on server 
>>>>>>> to see if brick process is running.
>>>>>>>
>>>>>>> [2015-03-16 01:37:07.126587] W [rdma.c:4273:gf_rdma_disconnect] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f) 
>>>>>>> [0x7fd9c557bccf] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) 
>>>>>>> [0x7fd9c557a995] 
>>>>>>> (-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea) 
>>>>>>> [0x7fd9c0d8fb9a]))) 0-gluster_disk-client-0: disconnect called 
>>>>>>> (peer:10.10.10.1:24008)
>>>>>>>
>>>>>>> [2015-03-16 01:37:07.126687] E 
>>>>>>> [client-handshake.c:1760:client_query_portmap_cbk] 
>>>>>>> 0-gluster_disk-client-1: failed to get the port number for 
>>>>>>> remote subvolume. Please run 'gluster volume status' on server 
>>>>>>> to see if brick process is running.
>>>>>>>
>>>>>>> [2015-03-16 01:37:07.126737] W [rdma.c:4273:gf_rdma_disconnect] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f) 
>>>>>>> [0x7fd9c557bccf] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) 
>>>>>>> [0x7fd9c557a995] 
>>>>>>> (-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea) 
>>>>>>> [0x7fd9c0d8fb9a]))) 0-gluster_disk-client-1: disconnect called 
>>>>>>> (peer:10.10.10.2:24008)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.730165] I 
>>>>>>> [rpc-clnt.c:1729:rpc_clnt_reconfig] 0-gluster_disk-client-0: 
>>>>>>> changing port to 49152 (from 0)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.730276] W [rdma.c:4273:gf_rdma_disconnect] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f) 
>>>>>>> [0x7fd9c557bccf] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) 
>>>>>>> [0x7fd9c557a995] 
>>>>>>> (-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea) 
>>>>>>> [0x7fd9c0d8fb9a]))) 0-gluster_disk-client-0: disconnect called 
>>>>>>> (peer:10.10.10.1:24008)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.739500] I 
>>>>>>> [rpc-clnt.c:1729:rpc_clnt_reconfig] 0-gluster_disk-client-1: 
>>>>>>> changing port to 49152 (from 0)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.739560] W [rdma.c:4273:gf_rdma_disconnect] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x13f) 
>>>>>>> [0x7fd9c557bccf] 
>>>>>>> (-->/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) 
>>>>>>> [0x7fd9c557a995] 
>>>>>>> (-->/usr/lib64/glusterfs/3.5.3/xlator/protocol/client.so(client_query_portmap_cbk+0x1ea) 
>>>>>>> [0x7fd9c0d8fb9a]))) 0-gluster_disk-client-1: disconnect called 
>>>>>>> (peer:10.10.10.2:24008)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.741883] I 
>>>>>>> [client-handshake.c:1677:select_server_supported_programs] 
>>>>>>> 0-gluster_disk-client-0: Using Program GlusterFS 3.3, Num 
>>>>>>> (1298437), Version (330)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.744524] I 
>>>>>>> [client-handshake.c:1462:client_setvolume_cbk] 
>>>>>>> 0-gluster_disk-client-0: Connected to 10.10.10.1:49152, attached 
>>>>>>> to remote volume '/bricks/brick1'.
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.744537] I 
>>>>>>> [client-handshake.c:1474:client_setvolume_cbk] 
>>>>>>> 0-gluster_disk-client-0: Server and Client lk-version numbers 
>>>>>>> are not same, reopening the fds
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.744566] I [afr-common.c:4267:afr_notify] 
>>>>>>> 0-gluster_disk-replicate-0: Subvolume 'gluster_disk-client-0' 
>>>>>>> came back up; going online.
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.744627] I 
>>>>>>> [client-handshake.c:450:client_set_lk_version_cbk] 
>>>>>>> 0-gluster_disk-client-0: Server lk version = 1
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.753037] I 
>>>>>>> [client-handshake.c:1677:select_server_supported_programs] 
>>>>>>> 0-gluster_disk-client-1: Using Program GlusterFS 3.3, Num 
>>>>>>> (1298437), Version (330)
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.755657] I 
>>>>>>> [client-handshake.c:1462:client_setvolume_cbk] 
>>>>>>> 0-gluster_disk-client-1: Connected to 10.10.10.2:49152, attached 
>>>>>>> to remote volume '/bricks/brick1'.
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.755676] I 
>>>>>>> [client-handshake.c:1474:client_setvolume_cbk] 
>>>>>>> 0-gluster_disk-client-1: Server and Client lk-version numbers 
>>>>>>> are not same, reopening the fds
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.761945] I 
>>>>>>> [fuse-bridge.c:5016:fuse_graph_setup] 0-fuse: switched to graph 0
>>>>>>>
>>>>>>> [2015-03-16 01:37:10.762144] I 
>>>>>>> [client-handshake.c:450:client_set_lk_version_cbk] 
>>>>>>> 0-gluster_disk-client-1: Server lk version = 1
>>>>>>>
>>>>>>> [*2015-03-16 01:37:10.762279*] I [fuse-bridge.c:3953:fuse_init] 
>>>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 
>>>>>>> 7.22 kernel 7.14
>>>>>>>
>>>>>>> [*2015-03-16 01:59:26.098670*] W 
>>>>>>> [fuse-bridge.c:2242:fuse_writev_cbk] 0-glusterfs-fuse: 292084: 
>>>>>>> WRITE => -1 (Input/output error)
>>>>>>>
>>>>>>>>>>>>>>
>>>>>>> I’ve seen no indication of split-brain on any files at any point 
>>>>>>> in this (ever since downdating from 3.6.2 to 3.5.3, which is 
>>>>>>> when this particular issue started):
>>>>>>>
>>>>>>> [root at duke gfapi-module-for-linux-target-driver-]# gluster v 
>>>>>>> heal gluster_disk info
>>>>>>>
>>>>>>> Brick duke.jonheese.local:/bricks/brick1/
>>>>>>>
>>>>>>> Number of entries: 0
>>>>>>>
>>>>>>> Brick duchess.jonheese.local:/bricks/brick1/
>>>>>>>
>>>>>>> Number of entries: 0
>>>>>>>
>>>>>>> Thanks.
>>>>>>>
>>>>>>> /Jon Heese/
>>>>>>> /Systems Engineer/
>>>>>>> *INetU Managed Hosting*
>>>>>>> P: 610.266.7441 x 261
>>>>>>> F: 610.266.7434
>>>>>>> www.inetu.net <https://www.inetu.net/>
>>>>>>>
>>>>>>> /** This message contains confidential information, which also 
>>>>>>> may be privileged, and is intended only for the person(s) 
>>>>>>> addressed above. Any unauthorized use, distribution, copying or 
>>>>>>> disclosure of confidential and/or privileged information is 
>>>>>>> strictly prohibited. If you have received this communication in 
>>>>>>> error, please erase all copies of the message and its 
>>>>>>> attachments and notify the sender immediately via reply e-mail. **/
>>>>>>>
>>>>>>> *From:*Ravishankar N [mailto:ravishankar at redhat.com]
>>>>>>> *Sent:* Tuesday, March 17, 2015 12:35 AM
>>>>>>> *To:* Jonathan Heese; gluster-users at gluster.org 
>>>>>>> <mailto:gluster-users at gluster.org>
>>>>>>> *Subject:* Re: [Gluster-users] I/O error on replicated volume
>>>>>>>
>>>>>>> On 03/17/2015 02:14 AM, Jonathan Heese wrote:
>>>>>>>
>>>>>>>     Hello,
>>>>>>>
>>>>>>>     So I resolved my previous issue with split-brains and the
>>>>>>>     lack of self-healing by dropping my installed glusterfs*
>>>>>>>     packages from 3.6.2 to 3.5.3, but now I've picked up a new
>>>>>>>     issue, which actually makes normal use of the volume
>>>>>>>     practically impossible.
>>>>>>>
>>>>>>>     A little background for those not already paying close
>>>>>>>     attention:
>>>>>>>     I have a 2 node 2 brick replicating volume whose purpose in
>>>>>>>     life is to hold iSCSI target files, primarily for use to
>>>>>>>     provide datastores to a VMware ESXi cluster.  The plan is to
>>>>>>>     put a handful of image files on the Gluster volume, mount
>>>>>>>     them locally on both Gluster nodes, and run tgtd on both,
>>>>>>>     pointed to the image files on the mounted gluster volume.
>>>>>>>     Then the ESXi boxes will use multipath (active/passive)
>>>>>>>     iSCSI to connect to the nodes, with automatic failover in
>>>>>>>     case of planned or unplanned downtime of the Gluster nodes.
>>>>>>>
>>>>>>>     In my most recent round of testing with 3.5.3, I'm seeing a
>>>>>>>     massive failure to write data to the volume after about 5-10
>>>>>>>     minutes, so I've simplified the scenario a bit (to minimize
>>>>>>>     the variables) to: both Gluster nodes up, only one node
>>>>>>>     (duke) mounted and running tgtd, and just regular (single
>>>>>>>     path) iSCSI from a single ESXi server.
>>>>>>>
>>>>>>>     About 5-10 minutes into migration a VM onto the test
>>>>>>>     datastore, /var/log/messages on duke gets blasted with a ton
>>>>>>>     of messages exactly like this:
>>>>>>>
>>>>>>>     Mar 15 22:24:06 duke tgtd: bs_rdwr_request(180) io error
>>>>>>>     0x1781e00 2a -1 512 22971904, Input/output error
>>>>>>>
>>>>>>>     And /var/log/glusterfs/mnt-gluster_disk.log gets blased with
>>>>>>>     a ton of messages exactly like this:
>>>>>>>
>>>>>>>     [2015-03-16 02:24:07.572279] W
>>>>>>>     [fuse-bridge.c:2242:fuse_writev_cbk] 0-glusterfs-fuse:
>>>>>>>     635299: WRITE => -1 (Input/output error)
>>>>>>>
>>>>>>>
>>>>>>> Are there any messages in the mount log from AFR about 
>>>>>>> split-brain just before the above line appears?
>>>>>>> Does `gluster v heal <VOLNAME> info` show any files? Performing 
>>>>>>> I/O on files that are in split-brain fail with EIO.
>>>>>>>
>>>>>>> -Ravi
>>>>>>>
>>>>>>>     And the write operation from VMware's side fails as soon as
>>>>>>>     these messages start.
>>>>>>>
>>>>>>>     I don't see any other errors (in the log files I know of)
>>>>>>>     indicating the root cause of these i/o errors.  I'm sure
>>>>>>>     that this is not enough information to tell what's going on,
>>>>>>>     but can anyone help me figure out what to look at next to
>>>>>>>     figure this out?
>>>>>>>
>>>>>>>     I've also considered using Dan Lambright's libgfapi gluster
>>>>>>>     module for tgtd (or something similar) to avoid going
>>>>>>>     through FUSE, but I'm not sure whether that would be
>>>>>>>     irrelevant to this problem, since I'm not 100% sure if it
>>>>>>>     lies in FUSE or elsewhere.
>>>>>>>
>>>>>>>     Thanks!
>>>>>>>
>>>>>>>     /Jon Heese/
>>>>>>>     /Systems Engineer/
>>>>>>>     *INetU Managed Hosting*
>>>>>>>     P: 610.266.7441 x 261
>>>>>>>     F: 610.266.7434
>>>>>>>     www.inetu.net <https://www.inetu.net/>
>>>>>>>
>>>>>>>     /** This message contains confidential information, which
>>>>>>>     also may be privileged, and is intended only for the
>>>>>>>     person(s) addressed above. Any unauthorized use,
>>>>>>>     distribution, copying or disclosure of confidential and/or
>>>>>>>     privileged information is strictly prohibited. If you have
>>>>>>>     received this communication in error, please erase all
>>>>>>>     copies of the message and its attachments and notify the
>>>>>>>     sender immediately via reply e-mail. **/
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>     _______________________________________________
>>>>>>>
>>>>>>>     Gluster-users mailing list
>>>>>>>
>>>>>>>     Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>>>>>>>
>>>>>>>     http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Gluster-users mailing list
>>>>>>> Gluster-users at gluster.org
>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>
>>>>
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150326/126ffb8e/attachment.html>


More information about the Gluster-users mailing list