[Gluster-devel] Messup with peer status!!
Atin Mukherjee
amukherj at redhat.com
Wed Mar 16 05:53:02 UTC 2016
[1970-01-01 00:02:05.860202] D [MSGID: 0]
[store.c:501:gf_store_iter_new] 0-: Returning with 0
[1970-01-01 00:02:05.860518] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860545] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= type value = 2
[1970-01-01 00:02:05.860583] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860609] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= count value = 2
[1970-01-01 00:02:05.860650] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860676] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= status value = 1
[1970-01-01 00:02:05.860717] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860743] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= sub_count value = 2
[1970-01-01 00:02:05.860780] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860806] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= stripe_count value = 1
[1970-01-01 00:02:05.860842] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860868] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= replica_count value = 2
[1970-01-01 00:02:05.860905] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860931] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= disperse_count value = 0
[1970-01-01 00:02:05.860967] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.860994] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= redundancy_count value = 0
[1970-01-01 00:02:05.861030] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861056] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= version value = 42
[1970-01-01 00:02:05.861093] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861118] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= transport-type value = 0
[1970-01-01 00:02:05.861155] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861182] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= volume-id value = d86e215c-1710-4b33-8076-fbf8e075d3e7
[1970-01-01 00:02:05.861290] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861317] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= username value = db1d21cb-3feb-41da-88d0-2fc7a34cdb3a
[1970-01-01 00:02:05.861361] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861387] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= password value = df5bf0b7-34dd-4f0d-a01b-62d2b67aa8b0
[1970-01-01 00:02:05.861426] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861455] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= op-version value = 3
[1970-01-01 00:02:05.861503] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861530] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= client-op-version value = 3
[1970-01-01 00:02:05.861568] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861594] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= quota-version value = 0
[1970-01-01 00:02:05.861632] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861658] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= parent_volname value = N/A
[1970-01-01 00:02:05.861696] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861722] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= restored_from_snap value = 00000000-0000-0000-0000-000000000000
[1970-01-01 00:02:05.861762] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861788] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= snap-max-hard-limit value = 256
[1970-01-01 00:02:05.861825] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.861851] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= nfs.disable value = on
[1970-01-01 00:02:05.861940] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=nfs.disable,value:on
[1970-01-01 00:02:05.861978] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862004] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= network.ping-timeout value = 4
[1970-01-01 00:02:05.862039] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=network.ping-timeout,value:4
[1970-01-01 00:02:05.862077] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862104] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= performance.readdir-ahead value = on
[1970-01-01 00:02:05.862140] D [MSGID: 0]
[glusterd-store.c:2725:glusterd_store_update_volinfo] 0-management:
Parsed as Volume-set:key=performance.readdir-ahead,value:on
[1970-01-01 00:02:05.862178] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862217] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= brick-0 value = 10.32.0.48:-opt-lvmdir-c2-brick
[1970-01-01 00:02:05.862257] D [MSGID: 0]
[store.c:613:gf_store_iter_get_next] 0-: Returning with 0
[1970-01-01 00:02:05.862283] D [MSGID: 0]
[glusterd-store.c:2567:glusterd_store_update_volinfo] 0-management: key
= brick-1 value = 10.32.1.144:-opt-lvmdir-c2-brick
On 03/16/2016 11:04 AM, ABHISHEK PALIWAL wrote:
> Hi Atin,
>
> Please tell me the line number where you areseeing that glusterd has
> restored value from the disk files in Board B file.
>
> Regards,
> Abhishek
>
> On Tue, Mar 15, 2016 at 11:31 AM, ABHISHEK PALIWAL
> <abhishpaliwal at gmail.com <mailto:abhishpaliwal at gmail.com>> wrote:
>
>
>
> On Tue, Mar 15, 2016 at 11:10 AM, Atin Mukherjee
> <amukherj at redhat.com <mailto:amukherj at redhat.com>> wrote:
>
>
>
> On 03/15/2016 10:54 AM, ABHISHEK PALIWAL wrote:
> > Hi Atin,
> >
> > Is these files are ok? or you need some other files.
> I just started going through the log files you shared. I've few
> questions for you looking at the log:
> 1. Are you sure the log what you have provided from board B is
> post a
> reboot? If you claim that a reboot wipes of /var/lib/glusterd/
> then why
> am I seeing that glusterd has restored value from the disk files?
>
>
> Yes these logs from Board B after reboot. Could you please explain
> me the line number where you are seeing that glusterd has restored
> value from the disk files.
>
>
> 2. From the content of glusterd configurations which you shared
> earlier
> the peer UUIDs are 4bf982c0-b21b-415c-b870-e72f36c7f2e7,
> 4bf982c0-b21b-415c-b870-e72f36c7f2e7 002500/glusterd/peers &
> c6b64e36-76da-4e98-a616-48e0e52c7006 from 000300/glusterd/peers.
> They
> don't even exist in glusterd.log.
>
> Somehow I have a feeling that the sequence of log and configurations
> files you shared don't match!
>
>
> There is two UUID file present in 002500/glusterd/peers
> 1. 4bf982c0-b21b-415c-b870-e72f36c7f2e7
> Content of this file is:
> uuid=4bf982c0-b21b-415c-b870-e72f36c7f2e7
> state=10
> hostname1=10.32.0.48
> I have a question from where this UUID is coming?
>
> 2. 98a28041-f853-48ac-bee0-34c592eeb827
> Content of this file is:
> uuid=f4ebe3c5-b6a4-4795-98e0-732337f76faf //This uuid is belogs to
> 000300(10.32.0.48) board you can check this in both of the glusterd
> log file
> state=4 //what this state field display in this file?
> hostname1=10.32.0.48
>
>
> There is only one UUID file is present on 00030/glusterd/peers
>
> c6b64e36-76da-4e98-a616-48e0e52c7006 //This is the old UUID of the
> 002500 board before reboot
>
> content of this file is:
>
> uuid=267a92c3-fd28-4811-903c-c1d54854bda9 //This is new UUID
> generated by the 002500 board after reboot you can check this as
> well in glusterd file of 00030 board.
> state=3
> hostname1=10.32.1.144
>
>
> ~Atin
>
> >
> > Regards,
> > Abhishek
> >
> > On Mon, Mar 14, 2016 at 6:12 PM, ABHISHEK PALIWAL
> > <abhishpaliwal at gmail.com <mailto:abhishpaliwal at gmail.com>
> <mailto:abhishpaliwal at gmail.com
> <mailto:abhishpaliwal at gmail.com>>> wrote:
> >
> > You mean etc*-glusterd-*.log file from both of the boards?
> >
> > if yes please find the attachment for the same.
> >
> > On Mon, Mar 14, 2016 at 5:27 PM, Atin Mukherjee <amukherj at redhat.com <mailto:amukherj at redhat.com>
> > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>> wrote:
> >
> >
> >
> > On 03/14/2016 05:09 PM, ABHISHEK PALIWAL wrote:
> > > I am not getting you which glusterd directory you are asking. if you are
> > > asking about the /var/lib/glusterd directory then which I shared earlier
> > > is the same.
> > 1. Go to /var/log/glusterfs directory
> > 2. Look for glusterd log file
> > 3. attach the log
> > Do it for both the boards.
> > >
> > > I have two directories related to gluster
> > >
> > > 1. /var/log/glusterfs
> > > 2./var/lib/glusterd
> > >
> > > On Mon, Mar 14, 2016 at 4:12 PM, Atin Mukherjee <amukherj at redhat.com <mailto:amukherj at redhat.com>
> <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>
> > > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>
> <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>>> wrote:
> > >
> > >
> > >
> > > On 03/14/2016 03:59 PM, ABHISHEK PALIWAL wrote:
> > > > I have only these glusterd files available on the nodes
> > > Look for etc-*-glusterd*.log in /var/log/glusterfs, that represents the
> > > glusterd log file.
> > > >
> > > > Regards,
> > > > Abhishek
> > > >
> > > > On Mon, Mar 14, 2016 at 3:43 PM, Atin Mukherjee <amukherj at redhat.com <mailto:amukherj at redhat.com>
> <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>
> > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>
> <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>>
> > > > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>
> > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>>>> wrote:
> > > >
> > > >
> > > >
> > > > On 03/14/2016 02:18 PM, ABHISHEK PALIWAL
> wrote:
> > > > >
> > > > >
> > > > > On Mon, Mar 14, 2016 at 12:12 PM, Atin
> Mukherjee
> > <amukherj at redhat.com <mailto:amukherj at redhat.com>
> <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>
> > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>>
> > > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>
> > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>>>
> > > > > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>
> > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>
> > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>>
> > > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>
> > <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com> <mailto:amukherj at redhat.com
> <mailto:amukherj at redhat.com>>>>>> wrote:
> > > > >
> > > > >
> > > > >
> > > > > On 03/14/2016 10:52 AM, ABHISHEK
> PALIWAL wrote:
> > > > > > Hi Team,
> > > > > >
> > > > > > I am facing some issue with peer
> status and
> > because of
> > > that
> > > > remove-brick
> > > > > > on replica volume is getting failed.
> > > > > >
> > > > > > Here. is the scenario what I am
> doing with
> > gluster:
> > > > > >
> > > > > > 1. I have two boards A & B and
> gluster is
> > running on
> > > both of
> > > > the boards.
> > > > > > 2. On board I have created a
> replicated
> > volume with one
> > > > brick on each
> > > > > > board.
> > > > > > 3. Created one glusterfs mount
> point where
> > both of
> > > brick are
> > > > mounted.
> > > > > > 4. start the volume with
> nfs.disable=true.
> > > > > > 5. Till now everything is in sync
> between
> > both of bricks.
> > > > > >
> > > > > > Now when I manually plug-out the
> board B
> > from the slot and
> > > > plug-in it again.
> > > > > >
> > > > > > 1. After bootup the board B I have
> started
> > the glusted on
> > > > the board B.
> > > > > >
> > > > > > Following are the some gluster command
> > output on Board B
> > > > after the step 1.
> > > > > >
> > > > > > # gluster peer status
> > > > > > Number of Peers: 2
> > > > > >
> > > > > > Hostname: 10.32.0.48
> > > > > > Uuid:
> f4ebe3c5-b6a4-4795-98e0-732337f76faf
> > > > > > State: Accepted peer request
> (Connected)
> > > > > >
> > > > > > Hostname: 10.32.0.48
> > > > > > Uuid:
> 4bf982c0-b21b-415c-b870-e72f36c7f2e7
> > > > > > State: Peer is connected and Accepted
> > (Connected)
> > > > > >
> > > > > > Why this peer status is showing
> two peer with
> > > different UUID?
> > > > > GlusterD doesn't generate a new UUID
> on init
> > if it has
> > > already
> > > > generated
> > > > > an UUID earlier. This clearly
> indicates that
> > on reboot
> > > of board B
> > > > > content of /var/lib/glusterd were
> wiped off.
> > I've asked this
> > > > question to
> > > > > you multiple times that is it the case?
> > > > >
> > > > >
> > > > > Yes I am following the same which is
> mentioned in
> > the link:
> > > > >
> > > > >
> > > >
> > >
> >
> http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
> > > > >
> > > > > but why it is showing two peer enteries?
> > > > >
> > > > > >
> > > > > > # gluster volume info
> > > > > >
> > > > > > Volume Name: c_glusterfs
> > > > > > Type: Replicate
> > > > > > Volume ID:
> c11f1f13-64a0-4aca-98b5-91d609a4a18d
> > > > > > Status: Started
> > > > > > Number of Bricks: 1 x 2 = 2
> > > > > > Transport-type: tcp
> > > > > > Bricks:
> > > > > > Brick1:
> 10.32.0.48:/opt/lvmdir/c2/brick
> > > > > > Brick2:
> 10.32.1.144:/opt/lvmdir/c2/brick
> > > > > > Options Reconfigured:
> > > > > > performance.readdir-ahead: on
> > > > > > network.ping-timeout: 4
> > > > > > nfs.disable: on
> > > > > > # gluster volume heal c_glusterfs info
> > > > > > c_glusterfs: Not able to fetch
> volfile from
> > glusterd
> > > > > > Volume heal failed.
> > > > > > # gluster volume status c_glusterfs
> > > > > > Status of volume: c_glusterfs
> > > > > > Gluster process
> > TCP Port
> > > RDMA Port
> > > > > Online
> > > > > > Pid
> > > > > >
> > > > >
> > > >
> > >
> >
> ------------------------------------------------------------------------------
> > > > > >
> > > > > > Brick 10.32.1.144:/opt/lvmdir/c2/brick
> > N/A N/A
> > > > N
> > > > > > N/A
> > > > > > Self-heal Daemon on localhost
> > N/A N/A
> > > > Y
> > > > > > 3922
> > > > > >
> > > > > > Task Status of Volume c_glusterfs
> > > > > >
> > > > >
> > > >
> > >
> >
> ------------------------------------------------------------------------------
> > > > > >
> > > > > > There are no active volume tasks
> > > > > > --
> > > > > >
> > > > > > At the same time Board A have the
> following
> > gluster
> > > commands
> > > > outcome:
> > > > > >
> > > > > > # gluster peer status
> > > > > > Number of Peers: 1
> > > > > >
> > > > > > Hostname: 10.32.1.144
> > > > > > Uuid:
> c6b64e36-76da-4e98-a616-48e0e52c7006
> > > > > > State: Peer in Cluster (Connected)
> > > > > >
> > > > > > Why it is showing the older UUID
> of host
> > 10.32.1.144
> > > when this
> > > > > UUID has
> > > > > > been changed and new UUID is
> > > > 267a92c3-fd28-4811-903c-c1d54854bda9
> > > > > >
> > > > > >
> > > > > > # gluster volume heal c_glusterfs info
> > > > > > c_glusterfs: Not able to fetch
> volfile from
> > glusterd
> > > > > > Volume heal failed.
> > > > > > # gluster volume status c_glusterfs
> > > > > > Status of volume: c_glusterfs
> > > > > > Gluster process
> > TCP Port
> > > RDMA Port
> > > > > Online
> > > > > > Pid
> > > > > >
> > > > >
> > > >
> > >
> >
> ------------------------------------------------------------------------------
> > > > > >
> > > > > > Brick 10.32.0.48:/opt/lvmdir/c2/brick
> > 49169 0
> > > > Y
> > > > > > 2427
> > > > > > Brick 10.32.1.144:/opt/lvmdir/c2/brick
> > N/A N/A
> > > > N
> > > > > > N/A
> > > > > > Self-heal Daemon on localhost
> > N/A N/A
> > > > Y
> > > > > > 3388
> > > > > > Self-heal Daemon on 10.32.1.144
> > N/A N/A
> > > > Y
> > > > > > 3922
> > > > > >
> > > > > > Task Status of Volume c_glusterfs
> > > > > >
> > > > >
> > > >
> > >
> >
> ------------------------------------------------------------------------------
> > > > > >
> > > > > > There are no active volume tasks
> > > > > >
> > > > > > As you see in the "gluster volume
> status"
> > showing that
> > > Brick
> > > > > > "10.32.1.144:/opt/lvmdir/c2/brick " is
> > offline so We have
> > > > tried to
> > > > > > remove it but getting "volume
> remove-brick
> > c_glusterfs
> > > replica 1
> > > > > > 10.32.1.144:/opt/lvmdir/c2/brick
> force :
> > FAILED :
> > > Incorrect
> > > > brick
> > > > > > 10.32.1.144:/opt/lvmdir/c2/brick
> for volume
> > c_glusterfs"
> > > > error on the
> > > > > > Board A.
> > > > > >
> > > > > > Please reply on this post because I am
> > always getting
> > > this error
> > > > > in this
> > > > > > scenario.
> > > > > >
> > > > > > For more detail I am also adding
> the logs of
> > both of the
> > > > board which
> > > > > > having some manual created file in
> which you
> > can find the
> > > > output of
> > > > > > glulster command from both of the
> boards
> > > > > >
> > > > > > in logs
> > > > > > 00030 is board A
> > > > > > 00250 is board B.
> > > > > This attachment doesn't help much.
> Could you
> > attach full
> > > > glusterd log
> > > > > files from both the nodes?
> > > > > >
> > > > >
> > > > > inside this attachment you will found full
> > glusterd log file
> > > > > 00300/glusterd/ and 002500/glusterd/
> > > > No, that contains the configuration files.
> > > > >
> > > > > > Thanks in advance waiting for the
> reply.
> > > > > >
> > > > > > Regards,
> > > > > > Abhishek
> > > > > >
> > > > > >
> > > > > > Regards
> > > > > > Abhishek Paliwal
> > > > > >
> > > > > >
> > > > > >
> _______________________________________________
> > > > > > Gluster-devel mailing list
> > > > > > Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>
> > > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>
> > > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>>>
> > > > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>
> > > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>
> > > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>
> > <mailto:Gluster-devel at gluster.org
> <mailto:Gluster-devel at gluster.org>>>>>
> > > > > >
> > http://www.gluster.org/mailman/listinfo/gluster-devel
> > > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > Regards
> > > > > Abhishek Paliwal
> > > >
> > > >
> > > >
> > > >
> > > > --
> > > >
> > > >
> > > >
> > > >
> > > > Regards
> > > > Abhishek Paliwal
> > >
> > >
> > >
> > >
> > > --
> > >
> > >
> > >
> > >
> > > Regards
> > > Abhishek Paliwal
> >
> >
> >
> >
> > --
> >
> >
> >
> >
> > Regards
> > Abhishek Paliwal
> >
> >
> >
> >
> > --
> >
> >
> >
> >
> > Regards
> > Abhishek Paliwal
>
>
>
>
> --
>
>
>
>
> Regards
> Abhishek Paliwal
>
>
>
>
> --
>
>
>
>
> Regards
> Abhishek Paliwal
More information about the Gluster-devel
mailing list