[Gluster-users] Quorum For replica 3 storage

Punit Dambiwal hypunit at gmail.com
Mon Jul 20 02:44:37 UTC 2015


HI Atin,

Apologies for the delay response...

1. When you added the brick was the command successful?
>> Yes..it was successful..
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?
>> I will reproduce the issue again and update you..

On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj at redhat.com>
wrote:

>
>
> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> > Hi Sathees,
> >
> > With 3 bricks i can get the gluster volume status....but after added more
> > bricks....can not get gluster volume status....
> The information is still incomplete in respect to analyze the problem.
> Further questions:
>
> 1. When you added the brick was the command successful?
> 2. If volume status is failing what's output its throwing in the console
> and how about the glusterd log?
>
> ~Atin
> >
> > On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar at redhat.com>
> wrote:
> >
> >> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
> >>
> >>>
> >>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
> >>>
> >>>> Hi,
> >>>>
> >>>> I have deployed one replica 3 storage...but i am facing some issue
> with
> >>>> quorum...
> >>>>
> >>>> Let me elaborate more :-
> >>>>
> >>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
> >>>> RAID...Just JBOD...
> >>>> 2. Gluster working fine when just add 3 HDD as below :-
> >>>>
> >>>> B HDD from server 1
> >>>> B HDD from server 2
> >>>> B HDD from server 3
> >>>>
> >>>> But when i add more bricks as below :-
> >>>>
> >>>> -----------------------
> >>>> [root at stor1 ~]# gluster volume info
> >>>>
> >>>> Volume Name: 3TB
> >>>> Type: Distributed-Replicate
> >>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
> >>>> Status: Stopped
> >>>> Number of Bricks: 5 x 3 = 15
> >>>> Transport-type: tcp
> >>>> Bricks:
> >>>> Brick1: stor1:/bricks/b/vol1
> >>>> Brick2: stor2:/bricks/b/vol1
> >>>> Brick3: stor3:/bricks/b/vol1
> >>>> Brick4: stor1:/bricks/c/vol1
> >>>> Brick5: stor2:/bricks/c/vol1
> >>>> Brick6: stor3:/bricks/c/vol1
> >>>> Brick7: stor1:/bricks/d/vol1
> >>>> Brick8: stor2:/bricks/d/vol1
> >>>> Brick9: stor3:/bricks/d/vol1
> >>>> Brick10: stor1:/bricks/e/vol1
> >>>> Brick11: stor2:/bricks/e/vol1
> >>>> Brick12: stor3:/bricks/e/vol1
> >>>> Brick13: stor1:/bricks/f/vol1
> >>>> Brick14: stor2:/bricks/f/vol1
> >>>> Brick15: stor3:/bricks/f/vol1
> >>>> Options Reconfigured:
> >>>> nfs.disable: off
> >>>> user.cifs: enable
> >>>> auth.allow: *
> >>>> performance.quick-read: off
> >>>> performance.read-ahead: off
> >>>> performance.io-cache: off
> >>>> performance.stat-prefetch: off
> >>>> cluster.eager-lock: enable
> >>>> network.remote-dio: enable
> >>>> cluster.quorum-type: auto
> >>>> cluster.server-quorum-type: server
> >>>> storage.owner-uid: 36
> >>>> storage.owner-gid: 36
> >>>> --------------------------------
> >>>>
> >>>> Brick added successfully without any error but after 1 min quorum
> failed
> >>>> and gluster stop working...
> >>>>
> >>> Punit,
> >>
> >> And what do you mean by quorum failed ?
> >> What is effect that you are seeing ?
> >> Could you provide output of 'gluster volume status' as well ?
> >>
> >> -- Sathees
> >>
> >>
> >>  What do log files say?
> >>>
> >>>> Thanks,
> >>>> Punit
> >>>>
> >>>>
> >>>>
> >>>> _______________________________________________
> >>>> Gluster-users mailing list
> >>>> Gluster-users at gluster.org
> >>>> http://www.gluster.org/mailman/listinfo/gluster-users
> >>>>
> >>>>
> >>
> >
>
> --
> ~Atin
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/76000897/attachment.html>


More information about the Gluster-users mailing list