[Gluster-users] Quorum For replica 3 storage

Punit Dambiwal hypunit at gmail.com
Mon Jul 20 07:14:00 UTC 2015


Hi Atin,

Earlier i was adding the brick and create the volume through ovirt
itself....this time i did it through commandline...i added all 15 bricks
5*3 through command line....

It seems before optimize for virt storage it all ok...but after that one
quorum start and it failed the whole...

On Mon, Jul 20, 2015 at 3:07 PM, Atin Mukherjee <amukherj at redhat.com> wrote:

> In one of your earlier email you mentioned that after adding a brick
> volume status stopped working. Can you point me to the glusterd log for
> that transaction?
>
> ~Atin
>
> On 07/20/2015 12:11 PM, Punit Dambiwal wrote:
> > Hi Atin,
> >
> > Please find the below details :-
> >
> > [image: Inline image 1]
> >
> > [image: Inline image 2]
> >
> > Now when i set the optimize for the virt storage under ovirt and restart
> > glusterd service on any node...it start failing the quorum..
> >
> > [image: Inline image 3]
> >
> > [image: Inline image 4]
> >
> > Thanks,
> > Punit
> >
> > On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit at gmail.com>
> wrote:
> >
> >> HI Atin,
> >>
> >> Apologies for the delay response...
> >>
> >> 1. When you added the brick was the command successful?
> >>>> Yes..it was successful..
> >> 2. If volume status is failing what's output its throwing in the console
> >> and how about the glusterd log?
> >>>> I will reproduce the issue again and update you..
> >>
> >> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj at redhat.com>
> >> wrote:
> >>
> >>>
> >>>
> >>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> >>>> Hi Sathees,
> >>>>
> >>>> With 3 bricks i can get the gluster volume status....but after added
> >>> more
> >>>> bricks....can not get gluster volume status....
> >>> The information is still incomplete in respect to analyze the problem.
> >>> Further questions:
> >>>
> >>> 1. When you added the brick was the command successful?
> >>> 2. If volume status is failing what's output its throwing in the
> console
> >>> and how about the glusterd log?
> >>>
> >>> ~Atin
> >>>>
> >>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar at redhat.com>
> >>> wrote:
> >>>>
> >>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
> >>>>>
> >>>>>>
> >>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
> >>>>>>
> >>>>>>> Hi,
> >>>>>>>
> >>>>>>> I have deployed one replica 3 storage...but i am facing some issue
> >>> with
> >>>>>>> quorum...
> >>>>>>>
> >>>>>>> Let me elaborate more :-
> >>>>>>>
> >>>>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
> >>>>>>> RAID...Just JBOD...
> >>>>>>> 2. Gluster working fine when just add 3 HDD as below :-
> >>>>>>>
> >>>>>>> B HDD from server 1
> >>>>>>> B HDD from server 2
> >>>>>>> B HDD from server 3
> >>>>>>>
> >>>>>>> But when i add more bricks as below :-
> >>>>>>>
> >>>>>>> -----------------------
> >>>>>>> [root at stor1 ~]# gluster volume info
> >>>>>>>
> >>>>>>> Volume Name: 3TB
> >>>>>>> Type: Distributed-Replicate
> >>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
> >>>>>>> Status: Stopped
> >>>>>>> Number of Bricks: 5 x 3 = 15
> >>>>>>> Transport-type: tcp
> >>>>>>> Bricks:
> >>>>>>> Brick1: stor1:/bricks/b/vol1
> >>>>>>> Brick2: stor2:/bricks/b/vol1
> >>>>>>> Brick3: stor3:/bricks/b/vol1
> >>>>>>> Brick4: stor1:/bricks/c/vol1
> >>>>>>> Brick5: stor2:/bricks/c/vol1
> >>>>>>> Brick6: stor3:/bricks/c/vol1
> >>>>>>> Brick7: stor1:/bricks/d/vol1
> >>>>>>> Brick8: stor2:/bricks/d/vol1
> >>>>>>> Brick9: stor3:/bricks/d/vol1
> >>>>>>> Brick10: stor1:/bricks/e/vol1
> >>>>>>> Brick11: stor2:/bricks/e/vol1
> >>>>>>> Brick12: stor3:/bricks/e/vol1
> >>>>>>> Brick13: stor1:/bricks/f/vol1
> >>>>>>> Brick14: stor2:/bricks/f/vol1
> >>>>>>> Brick15: stor3:/bricks/f/vol1
> >>>>>>> Options Reconfigured:
> >>>>>>> nfs.disable: off
> >>>>>>> user.cifs: enable
> >>>>>>> auth.allow: *
> >>>>>>> performance.quick-read: off
> >>>>>>> performance.read-ahead: off
> >>>>>>> performance.io-cache: off
> >>>>>>> performance.stat-prefetch: off
> >>>>>>> cluster.eager-lock: enable
> >>>>>>> network.remote-dio: enable
> >>>>>>> cluster.quorum-type: auto
> >>>>>>> cluster.server-quorum-type: server
> >>>>>>> storage.owner-uid: 36
> >>>>>>> storage.owner-gid: 36
> >>>>>>> --------------------------------
> >>>>>>>
> >>>>>>> Brick added successfully without any error but after 1 min quorum
> >>> failed
> >>>>>>> and gluster stop working...
> >>>>>>>
> >>>>>> Punit,
> >>>>>
> >>>>> And what do you mean by quorum failed ?
> >>>>> What is effect that you are seeing ?
> >>>>> Could you provide output of 'gluster volume status' as well ?
> >>>>>
> >>>>> -- Sathees
> >>>>>
> >>>>>
> >>>>>  What do log files say?
> >>>>>>
> >>>>>>> Thanks,
> >>>>>>> Punit
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> _______________________________________________
> >>>>>>> Gluster-users mailing list
> >>>>>>> Gluster-users at gluster.org
> >>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
> >>>>>>>
> >>>>>>>
> >>>>>
> >>>>
> >>>
> >>> --
> >>> ~Atin
> >>>
> >>
> >>
> >
>
> --
> ~Atin
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/2a027187/attachment.html>


More information about the Gluster-users mailing list