[Gluster-users] Quorum For replica 3 storage

Ramesh Nachimuthu rnachimu at redhat.com
Mon Jul 20 09:12:32 UTC 2015



On 07/20/2015 12:44 PM, Punit Dambiwal wrote:
> Hi Atin,
>
> Earlier i was adding the brick and create the volume through ovirt 
> itself....this time i did it through commandline...i added all 15 
> bricks 5*3 through command line....
>
> It seems before optimize for virt storage it all ok...but after that 
> one quorum start and it failed the whole...
>

"Optimize for Virt" action in oVirt sets the following volume options.

group = > virt
storage.owner-uid => 36
storage.owner-gid => 36
server.allow-insecure => on
network.ping-timeout => 10

Regards,
Ramesh

> On Mon, Jul 20, 2015 at 3:07 PM, Atin Mukherjee <amukherj at redhat.com 
> <mailto:amukherj at redhat.com>> wrote:
>
>     In one of your earlier email you mentioned that after adding a brick
>     volume status stopped working. Can you point me to the glusterd
>     log for
>     that transaction?
>
>     ~Atin
>
>     On 07/20/2015 12:11 PM, Punit Dambiwal wrote:
>     > Hi Atin,
>     >
>     > Please find the below details :-
>     >
>     > [image: Inline image 1]
>     >
>     > [image: Inline image 2]
>     >
>     > Now when i set the optimize for the virt storage under ovirt and
>     restart
>     > glusterd service on any node...it start failing the quorum..
>     >
>     > [image: Inline image 3]
>     >
>     > [image: Inline image 4]
>     >
>     > Thanks,
>     > Punit
>     >
>     > On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit at gmail.com <mailto:hypunit at gmail.com>> wrote:
>     >
>     >> HI Atin,
>     >>
>     >> Apologies for the delay response...
>     >>
>     >> 1. When you added the brick was the command successful?
>     >>>> Yes..it was successful..
>     >> 2. If volume status is failing what's output its throwing in
>     the console
>     >> and how about the glusterd log?
>     >>>> I will reproduce the issue again and update you..
>     >>
>     >> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee
>     <amukherj at redhat.com <mailto:amukherj at redhat.com>>
>     >> wrote:
>     >>
>     >>>
>     >>>
>     >>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
>     >>>> Hi Sathees,
>     >>>>
>     >>>> With 3 bricks i can get the gluster volume status....but
>     after added
>     >>> more
>     >>>> bricks....can not get gluster volume status....
>     >>> The information is still incomplete in respect to analyze the
>     problem.
>     >>> Further questions:
>     >>>
>     >>> 1. When you added the brick was the command successful?
>     >>> 2. If volume status is failing what's output its throwing in
>     the console
>     >>> and how about the glusterd log?
>     >>>
>     >>> ~Atin
>     >>>>
>     >>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN
>     <sasundar at redhat.com <mailto:sasundar at redhat.com>>
>     >>> wrote:
>     >>>>
>     >>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>     >>>>>
>     >>>>>>
>     >>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>     >>>>>>
>     >>>>>>> Hi,
>     >>>>>>>
>     >>>>>>> I have deployed one replica 3 storage...but i am facing
>     some issue
>     >>> with
>     >>>>>>> quorum...
>     >>>>>>>
>     >>>>>>> Let me elaborate more :-
>     >>>>>>>
>     >>>>>>> 1. I have 3 node machines and every machine has 5
>     HDD(Bricks)...No
>     >>>>>>> RAID...Just JBOD...
>     >>>>>>> 2. Gluster working fine when just add 3 HDD as below :-
>     >>>>>>>
>     >>>>>>> B HDD from server 1
>     >>>>>>> B HDD from server 2
>     >>>>>>> B HDD from server 3
>     >>>>>>>
>     >>>>>>> But when i add more bricks as below :-
>     >>>>>>>
>     >>>>>>> -----------------------
>     >>>>>>> [root at stor1 ~]# gluster volume info
>     >>>>>>>
>     >>>>>>> Volume Name: 3TB
>     >>>>>>> Type: Distributed-Replicate
>     >>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>     >>>>>>> Status: Stopped
>     >>>>>>> Number of Bricks: 5 x 3 = 15
>     >>>>>>> Transport-type: tcp
>     >>>>>>> Bricks:
>     >>>>>>> Brick1: stor1:/bricks/b/vol1
>     >>>>>>> Brick2: stor2:/bricks/b/vol1
>     >>>>>>> Brick3: stor3:/bricks/b/vol1
>     >>>>>>> Brick4: stor1:/bricks/c/vol1
>     >>>>>>> Brick5: stor2:/bricks/c/vol1
>     >>>>>>> Brick6: stor3:/bricks/c/vol1
>     >>>>>>> Brick7: stor1:/bricks/d/vol1
>     >>>>>>> Brick8: stor2:/bricks/d/vol1
>     >>>>>>> Brick9: stor3:/bricks/d/vol1
>     >>>>>>> Brick10: stor1:/bricks/e/vol1
>     >>>>>>> Brick11: stor2:/bricks/e/vol1
>     >>>>>>> Brick12: stor3:/bricks/e/vol1
>     >>>>>>> Brick13: stor1:/bricks/f/vol1
>     >>>>>>> Brick14: stor2:/bricks/f/vol1
>     >>>>>>> Brick15: stor3:/bricks/f/vol1
>     >>>>>>> Options Reconfigured:
>     >>>>>>> nfs.disable: off
>     >>>>>>> user.cifs: enable
>     >>>>>>> auth.allow: *
>     >>>>>>> performance.quick-read: off
>     >>>>>>> performance.read-ahead: off
>     >>>>>>> performance.io-cache: off
>     >>>>>>> performance.stat-prefetch: off
>     >>>>>>> cluster.eager-lock: enable
>     >>>>>>> network.remote-dio: enable
>     >>>>>>> cluster.quorum-type: auto
>     >>>>>>> cluster.server-quorum-type: server
>     >>>>>>> storage.owner-uid: 36
>     >>>>>>> storage.owner-gid: 36
>     >>>>>>> --------------------------------
>     >>>>>>>
>     >>>>>>> Brick added successfully without any error but after 1 min
>     quorum
>     >>> failed
>     >>>>>>> and gluster stop working...
>     >>>>>>>
>     >>>>>> Punit,
>     >>>>>
>     >>>>> And what do you mean by quorum failed ?
>     >>>>> What is effect that you are seeing ?
>     >>>>> Could you provide output of 'gluster volume status' as well ?
>     >>>>>
>     >>>>> -- Sathees
>     >>>>>
>     >>>>>
>     >>>>>  What do log files say?
>     >>>>>>
>     >>>>>>> Thanks,
>     >>>>>>> Punit
>     >>>>>>>
>     >>>>>>>
>     >>>>>>>
>     >>>>>>> _______________________________________________
>     >>>>>>> Gluster-users mailing list
>     >>>>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>     >>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>     >>>>>>>
>     >>>>>>>
>     >>>>>
>     >>>>
>     >>>
>     >>> --
>     >>> ~Atin
>     >>>
>     >>
>     >>
>     >
>
>     --
>     ~Atin
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150720/897b3856/attachment.html>


More information about the Gluster-users mailing list