[Gluster-users] Quorum For replica 3 storage

Atin Mukherjee amukherj at redhat.com
Mon Jul 13 03:46:06 UTC 2015



On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> Hi Sathees,
> 
> With 3 bricks i can get the gluster volume status....but after added more
> bricks....can not get gluster volume status....
The information is still incomplete in respect to analyze the problem.
Further questions:

1. When you added the brick was the command successful?
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?

~Atin
> 
> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar at redhat.com> wrote:
> 
>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>>
>>>
>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>>>
>>>> Hi,
>>>>
>>>> I have deployed one replica 3 storage...but i am facing some issue with
>>>> quorum...
>>>>
>>>> Let me elaborate more :-
>>>>
>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
>>>> RAID...Just JBOD...
>>>> 2. Gluster working fine when just add 3 HDD as below :-
>>>>
>>>> B HDD from server 1
>>>> B HDD from server 2
>>>> B HDD from server 3
>>>>
>>>> But when i add more bricks as below :-
>>>>
>>>> -----------------------
>>>> [root at stor1 ~]# gluster volume info
>>>>
>>>> Volume Name: 3TB
>>>> Type: Distributed-Replicate
>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>>>> Status: Stopped
>>>> Number of Bricks: 5 x 3 = 15
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: stor1:/bricks/b/vol1
>>>> Brick2: stor2:/bricks/b/vol1
>>>> Brick3: stor3:/bricks/b/vol1
>>>> Brick4: stor1:/bricks/c/vol1
>>>> Brick5: stor2:/bricks/c/vol1
>>>> Brick6: stor3:/bricks/c/vol1
>>>> Brick7: stor1:/bricks/d/vol1
>>>> Brick8: stor2:/bricks/d/vol1
>>>> Brick9: stor3:/bricks/d/vol1
>>>> Brick10: stor1:/bricks/e/vol1
>>>> Brick11: stor2:/bricks/e/vol1
>>>> Brick12: stor3:/bricks/e/vol1
>>>> Brick13: stor1:/bricks/f/vol1
>>>> Brick14: stor2:/bricks/f/vol1
>>>> Brick15: stor3:/bricks/f/vol1
>>>> Options Reconfigured:
>>>> nfs.disable: off
>>>> user.cifs: enable
>>>> auth.allow: *
>>>> performance.quick-read: off
>>>> performance.read-ahead: off
>>>> performance.io-cache: off
>>>> performance.stat-prefetch: off
>>>> cluster.eager-lock: enable
>>>> network.remote-dio: enable
>>>> cluster.quorum-type: auto
>>>> cluster.server-quorum-type: server
>>>> storage.owner-uid: 36
>>>> storage.owner-gid: 36
>>>> --------------------------------
>>>>
>>>> Brick added successfully without any error but after 1 min quorum failed
>>>> and gluster stop working...
>>>>
>>> Punit,
>>
>> And what do you mean by quorum failed ?
>> What is effect that you are seeing ?
>> Could you provide output of 'gluster volume status' as well ?
>>
>> -- Sathees
>>
>>
>>  What do log files say?
>>>
>>>> Thanks,
>>>> Punit
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>
> 

-- 
~Atin


More information about the Gluster-users mailing list