[Gluster-users] volume not working after yum update - gluster 3.6.3
Atin Mukherjee
amukherj at redhat.com
Mon Aug 10 10:21:27 UTC 2015
On 08/10/2015 03:35 PM, Kingsley wrote:
> Sorry for the blind panic - restarting the volume seems to have fixed
> it.
>
> But then my next question - why is this necessary? Surely it undermines
> the whole point of a high availability system?
>
> Cheers,
> Kingsley.
>
> On Mon, 2015-08-10 at 10:53 +0100, Kingsley wrote:
>> Hi,
>>
>> We have a 4 way replicated volume using gluster 3.6.3 on CentOS 7.
>>
>> Over the weekend I did a yum update on each of the bricks in turn, but
>> now when clients (using fuse mounts) try to access the volume, it hangs.
What does mount log file say when you tried to access the volume? Can
you attach the mount log file?
>> Gluster itself wasn't updated (we've disabled that repo so that we keep
>> to 3.6.3 for now).
>>
>> This was what I did:
>>
>> * on first brick, "yum update"
>> * reboot brick
>> * watch "gluster volume status" on another brick and wait for it
>> to say all 4 bricks are online before proceeding to update the
>> next brick
>>
>> I was expecting the clients might pause 30 seconds while they notice a
>> brick is offline, but then recover.
>>
>> I've tried re-mounting clients, but that hasn't helped.
>>
>> I can't see much data in any of the log files.
>>
>> I've tried "gluster volume heal callrec" but it doesn't seem to have
>> helped.
>>
>> What shall I do next?
>>
>> I've pasted some stuff below in case any of it helps.
>>
>> Cheers,
>> Kingsley.
>>
>> [root at gluster1b-1 ~]# gluster volume info callrec
>>
>> Volume Name: callrec
>> Type: Replicate
>> Volume ID: a39830b7-eddb-4061-b381-39411274131a
>> Status: Started
>> Number of Bricks: 1 x 4 = 4
>> Transport-type: tcp
>> Bricks:
>> Brick1: gluster1a-1:/data/brick/callrec
>> Brick2: gluster1b-1:/data/brick/callrec
>> Brick3: gluster2a-1:/data/brick/callrec
>> Brick4: gluster2b-1:/data/brick/callrec
>> Options Reconfigured:
>> performance.flush-behind: off
>> [root at gluster1b-1 ~]#
>>
>>
>> [root at gluster1b-1 ~]# gluster volume status callrec
>> Status of volume: callrec
>> Gluster process Port Online Pid
>> ------------------------------------------------------------------------------
>> Brick gluster1a-1:/data/brick/callrec 49153 Y 6803
>> Brick gluster1b-1:/data/brick/callrec 49153 Y 2614
>> Brick gluster2a-1:/data/brick/callrec 49153 Y 2645
>> Brick gluster2b-1:/data/brick/callrec 49153 Y 4325
>> NFS Server on localhost 2049 Y 2769
>> Self-heal Daemon on localhost N/A Y 2789
>> NFS Server on gluster2a-1 2049 Y 2857
>> Self-heal Daemon on gluster2a-1 N/A Y 2814
>> NFS Server on 88.151.41.100 2049 Y 6833
>> Self-heal Daemon on 88.151.41.100 N/A Y 6824
>> NFS Server on gluster2b-1 2049 Y 4428
>> Self-heal Daemon on gluster2b-1 N/A Y 4387
>>
>> Task Status of Volume callrec
>> ------------------------------------------------------------------------------
>> There are no active volume tasks
>>
>> [root at gluster1b-1 ~]#
>>
>>
>> [root at gluster1b-1 ~]# gluster volume heal callrec info
>> Brick gluster1a-1.dns99.co.uk:/data/brick/callrec/
>> /to_process - Possibly undergoing heal
>>
>> Number of entries: 1
>>
>> Brick gluster1b-1.dns99.co.uk:/data/brick/callrec/
>> Number of entries: 0
>>
>> Brick gluster2a-1.dns99.co.uk:/data/brick/callrec/
>> /to_process - Possibly undergoing heal
>>
>> Number of entries: 1
>>
>> Brick gluster2b-1.dns99.co.uk:/data/brick/callrec/
>> Number of entries: 0
>>
>> [root at gluster1b-1 ~]#
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
--
~Atin
More information about the Gluster-users
mailing list