[Gluster-users] Change underlying brick on node

David Gossage dgossage at carouselchecks.com
Mon Aug 8 21:56:04 UTC 2016


On Mon, Aug 8, 2016 at 4:37 PM, David Gossage <dgossage at carouselchecks.com>
wrote:

> On Mon, Aug 8, 2016 at 4:23 PM, Joe Julian <joe at julianfamily.org> wrote:
>
>>
>>
>> On 08/08/2016 01:39 PM, David Gossage wrote:
>>
>> So now that I have my cluster on 3.7.14 and sharded and working I am of
>> course looking for what to break next.
>>
>> Currently each of 3 nodes is on a 6 disk (WD Red 1TB) raidz6 (zil on
>> mirrored ssd), which I am thinking is more protection than I may need with
>> a 3 way replica.  I was going to one by one change them to basically raid10
>> letting it heal in between.
>>
>> Is best way to do that a systemctl stop glusterd, should I just kill the
>> brick process to simulate a brick dying, or is their an actual brick
>> maintenance command?
>>
>>
>> Just kill (-15) the brick process. That'll close the TCP connections and
>> the clients will just go right on functioning off the remaining replica.
>> When you format and recreate your filesystem, it'll be missing the
>> volume-id extended attributes so to start it you'll need to force it:
>>
>
Also could I just do this from different node?

getfattr  -n trusted.glusterfs.volume-id /srv/.bricks/www

Then on node with new raid10 backed disks

setfattr -n trusted.glusterfs.volume-id -v 'value_from_other_brick'
/srv/.bricks/www




>>    gluster volume start $volname start force
>>
>
> If I left volume started when brick process is killed and clients are
> still (in theory) connected to volume wouldn't that just give me an error
> that volume is already started?
>
>
> Likely I would shut down the volume and do downtime for this anyway though
> letting heals go on with VM's off.
>
>
>>
>> If /etc/glusterfs is unchanged and /var/lib/glusterd is unchanged will
>> doing a heal full after reboot or restarting glusterd take care of
>> everything if I recreate the expected brick path first?
>>
>>
>> Once started, perform a full heal to re-replicate.
>>
>>
>> Are the improvements in 3.8 for sharding significant enough I should
>> first look at updating to 3.8.2 when released in few days?
>>
>>
>> Yes.
>>
>>
>>
>> *David Gossage*
>> *Carousel Checks Inc. | System Administrator*
>> *Office* 708.613.2284
>>
>>
>> _______________________________________________
>> Gluster-users mailing listGluster-users at gluster.orghttp://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160808/d5f36b0d/attachment.html>


More information about the Gluster-users mailing list