[Gluster-users] gluster remove-brick
Nithya Balachandran
nbalacha at redhat.com
Mon Feb 4 11:37:17 UTC 2019
Hi,
On Mon, 4 Feb 2019 at 16:39, mohammad kashif <kashif.alig at gmail.com> wrote:
> Hi Nithya
>
> Thanks for replying so quickly. It is very much appreciated.
>
> There are lots if " [No space left on device] " errors which I can not
> understand as there are much space on all of the nodes.
>
This means that Gluster could not find sufficient space for the file. Would
you be willing to share your rebalance log file?
Please provide the following information:
- The gluster version
- The gluster volume info for the volume
- How full are the individual bricks for the volume?
> A little bit of background will be useful in this case. I had cluster of
> seven nodes of varying capacity(73, 73, 73, 46, 46, 46,46 TB) . The
> cluster was almost 90% full so every node has almost 8 to 15 TB free
> space. I added two new nodes with 100TB each and ran fix-layout which
> completed successfully.
>
> After that I started remove-brick operation. I don't think that any point
> , any of the nodes were 100% full. Looking at my ganglia graph, there is
> minimum 5TB always available at every node.
>
> I was keeping an eye on remove-brick status and for very long time there
> was no failures and then at some point these 17000 failures appeared and it
> stayed like that.
>
> Thanks
>
> Kashif
>
>
>
>
>
> Let me explain a little bit of background.
>
>
> On Mon, Feb 4, 2019 at 5:09 AM Nithya Balachandran <nbalacha at redhat.com>
> wrote:
>
>> Hi,
>>
>> The status shows quite a few failures. Please check the rebalance logs to
>> see why that happened. We can decide what to do based on the errors.
>> Once you run a commit, the brick will no longer be part of the volume and
>> you will not be able to access those files via the client.
>> Do you have sufficient space on the remaining bricks for the files on the
>> removed brick?
>>
>> Regards,
>> Nithya
>>
>> On Mon, 4 Feb 2019 at 03:50, mohammad kashif <kashif.alig at gmail.com>
>> wrote:
>>
>>> Hi
>>>
>>> I have a pure distributed gluster volume with nine nodes and trying to
>>> remove one node, I ran
>>> gluster volume remove-brick atlasglust
>>> nodename:/glusteratlas/brick007/gv0 start
>>>
>>> It completed but with around 17000 failures
>>>
>>> Node Rebalanced-files size scanned failures
>>> skipped status run time in h:m:s
>>> --------- ----------- -----------
>>> ----------- ----------- ----------- ------------
>>> --------------
>>> nodename 4185858 27.5TB 6746030
>>> 17488 0 completed 405:15:34
>>>
>>> I can see that there is still 1.5 TB of data on the node which I was
>>> trying to remove.
>>>
>>> I am not sure what to do now? Should I run remove-brick command again
>>> so the files which has been failed can be tried again?
>>>
>>> or should I run commit first and then try to remove node again?
>>>
>>> Please advise as I don't want to remove files.
>>>
>>> Thanks
>>>
>>> Kashif
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190204/b819b55e/attachment.html>
More information about the Gluster-users
mailing list