[Gluster-users] rebalance and volume commit hash
Shyam
srangana at redhat.com
Thu Jan 19 19:45:08 UTC 2017
On 01/17/2017 11:40 AM, Piotr Misiak wrote:
>
> 17 sty 2017 17:10 Jeff Darcy <jdarcy at redhat.com> napisał(a):
>>
>>> Do you think that is wise to run rebalance process manually on every
>>> brick with the actual commit hash value?
>>>
>>> I didn't do anything with bricks after previous rebalance run and I have
>>> cluster.weighted-rebalance=off.
>>>
>>> My problem is that I have a very big directory structure (millions of
>>> directories and files) and I haven't ever completed rebalance process
>>> once, because it will take I guess weeks or months. I'd like to speed it
>>> up a bit by not generating new commit hash for volume during new
>>> rebalance run. Then directories rebalanced in the previous run will be
>>> untouched during the new run. Is it possible?
>>
>> I'm pretty sure that trying to rebalance on each brick separately will
>> not work correctly. Rebalancing smaller pieces of the directory
>> hierarchy separately, by issuing the appropriate setxattr calls on them
>> instead of using the CLI, *might* work. Either way, I think the DHT
>> experts could provide a better answer.
>>
>
> Is it possible to start rebalancing from a particular subdirecory? Do you know how? It would be very useful for me.
Rebalancing a sub-directory is not supported.
Having said that, if we were to support it, then we could rebalance at a
sub-directory level and retain the volume commit hash as is. The commit
hash states truth about a directory and its immediate children, so
retaining the volume commit-hash when rebalancing a sub-dir is a
possibility (needs a little more thought, but at the outset is possible).
Further, for your case, it looks like rebalance takes a long time. So
one other option could be to just create the link-to files (or complete
a tree walk and lookup everything) and not move any data. This should be
faster than moving data around, and provides enough pre-condition for
the lookup-optimize to function. Of course, this will not balance the
data, so if are really looking to expand the volume size a full
rebalance maybe required.
May I request a couple of issues raised for these here [1], and based on
other DHT work we can check and see when we can accommodate this (or as
always patches are welcome).
[1] https://github.com/gluster/glusterfs/issues/new
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list