[Gluster-users] dht log entries in fuse client after successful expansion/rebalance

Tim Robinson Tim.Robinson at humedica.com
Fri Mar 9 18:11:51 UTC 2012

Thanks for breaking it down Jeff.  I can live with a remount on the rare
occasion we may resize just wanted to make sure my test environment was
exhibiting the expected behavior.


On 3/9/12 12:11 PM, "Jeff Darcy" <jdarcy at redhat.com> wrote:

>On Fri, 9 Mar 2012 15:44:48 +0000
>Tim Robinson <Tim.Robinson at humedica.com> wrote:
>> mismatching layouts for .../bench
>> subvol: bfd-replicate-2;
>>	inode layout - 0 - 0;
>>	disk layout - 2863311530 - 4294967295
>> mismatching layouts for .../bench
>> subvol: bfd-replicate-0;
>>	inode layout - 0 - 2147483646;
>>	disk layout - 0 - 1431655764
>> mismatching layouts .../bench
>> subvol: bfd-replicate-0;
>>	inode layout - 2147483647 - 4294967295;
>>	disk layout - 0 - 1431655764
>> mismatching layouts for .../debug
>> subvol: bfd-replicate-1;
>>	inode layout - 0 - 2147483646;
>>	disk layout - 1431655765 - 2863311529
>It looks like either your mail program or mine scrambled these lines a
>bit, so I've edited to show the most important bits.  Basically there
>are three messages for .../bench, corresponding to the three
>subvolumes, and then messages for .../debug on two out of three
>subvolumes.  I'd be worried if we saw these repeated for the same
>directory on the same subvolume, but (so far) that doesn't seem to be
>the case.  AFAICT the problem is just that the layout-revalidation code
>is being a lot more verbose than necessary.  This activity should tail
>off pretty quickly.
>> I have tried doing a self-heal on the client and rebalancing the
>> volume again but the messages persist.  After remounting the volume
>> the messages stop.  The rebalance reports success in adjusting the
>> layout and redistributing files and I can see from looking at the
>> bricks that pre-expansion files have been moved to the new brick and
>> post expansion files are going to all three but the excessive client
>> logging effects performance and would take up lots of space when
>> under heavy use.  Does anybody know what might be happening here and
>> how I can avoid these messages after expansion without remounting or
>> turning off/down client logging?
>We support dynamic reconfiguration on the servers, but not on the
>clients, so without remounting there's no good way to reduce the client
>log level.  I'd do it with gdb, but I don't think I can recommend that
>for the sort of environment where a remount would be precluded.

More information about the Gluster-users mailing list