[Gluster-users] Getting timedout error while rebalancing

Atin Mukherjee amukherj at redhat.com
Wed Feb 6 08:31:36 UTC 2019


On Tue, Feb 5, 2019 at 8:43 PM Nithya Balachandran <nbalacha at redhat.com>
wrote:

>
>
> On Tue, 5 Feb 2019 at 17:26, deepu srinivasan <sdeepugd at gmail.com> wrote:
>
>> HI Nithya
>> We have a test gluster setup.We are testing the rebalancing option of
>> gluster. So we started the volume which have 1x3 brick with some data on it
>> .
>> command : gluster volume create test-volume replica 3
>> 192.168.xxx.xx1:/home/data/repl 192.168.xxx.xx2:/home/data/repl
>> 192.168.xxx.xx3:/home/data/repl.
>>
>> Now we tried to expand the cluster storage by adding three more bricks.
>> command : gluster volume add-brick test-volume 192.168.xxx.xx4:/home/data/repl
>> 192.168.xxx.xx5:/home/data/repl 192.168.xxx.xx6:/home/data/repl
>>
>> So after the brick addition we tried to rebalance the layout and the data.
>> command : gluster volume rebalance test-volume fix-layout start.
>> The command exited with status "Error : Request timed out".
>>
>
> This sounds like an error in the cli or glusterd. Can you send the
> glusterd.log from the node on which you ran the command?
>

It seems to me that glusterd took more than 120 seconds to process the
command and hence cli timed out. We can confirm the same by checking the
status of the rebalance below which indicates rebalance did kick in and
eventually completed. We need to understand why did it take such longer, so
please pass on the cli and glusterd log from all the nodes as Nithya
requested for.


> regards,
> Nithya
>
>>
>> After the failure of the command, we tried to view the status of the
>> command and it is something like this :
>>
>>                                     Node Rebalanced-files          size
>>     scanned      failures       skipped               status  run time
>> in h:m:s
>>
>>                                ---------      -----------   -----------
>> -----------   -----------   -----------         ------------
>> --------------
>>
>>                                localhost               41        41.0MB
>>         8200             0             0            completed
>> 0:00:09
>>
>>                          192.168.xxx.xx4               79        79.0MB
>>         8231             0             0            completed
>> 0:00:12
>>
>>                          192.168.xxx.xx6               58        58.0MB
>>         8281             0             0            completed
>> 0:00:10
>>
>>                          192.168.xxx.xx2              136       136.0MB
>>         8566             0           136            completed
>> 0:00:07
>>
>>                          192.168.xxx.xx4              129       129.0MB
>>         8566             0           129            completed
>> 0:00:07
>>
>>                          192.168.xxx.xx6              201       201.0MB
>>         8566             0           201            completed
>> 0:00:08
>>
>> Is the rebalancing option working fine? Why did gluster  throw the error
>> saying that "Error : Request timed out"?
>> .On Tue, Feb 5, 2019 at 4:23 PM Nithya Balachandran <nbalacha at redhat.com>
>> wrote:
>>
>>> Hi,
>>> Please provide the exact step at which you are seeing the error. It
>>> would be ideal if you could copy-paste the command and the error.
>>>
>>> Regards,
>>> Nithya
>>>
>>>
>>>
>>> On Tue, 5 Feb 2019 at 15:24, deepu srinivasan <sdeepugd at gmail.com>
>>> wrote:
>>>
>>>> HI everyone. I am getting "Error : Request timed out " while doing
>>>> rebalance . I have aded new bricks to my replicated volume.i.e. First it
>>>> was 1x3 volume and added three more bricks to make it
>>>> distributed-replicated volume(2x3) . What should i do for the timeout error
>>>> ?
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190206/e425a015/attachment.html>


More information about the Gluster-users mailing list