[Gluster-users] Fwd: DF reports incorrect sizes

Stas Oskin stas.oskin at gmail.com
Sun Mar 29 22:20:55 UTC 2009


---------- Forwarded message ----------
From: Stas Oskin <stas.oskin at gmail.com>
Date: 2009/3/29
Subject: Re: [Gluster-users] DF reports incorrect sizes
To: Vikas Gorur ‏vikas at zresearch.com‏


Hi.

After erasing all the data from my lab setup, and restarting all, it
happened again in less then 5 hours.

Here is what I see:

Client:
df -h: glusterfs              31G   29G     0 100% /mnt/media

Server 1:
df -h: /dev/hda4              31G   29G     0 100% /media

Server 2:
df -h: /dev/hda4              31G   20G  8.7G  70% /media

This means that again the server lost each other.

Perhaps it's related to the fact that the space go filled out.

Any idea how to diagnose it?


Regards.

2009/3/26 Stas Oskin <stas.oskin at gmail.com>

> Hi.
>
> It occurs that 1 of the 2 AFR volumes is not synchronized.
>
> Meaning erasing or creating files on mounts performed only on 1 node - but
> the free space reported from the both nodes.
>
> Any idea what's went wrong?
>
> Regards.
>
>
> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>
>
>> Hi.
>>
>> Same as advised on this list, see below.
>>
>> By the way, I restarted both the clients and servers, and the reported
>> size is still the same.
>> Whichever it is, it stuck quite persistently :).
>>
>> server.vol
>>
>> volume home1
>>  type storage/posix                   # POSIX FS translator
>>  option directory /media/storage        # Export this directory
>> end-volume
>>
>> volume posix-locks-home1
>>  type features/posix-locks
>>  option mandatory-locks on
>>  subvolumes home1
>> end-volume
>>
>> ### Add network serving capability to above home.
>> volume server
>>  type protocol/server
>>  option transport-type tcp
>>  subvolumes posix-locks-home1
>>  option auth.addr.posix-locks-home1.allow * # Allow access to "home1"
>> volume
>> end-volume
>>
>>
>> client.vol
>>
>> ## Reference volume "home1" from remote server
>> volume home1
>>  type protocol/client
>>  option transport-type tcp/client
>>  option remote-host 192.168.253.41      # IP address of remote host
>>  option remote-subvolume posix-locks-home1     # use home1 on remote host
>>  option transport-timeout 10           # value in seconds; it should be
>> set relatively low
>> end-volume
>>
>> ## Reference volume "home2" from remote server
>> volume home2
>>  type protocol/client
>>  option transport-type tcp/client
>>  option remote-host 192.168.253.42      # IP address of remote host
>>  option remote-subvolume posix-locks-home1     # use home1 on remote host
>>  option transport-timeout 10           # value in seconds; it should be
>> set relatively low
>> end-volume
>>
>> volume home
>>  type cluster/afr
>>  option metadata-self-heal on
>>  subvolumes home1 home2
>> end-volume
>>
>> volume writebehind
>>   type performance/write-behind
>>   option aggregate-size 128KB
>>   option window-size 1MB
>>   subvolumes home
>> end-volume
>>
>> volume cache
>>   type performance/io-cache
>>   option cache-size 512MB
>>   subvolumes writebehind
>> end-volume
>>
>>
>> Regards.
>>
>> 2009/3/26 Vikas Gorur <vikas at zresearch.com>
>>
>> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>:
>>> > Hi.
>>> >
>>> > We erased all the data from our mount point, but the df still reports
>>> > it's almost full:
>>> >
>>> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs
>>> >
>>> > Running du either in the mount point, or in the back-end directory,
>>> > reports 914M.
>>> >
>>> > How do we get the space back?
>>>
>>> What is your client and server configuration?
>>>
>>> Vikas
>>> --
>>> Engineer - Z Research
>>> http://gluster.com/
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090330/570576df/attachment.html>


More information about the Gluster-users mailing list