[Gluster-users] DF reports incorrect sizes
Stas Oskin
stas.oskin at gmail.com
Thu Mar 26 15:02:01 UTC 2009
Hi.
Same as advised on this list, see below.
By the way, I restarted both the clients and servers, and the reported size
is still the same.
Whichever it is, it stuck quite persistently :).
server.vol
volume home1
type storage/posix # POSIX FS translator
option directory /media/storage # Export this directory
end-volume
volume posix-locks-home1
type features/posix-locks
option mandatory-locks on
subvolumes home1
end-volume
### Add network serving capability to above home.
volume server
type protocol/server
option transport-type tcp
subvolumes posix-locks-home1
option auth.addr.posix-locks-home1.allow * # Allow access to "home1" volume
end-volume
client.vol
## Reference volume "home1" from remote server
volume home1
type protocol/client
option transport-type tcp/client
option remote-host 192.168.253.41 # IP address of remote host
option remote-subvolume posix-locks-home1 # use home1 on remote host
option transport-timeout 10 # value in seconds; it should be set
relatively low
end-volume
## Reference volume "home2" from remote server
volume home2
type protocol/client
option transport-type tcp/client
option remote-host 192.168.253.42 # IP address of remote host
option remote-subvolume posix-locks-home1 # use home1 on remote host
option transport-timeout 10 # value in seconds; it should be set
relatively low
end-volume
volume home
type cluster/afr
option metadata-self-heal on
subvolumes home1 home2
end-volume
volume writebehind
type performance/write-behind
option aggregate-size 128KB
option window-size 1MB
subvolumes home
end-volume
volume cache
type performance/io-cache
option cache-size 512MB
subvolumes writebehind
end-volume
Regards.
2009/3/26 Vikas Gorur <vikas at zresearch.com>
> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>:
> > Hi.
> >
> > We erased all the data from our mount point, but the df still reports
> > it's almost full:
> >
> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs
> >
> > Running du either in the mount point, or in the back-end directory,
> > reports 914M.
> >
> > How do we get the space back?
>
> What is your client and server configuration?
>
> Vikas
> --
> Engineer - Z Research
> http://gluster.com/
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090326/b3f08ef3/attachment.html>
More information about the Gluster-users
mailing list