[Gluster-users] DF reports incorrect sizes

Stas Oskin stas.oskin at gmail.com
Thu Mar 26 15:02:01 UTC 2009


Hi.

Same as advised on this list, see below.

By the way, I restarted both the clients and servers, and the reported size
is still the same.
Whichever it is, it stuck quite persistently :).

server.vol

volume home1
 type storage/posix                   # POSIX FS translator
 option directory /media/storage        # Export this directory
end-volume

volume posix-locks-home1
 type features/posix-locks
 option mandatory-locks on
 subvolumes home1
end-volume

### Add network serving capability to above home.
volume server
 type protocol/server
 option transport-type tcp
 subvolumes posix-locks-home1
 option auth.addr.posix-locks-home1.allow * # Allow access to "home1" volume
end-volume


client.vol

## Reference volume "home1" from remote server
volume home1
 type protocol/client
 option transport-type tcp/client
 option remote-host 192.168.253.41      # IP address of remote host
 option remote-subvolume posix-locks-home1     # use home1 on remote host
 option transport-timeout 10           # value in seconds; it should be set
relatively low
end-volume

## Reference volume "home2" from remote server
volume home2
 type protocol/client
 option transport-type tcp/client
 option remote-host 192.168.253.42      # IP address of remote host
 option remote-subvolume posix-locks-home1     # use home1 on remote host
 option transport-timeout 10           # value in seconds; it should be set
relatively low
end-volume

volume home
 type cluster/afr
 option metadata-self-heal on
 subvolumes home1 home2
end-volume

volume writebehind
  type performance/write-behind
  option aggregate-size 128KB
  option window-size 1MB
  subvolumes home
end-volume

volume cache
  type performance/io-cache
  option cache-size 512MB
  subvolumes writebehind
end-volume


Regards.

2009/3/26 Vikas Gorur <vikas at zresearch.com>

> 2009/3/26 Stas Oskin <stas.oskin at gmail.com>:
> > Hi.
> >
> > We erased all the data from our mount point, but the df still reports
> > it's almost full:
> >
> > glusterfs 31G 27G 2.5G 92% /mnt/glusterfs
> >
> > Running du either in the mount point, or in the back-end directory,
> > reports 914M.
> >
> > How do we get the space back?
>
> What is your client and server configuration?
>
> Vikas
> --
> Engineer - Z Research
> http://gluster.com/
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090326/b3f08ef3/attachment.html>


More information about the Gluster-users mailing list