[Gluster-users] Add single server

Shyam srangana at redhat.com
Mon May 1 18:44:49 UTC 2017


On 05/01/2017 02:42 PM, Pranith Kumar Karampuri wrote:
>
>
> On Tue, May 2, 2017 at 12:07 AM, Shyam <srangana at redhat.com
> <mailto:srangana at redhat.com>> wrote:
>
>     On 05/01/2017 02:23 PM, Pranith Kumar Karampuri wrote:
>
>
>
>         On Mon, May 1, 2017 at 11:43 PM, Shyam <srangana at redhat.com
>         <mailto:srangana at redhat.com>
>         <mailto:srangana at redhat.com <mailto:srangana at redhat.com>>> wrote:
>
>             On 05/01/2017 02:00 PM, Pranith Kumar Karampuri wrote:
>
>                     Splitting the bricks need not be a post factum
>         decision, we can
>                     start with larger brick counts, on a given node/disk
>         count, and
>                     hence spread these bricks to newer nodes/bricks as
>         they are
>                 added.
>
>
>                 Let's say we have 1 disk, we format it with say XFS and that
>                 becomes a
>                 brick at the moment. Just curious, what will be the
>         relationship
>                 between
>                 brick to disk in this case(If we leave out LVM for this
>         example)?
>
>
>             I would assume the relation is brick to provided FS
>         directory (not
>             brick to disk, we do not control that at the moment, other than
>             providing best practices around the same).
>
>
>         Hmmm... as per my understanding, if we do this then 'df' I guess
>         will
>         report wrong values? available-size/free-size etc will be
>         counted more
>         than once?
>
>
>     This is true even today, if anyone uses 2 bricks from the same mount.
>
>
> That is the reason why documentation is the way it is as far as I can
> remember.
>
>
>
>     I forgot a converse though, we could take a disk and partition it
>     (LVM thinp volumes) and use each of those partitions as bricks,
>     avoiding the problem of df double counting. Further thinp will help
>     us expand available space to other bricks on the same disk, as we
>     destroy older bricks or create new ones to accommodate the moving
>     pieces (needs more careful thought though, but for sure is a
>     nightmare without thinp).
>
>     I am not so much a fan of large number of thinp partitions, so as
>     long as that is reasonably in control, we can possibly still use it.
>     The big advantage though is, we nuke a thinp volume when the brick
>     that uses that partition, moves out of that disk, and we get the
>     space back, rather than having or to something akin to rm -rf on the
>     backend to reclaim space.
>
>
> Other way to achieve the same is to leverage the quota functionality of
> counting how much size is used under a directory.

Yes, I think this is the direction to solve the 2 bricks on a single FS 
as well. Also, IMO, the weight of accounting at each directory level 
that quota brings in seems/is heavyweight to solve just *this* problem.

>
>
>
>
>
>
>
>             Today, gluster takes in a directory on host as a brick, and
>         assuming
>             we retain that, we would need to split this into multiple
>         sub-dirs
>             and use each sub-dir as a brick internally.
>
>             All these sub-dirs thus created are part of the same volume
>         (due to
>             our current snapshot mapping requirements).
>
>
>
>
>         --
>         Pranith
>
>
>
>
> --
> Pranith


More information about the Gluster-users mailing list