[Gluster-users] Add single server
Pranith Kumar Karampuri
pkarampu at redhat.com
Mon May 1 18:42:27 UTC 2017
On Tue, May 2, 2017 at 12:07 AM, Shyam <srangana at redhat.com> wrote:
> On 05/01/2017 02:23 PM, Pranith Kumar Karampuri wrote:
>
>>
>>
>> On Mon, May 1, 2017 at 11:43 PM, Shyam <srangana at redhat.com
>> <mailto:srangana at redhat.com>> wrote:
>>
>> On 05/01/2017 02:00 PM, Pranith Kumar Karampuri wrote:
>>
>> Splitting the bricks need not be a post factum decision, we
>> can
>> start with larger brick counts, on a given node/disk count,
>> and
>> hence spread these bricks to newer nodes/bricks as they are
>> added.
>>
>>
>> Let's say we have 1 disk, we format it with say XFS and that
>> becomes a
>> brick at the moment. Just curious, what will be the relationship
>> between
>> brick to disk in this case(If we leave out LVM for this example)?
>>
>>
>> I would assume the relation is brick to provided FS directory (not
>> brick to disk, we do not control that at the moment, other than
>> providing best practices around the same).
>>
>>
>> Hmmm... as per my understanding, if we do this then 'df' I guess will
>> report wrong values? available-size/free-size etc will be counted more
>> than once?
>>
>
> This is true even today, if anyone uses 2 bricks from the same mount.
>
That is the reason why documentation is the way it is as far as I can
remember.
>
> I forgot a converse though, we could take a disk and partition it (LVM
> thinp volumes) and use each of those partitions as bricks, avoiding the
> problem of df double counting. Further thinp will help us expand available
> space to other bricks on the same disk, as we destroy older bricks or
> create new ones to accommodate the moving pieces (needs more careful
> thought though, but for sure is a nightmare without thinp).
>
> I am not so much a fan of large number of thinp partitions, so as long as
> that is reasonably in control, we can possibly still use it. The big
> advantage though is, we nuke a thinp volume when the brick that uses that
> partition, moves out of that disk, and we get the space back, rather than
> having or to something akin to rm -rf on the backend to reclaim space.
Other way to achieve the same is to leverage the quota functionality of
counting how much size is used under a directory.
>
>
>
>>
>>
>> Today, gluster takes in a directory on host as a brick, and assuming
>> we retain that, we would need to split this into multiple sub-dirs
>> and use each sub-dir as a brick internally.
>>
>> All these sub-dirs thus created are part of the same volume (due to
>> our current snapshot mapping requirements).
>>
>>
>>
>>
>> --
>> Pranith
>>
>
--
Pranith
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170502/bc2b6c6e/attachment.html>
More information about the Gluster-users
mailing list