[Gluster-users] Add single server
Pranith Kumar Karampuri
pkarampu at redhat.com
Mon May 1 18:00:14 UTC 2017
On Mon, May 1, 2017 at 11:20 PM, Shyam <srangana at redhat.com> wrote:
> On 05/01/2017 01:13 PM, Pranith Kumar Karampuri wrote:
>
>>
>>
>> On Mon, May 1, 2017 at 10:42 PM, Pranith Kumar Karampuri
>> <pkarampu at redhat.com <mailto:pkarampu at redhat.com>> wrote:
>>
>>
>>
>> On Mon, May 1, 2017 at 10:39 PM, Gandalf Corvotempesta
>> <gandalf.corvotempesta at gmail.com
>> <mailto:gandalf.corvotempesta at gmail.com>> wrote:
>>
>> 2017-05-01 18:57 GMT+02:00 Pranith Kumar Karampuri
>> <pkarampu at redhat.com <mailto:pkarampu at redhat.com>>:
>>
>> > Yes this is precisely what all the other SDS with metadata
>> servers kind of
>> > do. They kind of keep a map of on what all servers a particular
>> file/blob is
>> > stored in a metadata server.
>>
>> Not exactly. Other SDS has some servers dedicated to metadata and,
>> personally, I don't like that approach.
>>
>> > GlusterFS doesn't do that. In GlusterFS what
>> > bricks need to be replicated is always given and distribute
>> layer on top of
>> > these replication layer will do the job of distributing and
>> fetching the
>> > data. Because replication happens at a brick level and not at a
>> file level
>> > and distribute happens on top of replication and not at file
>> level. There
>> > isn't too much metadata that needs to be stored per file. Hence
>> no need for
>> > separate metadata servers.
>>
>> And this is great, that's why i'm talking about embedding a sort
>> of database
>> to be stored on all nodes. no metadata servers, only a mapping
>> between files
>> and servers.
>>
>> > If you know path of the file, you can always know where the
>> file is stored
>> > using pathinfo:
>> > Method-2 in the following link:
>> > https://gluster.readthedocs.io/en/latest/Troubleshooting/gfi
>> d-to-path/
>> <https://gluster.readthedocs.io/en/latest/Troubleshooting/gf
>> id-to-path/>
>> >
>> > You don't need any db.
>>
>> For the current gluster yes.
>> I'm talking about a different thing.
>>
>> In a RAID, you have data stored somewhere on the array, with
>> metadata
>> defining how this data should
>> be wrote or read. obviously, raid metadata must be stored in a
>> fixed
>> position, or you won't be able to read
>> that.
>>
>> Something similiar could be added in gluster (i don't know if it
>> would
>> be hard): you store a file mapping in a fixed
>> position in gluster, then all gluster clients will be able to know
>> where a file is by looking at this "metadata" stored in
>> the fixed position.
>>
>> Like ".gluster" directory. Gluster is using some "internal"
>> directories for internal operations (".shards", ".gluster",
>> ".trash")
>> A ".metadata" with file mapping would be hard to add ?
>>
>> > Basically what you want, if I understood correctly is:
>> > If we add a 3rd node with just one disk, the data should
>> automatically
>> > arrange itself splitting itself to 3 categories(Assuming
>> replica-2)
>> > 1) Files that are present in Node1, Node2
>> > 2) Files that are present in Node2, Node3
>> > 3) Files that are present in Node1, Node3
>> >
>> > As you can see we arrived at a contradiction where all the
>> nodes should have
>> > at least 2 bricks but there is only 1 disk. Hence the
>> contradiction. We
>> > can't do what you are asking without brick splitting. i.e. we
>> need to split
>> > the disk into 2 bricks.
>>
>
> Splitting the bricks need not be a post factum decision, we can start with
> larger brick counts, on a given node/disk count, and hence spread these
> bricks to newer nodes/bricks as they are added.
>
Let's say we have 1 disk, we format it with say XFS and that becomes a
brick at the moment. Just curious, what will be the relationship between
brick to disk in this case(If we leave out LVM for this example)?
> If I understand the ceph PG count, it works on a similar notion, till the
> cluster grows beyond the initial PG count (set for the pool) at which point
> there is a lot more data movement (as the pg count has to be increased, and
> hence existing PGs need to be further partitioned) . (just using ceph as an
> example, a similar approach exists for openstack swift with their partition
> power settings).
>
>
>> I don't think so.
>> Let's assume a replica 2.
>>
>> S1B1 + S2B1
>>
>> 1TB each, thus 1TB available (2TB/2)
>>
>> Adding a third 1TB disks should increase available space to
>> 1.5TB (3TB/2)
>>
>>
>> I agree it should. Question is how? What will be the resulting
>> brick-map?
>>
>>
>> I don't see any solution that we can do without at least 2 bricks on
>> each of the 3 servers.
>>
>>
>>
>>
>> --
>> Pranith
>>
>>
>>
>>
>> --
>> Pranith
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
--
Pranith
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170501/975f9ffc/attachment.html>
More information about the Gluster-users
mailing list