[Gluster-devel] Replace cluster wide gluster locks with volume wide locks

Kaushal M kshlmster at gmail.com
Fri Sep 13 07:26:53 UTC 2013


On Fri, Sep 13, 2013 at 12:14 PM, Avra Sengupta <asengupt at redhat.com> wrote:
> Hi,
>
> Please see comments inline >>>
>
>
> On 09/13/2013 10:58 AM, Vijay Bellur wrote:
>>
>> On 09/13/2013 12:30 AM, Avra Sengupta wrote:
>>>
>>> Hi,
>>>
>>> After having further discussions, we revisited the requirements and it
>>> looks possible to further improve them, as well
>>> as the design.
>>>
>>> 1. We classify all gluster operations in three different classes :
>>> Create volume, Delete volume, and volume specific
>>>     operations.
>>> 2. At any given point of time, we should allow two simultaneous
>>> operations (create, delete or volume specific), as long
>>>     as each both the operations are not happening on the same volume.
>>> 3. If two simultaneous operations are performed on the same volume, the
>>> operation which manages to acquire the volume
>>>     lock will succeed, while the other will fail.
>>>
>>> In order to achieve this, we propose a locking engine, which will
>>> receive lock requests from these three types of
>>> operations.
>>
>>
>> How is the locking engine proposed to be implemented? Is it part of
>> glusterd or a separate process?
>
>>>>The locking engine will be part of glusterd. Today glusterd on every node
>>>> holds a global lock(global to that node), for which
> every gluster command running on that node contests. We propose to use the
> same infra that is in place today(add a new
> rpc to accomodate the volume name in the new lock, instead of using the old
> rpc), and instead of a single global lock, maintain
> multiple volume locks(volume name and node-uuid), for which the respective
> volume operations will contest.
>
>>
>>> Each such request for a particular volume will contest for
>>> the same volume lock (based on the volume name
>>> and the node-uuid). For example, a delete volume command for volume1 and
>>> a volume status command for volume 1 will
>>> contest for the same lock (comprising of the volume name, and the uuid
>>> of the node winning the lock), in which case,
>>> one of these commands will succeed and the other one will not, failing
>>> to acquire the lock.
>>
>>
>> Will volume status need to hold a lock?
>
>>>>Commands like volume status which don't need to hold a lock, will be lock
>>>> less.
>
>>
>>>
>>> Whereas, if two operations are simultaneously performed on a different
>>> volumes they should happen smoothly, as both
>>> these operations would request the locking engine for two different
>>> locks, and will succeed in locking them in parallel.
>>
>>
>> How do you propose to manage the op state machine? Right now it is global
>> in scope - how does that fit into this model?
>
>>>>Although the op state machine is different from the syncop framework,
>>>> that runs on the originator glusterd, it still goes through
> the same states, and also uses the same locking infra today. We propose to
> use the new locking engine for both the state machine
> and the syncop framework. Hetrogeneous clusters, running older versions we
> will use op-versioning to ensure that they use the
> cluster wide lock.

The op-sm currently uses a single global variable 'opinfo' to track
the progress of an operation, which means the currently the op-sm can
track only one operation at a time. I think this is the issue Vijay
was pointing to. We will need to come up with a solution for this as
well.

>
> Regards,
> Avra
>
>>
>> -Vijay
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> https://lists.nongnu.org/mailman/listinfo/gluster-devel




More information about the Gluster-devel mailing list