[Gluster-devel] Simplify creation and set-up of meta-volume (shared storage)

Shubhendu Tripathi shtripat at redhat.com
Mon May 18 08:40:49 UTC 2015


Thanks Avra for the confirmation.

Regards,
Shubhendu

On 05/18/2015 02:04 PM, Avra Sengupta wrote:
> With this option, the volume will be created and explicitly mounted on 
> all the nodes, which are currently a part of the cluster. Please note 
> that new nodes added to the cluster will not have the meta volume 
> mounted explicitly.
>
> So in a case where the console tries to use the volume from a peer in 
> the cluster, which was added after the option was set it will not have 
> the mount available to it. Hence I feel its best that the console 
> continues with the explicit mounting, and showing explicit warning 
> during stop/remove-brick of the meta volume in console and allow the 
> operations.
>
> There is no other impact on the console, as far as this feature is 
> concerned.
>
> Regards,
> Avra
>
> On 05/18/2015 09:27 AM, Shubhendu Tripathi wrote:
>> Avra,
>>
>> We are planning to provide mounting of meta volume on the nodes of 
>> the cluster from Console as part of volume syncing and addition of 
>> new nodes to the cluster. Looks like if this option is set, the 
>> explicit mounting of the meta volume from console is not required and 
>> it would be taken care by gluster.
>>
>> Currently we show explicit warning during stop/remove-brick of the 
>> meta volume in console and allow the operations. I dont feel there 
>> would be any impact due to new feature.
>>
>> Kindly let us know if any other impact on console or we need to take 
>> care of anything else as result of this feature.
>>
>> Thanks and Regards,
>> Shubhendu
>>
>> On 05/15/2015 07:30 PM, Avra Sengupta wrote:
>>> Hi,
>>>
>>> A shared storage meta-volume is currently being used by 
>>> snapshot-scheduler, geo-replication, and nfs-ganesha. In order to 
>>> simplify the creation and set-up of the same, we are introducing a 
>>> global volume set option(cluster.enable-shared-storage).
>>>
>>> On enabling this option, the system analyzes the number of peers in 
>>> the cluster, which are currently connected, and chooses three such 
>>> peers(including the node the command is issued from). From these 
>>> peers a volume(gluster_shared_storage) is created. Depending on the 
>>> number of peers available the volume is either a replica 3 volume(if 
>>> there are 3 connected peers), a replica 2 volume(if there are 2 
>>> connected peers), or a single brick volume(if there is only one node 
>>> in the cluster). "/var/run/gluster/ss_brick" serves as the brick 
>>> path on each node for the shared storage volume. We also mount the 
>>> shared storage at "/var/run/gluster/shared_storage" on all the nodes 
>>> in the cluster as part of enabling this option.
>>>
>>> Once the volume is created, and mounted the maintainance of the 
>>> volume like adding-bricks, removing bricks etc., is expected to be 
>>> the onus of the user.
>>>
>>> On disabling the option, we provide the user a warning, and on 
>>> affirmation from the user we stop the shared storage volume, and 
>>> unmount it from all the nodes in the cluster.
>>>
>>> These are achieved with hook-scripts as part of the volume set 
>>> option. If anyone is interested in having a look at the patch, it's 
>>> available for review at http://review.gluster.org/#/c/10793/ . If 
>>> there is any feedback and suggestion regarding the same, please feel 
>>> free to share it.
>>>
>>> Regards,
>>> Avra
>>>
>>> _______________________________________________
>>> Gluster-devel mailing list
>>> Gluster-devel at gluster.org
>>> http://www.gluster.org/mailman/listinfo/gluster-devel
>>
>
>
>



More information about the Gluster-devel mailing list