[Gluster-devel] Simplify creation and set-up of meta-volume (shared storage)
Avra Sengupta
asengupt at redhat.com
Mon May 18 08:34:49 UTC 2015
With this option, the volume will be created and explicitly mounted on
all the nodes, which are currently a part of the cluster. Please note
that new nodes added to the cluster will not have the meta volume
mounted explicitly.
So in a case where the console tries to use the volume from a peer in
the cluster, which was added after the option was set it will not have
the mount available to it. Hence I feel its best that the console
continues with the explicit mounting, and showing explicit warning
during stop/remove-brick of the meta volume in console and allow the
operations.
There is no other impact on the console, as far as this feature is
concerned.
Regards,
Avra
On 05/18/2015 09:27 AM, Shubhendu Tripathi wrote:
> Avra,
>
> We are planning to provide mounting of meta volume on the nodes of the
> cluster from Console as part of volume syncing and addition of new
> nodes to the cluster. Looks like if this option is set, the explicit
> mounting of the meta volume from console is not required and it would
> be taken care by gluster.
>
> Currently we show explicit warning during stop/remove-brick of the
> meta volume in console and allow the operations. I dont feel there
> would be any impact due to new feature.
>
> Kindly let us know if any other impact on console or we need to take
> care of anything else as result of this feature.
>
> Thanks and Regards,
> Shubhendu
>
> On 05/15/2015 07:30 PM, Avra Sengupta wrote:
>> Hi,
>>
>> A shared storage meta-volume is currently being used by
>> snapshot-scheduler, geo-replication, and nfs-ganesha. In order to
>> simplify the creation and set-up of the same, we are introducing a
>> global volume set option(cluster.enable-shared-storage).
>>
>> On enabling this option, the system analyzes the number of peers in
>> the cluster, which are currently connected, and chooses three such
>> peers(including the node the command is issued from). From these
>> peers a volume(gluster_shared_storage) is created. Depending on the
>> number of peers available the volume is either a replica 3 volume(if
>> there are 3 connected peers), a replica 2 volume(if there are 2
>> connected peers), or a single brick volume(if there is only one node
>> in the cluster). "/var/run/gluster/ss_brick" serves as the brick path
>> on each node for the shared storage volume. We also mount the shared
>> storage at "/var/run/gluster/shared_storage" on all the nodes in the
>> cluster as part of enabling this option.
>>
>> Once the volume is created, and mounted the maintainance of the
>> volume like adding-bricks, removing bricks etc., is expected to be
>> the onus of the user.
>>
>> On disabling the option, we provide the user a warning, and on
>> affirmation from the user we stop the shared storage volume, and
>> unmount it from all the nodes in the cluster.
>>
>> These are achieved with hook-scripts as part of the volume set
>> option. If anyone is interested in having a look at the patch, it's
>> available for review at http://review.gluster.org/#/c/10793/ . If
>> there is any feedback and suggestion regarding the same, please feel
>> free to share it.
>>
>> Regards,
>> Avra
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-devel
>
More information about the Gluster-devel
mailing list