[Gluster-users] Is this viable / useful / silly
Russell Purinton
russell.purinton at gmail.com
Tue Apr 5 14:15:18 UTC 2016
For what it’s worth, I once created a stripe volume on ram disks. After the initial creation of the bricks, I made a copy of all the files gluster created. After reboot, the files are copied back to the ramdisk before gluster starts, so basically after a reboot you have an empty gluster volume once again.
The performance was really good. Maxed out the dual 10GbE on each server.
If you need really-high IOPS to a file that may be too big for a ramdisk in 1 machine, consider a stripe volume of multiple ram disks.
> On Apr 5, 2016, at 8:53 AM, Sean Delaney <sdelaney at cp.dias.ie> wrote:
>
> Hi all,
>
> I'm considering using my cluster's local scratch SSDs as a shared filesystem. I'd like to be able to start glusterfs on a few nodes (say 16), run a HPC job on those same nodes (reading/writing on glusterfs), copy the final result off to the panasas storage, and shut down glusterfs until next time.
>
> I'm interested in this because my workload has shown strong performance on the SSDs, which I'd like to scale out a little.
>
> Ultimately, I might be interested in setting up a tiered glusterfs using the SSDs as the hot tier. Again, the ability to bring the filesystem up and down easily would be of interest.
>
> Example cluster: 32 nodes, 1.5 TB SSD (xfs) per node, separate HDD for OS, panasas storage.
>
> Thanks
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160405/8d8669fb/attachment.html>
More information about the Gluster-users
mailing list