[Gluster-users] Integration of GPU with glusterfs
jim.kinney at gmail.com
Thu Jan 11 13:03:49 UTC 2018
I like the idea immensely. As long as the gpu usage can be specified for server-only, client and server, client and server with a client limit of X. Don't want to take gpu cycles away from machine learning for file IO.
Also must support multiple GPUs and GPU pinning. Really useful for encryption/decryption on client side.
On January 11, 2018 1:12:43 AM EST, Ashish Pandey <aspandey at redhat.com> wrote:
>We have been thinking of exploiting GPU capabilities to enhance
>performance of glusterfs. We would like to know others thoughts on
>In EC, we have been doing CPU intensive computations to encode and
>decode data before writing and reading. This requires a lot of CPU
>cycles and we have
>been observing 100% CPU usage on client side. Data healing will also
>have the same impact as it also needs to do read-decode-encode-write
>As most of the modern servers comes with GPU feature, having glusterfs
>GPU ready might give us performance improvements.
>This is not only specific to EC volume, there are other features which
>will require a lot of computations and could use this capability; For
>1 - Encryption/Decryption
>2 - Compression and de-duplication
>3 - Hashing
>4 - Any other? [Please add if you have something in mind]
>Before proceeding further we would like to have your inputs on this.
>Do you have any other use case (existing or future) which could perform
>better on GPU?
>Do you think that it is worth to integrate GPU with glusterfs? The
>effort to have this performance gain could be achieved by some other
>Any input on the way we should implement it.
>There is a gihub issue opened for this. Please provide your comment or
>reply to this mail.
>A - https://github.com/gluster/glusterfs/issues/388
Sent from my Android device with K-9 Mail. All tyopes are thumb related and reflect authenticity.
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Gluster-users