[Gluster-devel] Suggestion needed to make use of iobuf_pool as rdma buffer.
Mohammed Rafi K C
rkavunga at redhat.com
Wed Jan 14 07:57:48 UTC 2015
On 01/14/2015 12:11 AM, Anand Avati wrote:
> 3) Why not have a separate iobuf pool for RDMA?
Since every fops are using the default iobuf_pool, if we go with another
iobuf_pool dedicated to rdma, we need to copy that buffer from default
pool to rdma or so, unless we are intelligently allocating the buffers
based on the transport which we are going to use. It is an extra level
copying in the I/O path.
>
> On Tue Jan 13 2015 at 6:30:09 AM Mohammed Rafi K C
> <rkavunga at redhat.com <mailto:rkavunga at redhat.com>> wrote:
>
> Hi All,
>
> When using RDMA protocol, we need to register the buffer which is
> going
> to send through rdma with rdma device. In fact, it is a costly
> operation, and a performance killer if it happened in I/O path. So our
> current plan is to register pre-allocated iobuf_arenas from
> iobuf_pool
> with rdma when rdma is getting initialized. The problem comes when all
> the iobufs are exhausted, then we need to dynamically allocate new
> arenas from libglusterfs module. Since it is created in
> libglusterfs, we
> can't make a call to rdma from libglusterfs. So we will force to
> register each of the iobufs from the newly created arenas with rdma in
> I/O path. If io-cache is turned on in client stack, then all the
> pre-registred arenas will use by io-cache as cache buffer. so we
> have to
> do the registration in rdma for each i/o call for every iobufs,
> eventually we cannot make use of pre registered arenas.
>
> To address the issue, we have two approaches in mind,
>
> 1) Register each dynamically created buffers in iobuf by bringing
> transport layer together with libglusterfs.
>
> 2) create a separate buffer for caching and offload the data from the
> read response to the cache buffer in background.
>
> If we could make use of preregister memory for every rdma call,
> then we
> will have approximately 20% increment for write and 25% of
> increment for
> read.
>
> Please give your thoughts to address the issue.
>
> Thanks & Regards
> Rafi KC
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at gluster.org <mailto:Gluster-devel at gluster.org>
> http://www.gluster.org/mailman/listinfo/gluster-devel
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-devel/attachments/20150114/99962606/attachment-0001.html>
More information about the Gluster-devel
mailing list