[Gluster-devel] FreeBSD port of GlusterFS racks up a lot of CPU usage

Rick Macklem rmacklem at uoguelph.ca
Wed Dec 30 23:26:17 UTC 2015

Jordan Hubbard wrote:
> > On Dec 30, 2015, at 2:31 AM, Niels de Vos <ndevos at redhat.com> wrote:
> > 
> >> I'm guessing that Linux uses the event-epoll stuff instead of event-poll,
> >> so it wouldn't exhibit this. Is that correct?
> > 
> > Well, both. most (if not all) Linux builds will use event-poll. But,
> > that calls epoll_wait() with a timeout of 1 millisecond as well.
> > 
> >> Thanks for any information on this, rick
> >> ps: I am tempted to just crank the timeout of 1msec up to 10 or 20msec.
> > 
> > Yes, that is probably what I would do too. And have both poll functions
> > use the same timeout, have it defined in libglusterfs/src/event.h. We
> > could make it a configurable option too, but I do not think it is very
> > useful to have.
> I guess this begs the question - what’s the actual purpose of polling for an
> event with a 1 millisecond timeout?  If it was some sort of heartbeat check,
> one might imagine that would be better served by a timer with nothing close
> to 1 millisecond as an interval (that would be one seriously aggressive
> heartbeat) and if filesystem events are incoming that glusterfs needs to
> respond to, why timeout at all?
If I understand the code (I probably don't) the timeout allows the loop
to call a function that may add new fd's to be polled. (If I'm right,
the new ones might not get serviced.)

I'll post once I've tried a longer timeout and if it seems ok, I will
put it in the Redhat bugs database (as mentioned in the last post).
In its current form, it's fine for testing.

> I also have a broader question to go with the specific one:  We (at
> iXsystems) were attempting to engage with some of the Red Hat folks back
> when the FreeBSD port was first done, in the hope of getting it more
> “officially supported” for FreeBSD and perhaps even donating some more
> serious stress-testing and integration work for it, but when those Red Hat
> folks moved on we lost continuity and the effort stalled.  Who at Red Hat
> would / could we work with in getting this back on track?  We’d like to
> integrate glusterfs with FreeNAS 10, and in fact have already done so but
> it’s still early days and we’re not even really sure what we have yet.
Just fyi..sofar, working with FreeBSD11/head and the port of 3.7.6 (the port tarball
is in FreeBSD PR#194409), the only GlusterFS problem I've encountered is
the above one. I'm not sure why this isn't in /usr/ports, but that would be
nice as it might get more people trying it. (I'm a src comitter, but not a
ports one.)

However, I have several patches for the FreeBSD fuse interface and for
a mount_glusterfs mount to work ok you need a couple of them.
1 - When an open decides to do DIRECT_IO after the file has done buffer
    cache I/O the buffer cache needs to be invalidated so you don't get
    stale cached data.
2 - For a WRONLY write, you need to force DIRECT_IO (or do a read/write open).
    If you don't do this, the buffer cache code will get stuck when trying
    to read a block in before writing a partial block. (I think this is
    what FreeBSD PR#194293 is caused by.)

Because I won't be able to do svn until April, these patches won't make it
into head for a while, but they will both be in PR#194293 within hours.

The others add features like extended attributes, advisory byte range locking
and the changes needed to export the fuse/glusterfs mount via the FreeBSD
kernel nfsd. If anyone wants/needs these patches, email and I can send you

A bit off your topic, but until you have the fixes for FreeBSD fuse, you
probably can't do a lot of serious testing.
(I don't know, but I'd guess that FreeNAS has about the same fuse module
 code as FreeBSD's head, since it hasn't been changed much in head recently.)

Thanks everyone for your help with this, rick

> Thanks,
> - Jordan

More information about the Gluster-devel mailing list