Re: FreeBSD port of GlusterFS racks up a lot of CPU usage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Fri, Jan 8, 2016 at 1:32 PM, Xavier Hernandez <xhernandez@xxxxxxxxxx> wrote:

On 08/01/16 05:42, Raghavendra G wrote:
Sorry for the delayed reply. Had missed out this mail. Please find my
comments inlined.

On Thu, Dec 31, 2015 at 4:56 AM, Rick Macklem <rmacklem@xxxxxxxxxxx
<mailto:rmacklem@xxxxxxxxxxx>> wrote:

    Jordan Hubbard wrote:
    >
    > > On Dec 30, 2015, at 2:31 AM, Niels de Vos <ndevos@xxxxxxxxxx <mailto:ndevos@xxxxxxxxxx>> wrote:
    > >
    > >> I'm guessing that Linux uses the event-epoll stuff instead of event-poll,
    > >> so it wouldn't exhibit this. Is that correct?
    > >
    > > Well, both. most (if not all) Linux builds will use event-poll. But,
    > > that calls epoll_wait() with a timeout of 1 millisecond as well.
    > >
    > >> Thanks for any information on this, rick
    > >> ps: I am tempted to just crank the timeout of 1msec up to 10 or 20msec.
    > >
    > > Yes, that is probably what I would do too. And have both poll functions
    > > use the same timeout, have it defined in libglusterfs/src/event.h. We
    > > could make it a configurable option too, but I do not think it is very
    > > useful to have.
    >
    > I guess this begs the question - what’s the actual purpose of polling for an
    > event with a 1 millisecond timeout?  If it was some sort of heartbeat check,
    > one might imagine that would be better served by a timer with nothing close
    > to 1 millisecond as an interval (that would be one seriously aggressive
    > heartbeat) and if filesystem events are incoming that glusterfs needs to
    > respond to, why timeout at all?
    >
    If I understand the code (I probably don't) the timeout allows the loop
    to call a function that may add new fd's to be polled. (If I'm right,
    the new ones might not get serviced.)


Yes, that's correct. Since in poll we pass the fds to be polled in an
array as an argument, the only place where we can add/remove fds to be
polled is at the time we call poll sycall. To make adding/removing fds
from polling to be more responsive, poll timeouts "frequently enough".
The trade-off we are considering here is between:

1. Number of calls to poll
            vs
2. Responsiveness of adding/removing a new fd from polling.

For clients, there is not much change of the list of fds that are
polled. However, for bricks/server this list can vary frequently as new
clients are connected/disconnected.

Since epoll provides a way to add new fds for polling while an
epoll_wait is in progress (unlike poll), the timeout of epoll_wait is
infinite. Also note that on systems where both epoll and poll are
available, epoll is preferred over poll.

I don't know anything about gluster's poll implementation so I may be totally wrong, but would it be possible to use an eventfd (or a pipe if eventfd is not supported) to signal the need to add more file descriptors to the poll call ?

The poll call should listen on this new fd. When we need to change the fd list, we should simply write to the eventfd or pipe from another thread. This will cause the poll call to return and we will be able to change the fd list without having a short timeout nor having to decide on any trade-off.

Thats a nice idea. Based on my understanding of why timeouts are being used, this approach can work.
 

Just an idea...

Xavi



    I'll post once I've tried a longer timeout and if it seems ok, I will
    put it in the Redhat bugs database (as mentioned in the last post).
    In its current form, it's fine for testing.

    > I also have a broader question to go with the specific one:  We (at
    > iXsystems) were attempting to engage with some of the Red Hat folks back
    > when the FreeBSD port was first done, in the hope of getting it more
    > “officially supported” for FreeBSD and perhaps even donating some more
    > serious stress-testing and integration work for it, but when those Red Hat
    > folks moved on we lost continuity and the effort stalled.  Who at Red Hat
    > would / could we work with in getting this back on track?  We’d like to
    > integrate glusterfs with FreeNAS 10, and in fact have already done so but
    > it’s still early days and we’re not even really sure what we have yet.
    >
    Just fyi..sofar, working with FreeBSD11/head and the port of 3.7.6
    (the port tarball
    is in FreeBSD PR#194409), the only GlusterFS problem I've encountered is
    the above one. I'm not sure why this isn't in /usr/ports, but that
    would be
    nice as it might get more people trying it. (I'm a src comitter, but
    not a
    ports one.)

    However, I have several patches for the FreeBSD fuse interface and for
    a mount_glusterfs mount to work ok you need a couple of them.
    1 - When an open decides to do DIRECT_IO after the file has done buffer
         cache I/O the buffer cache needs to be invalidated so you don't get
         stale cached data.
    2 - For a WRONLY write, you need to force DIRECT_IO (or do a
    read/write open).
         If you don't do this, the buffer cache code will get stuck when
    trying
         to read a block in before writing a partial block. (I think this is
         what FreeBSD PR#194293 is caused by.)

    Because I won't be able to do svn until April, these patches won't
    make it
    into head for a while, but they will both be in PR#194293 within hours.

    The others add features like extended attributes, advisory byte
    range locking
    and the changes needed to export the fuse/glusterfs mount via the
    FreeBSD
    kernel nfsd. If anyone wants/needs these patches, email and I can
    send you
    them.

    A bit off your topic, but until you have the fixes for FreeBSD fuse, you
    probably can't do a lot of serious testing.
    (I don't know, but I'd guess that FreeNAS has about the same fuse module
      code as FreeBSD's head, since it hasn't been changed much in head
    recently.)

    Thanks everyone for your help with this, rick

     > Thanks,
     >
     > - Jordan
     >
     >
    _______________________________________________
    Gluster-devel mailing list
    Gluster-devel@xxxxxxxxxxx <mailto:Gluster-devel@xxxxxxxxxxx>
    http://www.gluster.org/mailman/listinfo/gluster-devel




--
Raghavendra G


_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel

_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel



--
Raghavendra G
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel

[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux