Re: [RFC]New Message Implementation Based on Event

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



As for testing, now I mainly passed tests in src/tests such as
ceph_test_rados. Because of the lack of Messenger's unittest, I have
to deploy this branch into my dev cluster to test. I'm thinking in
make ms_inject* options available in this Messenger for failure
coverage.

On Mon, Sep 15, 2014 at 11:51 PM, Sage Weil <sweil@xxxxxxxxxx> wrote:
> Hi Haomai,
>
> On Fri, 12 Sep 2014, Haomai Wang wrote:
>> Hi all,
>>
>> Recently, I did some basic work on new message implementation based on
>> event(https://github.com/yuyuyu101/ceph/tree/msg-event). The basic
>> idea is that we use a Processor thread for each Messenger to monitor
>> all sockets and dispatch fd to threadpool. The event mechanism can be
>> epoll, kqueue, poll or select. The thread in threadpool will
>> read/write with this socket and dispatch message later.
>>
>> Now the branch has passed basic tests and before make it more stable
>> and pass more QA suites. I want to do some benchmark tests compared to
>> pipe implementation with large-scale cluster. I would like to use at
>> least 100 OSDs(SSD) and hundreds of clients to test it. And now the
>> benchmark for only one OSD, the client can get the same latency with
>> pipe implementation and the latency stdev will be smaller.
>>
>> The background for this implementation is that pipe implementation
>> consumes too much overhead on context switch and thread resource. In
>> our env, several ceph-osd is running on compute node which also runs
>> KVM process.
>>
>> Do you have any ideas about this, or some serious concerns compared to pipe.
>
> I haven't had time to look at this in much detail yet, but at a high
> level, this looks awesome!  It sounds like using an event lib for this is
> a good approach, and from a quick skim it looks like you've already done
> the hard work of breaking all of the logic in Pipe.cc into a state
> machine.
>
> How much testing have you done with this?
>
> I hope to find more time this week to look in more detail, but wanted to
> let you know I didn't miss this before that :)
>
> Cheers-
> sage



-- 
Best Regards,

Wheat
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux