Re: [GSOC] Introduction Mail : ceph-mgr: Smarter Reweight-by-Utilization

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Spandan,

Thanks for the response.

On Tue, Mar 28, 2017 at 12:46 AM, Spandan Kumar Sahu
<spandankumarsahu@xxxxxxxxx> wrote:
> Hi Karthik
>
> This is what Kefu Chai, the mentor of this project, suggested me :
> "you can take a
> look at http://tracker.ceph.com/issues/15653, which is already assigned
> to Loic, but by investigating on it, you can get more insight of the project
> you want to take."
>

I will look into it.


> On Tue, Mar 28, 2017 at 2:19 AM, Methuku Karthik <kmeth@xxxxxxxxxxxxxx> wrote:
>> Hi Everyone,
>>
>> My name is Karthik. I am a first year graduate student in Embedded
>> Systems at University of Pennsylvania. I am a avid c, c++ and python
>> programmer.I have 4 years of work experience as  Software developer at
>> Airbus.
>>
>> I have been working as research assistant in PRECISE lab at the
>> University of Pennsylvania to evaluate the performance of the Xen's
>> RTDS scheduler.
>>
>> Currently, I am doing a course on distributed systems. As a part of
>> that course ,I am building a small cloud platform using gRPC (Google's
>> high performance , open-source RPC framework) with the following
>> features:
>>
>> (1)Webmail service (SMTP & POP3) to send, receive and forward mails
>> (2)A fault-tolerant backend server that employs key-Value store
>> similar to Google's Bigtable.
>> (3)The entire Bigtable is distributed across multiple backend servers.
>> (4)Frontend Http server to process requests from a browser, retrieve
>> appropriate data from the backend server and construct the http
>> response for the GUI.
>> (5)Storage service (Similar to Google Drive) with support for
>> navigating the directories, folder creation and uploading and
>> downloading any file type.
>> (6)This system will be fault tolerant with quorum based causal
>> replication done across multiple nodes and load balancing done with
>> dynamic distribution of users among different groups.
>>
>> I compiled and hosted a small cluster to observe how ceph works in
>> storing the data and how the  distribution of the data is maintained
>> while ensuring fault tolerance.With the help of my friend Myna
>> (cc.ed), I could come to speed and performed few experiments to
>> observe how data is shuffled after bringing down one osd.
>>
>> I am currently doing literature review on crush algorithm and
>> understanding the Ceph Architecture.
>>
>> It would be exciting to work on project "ceph-mgr: Smarter
>> Reweight-by-Utilization"
>>
>> Can you point me to any resources that guide to evaluate performace of
>> storage system ?
>>
>> What kind of factors should one consider to evaluate performace of a
>> storage system ?
>> I could think of response time for reading, writing and deleting a
>> file or how quickly a node is configured into a cluster or how quickly
>> cluster heals after a node dies.
>>
>> Please suggest me some existing simple beginner bug which would give
>> me a chance to explore the code.
>>
>> I'm very much interested in Ceph. I want to become a Ceph contributor
>> in the near future.
>>
>> Thank you very much for your help!
>>
>> Best,
>> Karthik
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>
>
> --
> Spandan Kumar Sahu
> IIT Kharagpur

Best,
Karthik
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux