Hi Karthik This is what Kefu Chai, the mentor of this project, suggested me : "you can take a look at http://tracker.ceph.com/issues/15653, which is already assigned to Loic, but by investigating on it, you can get more insight of the project you want to take." On Tue, Mar 28, 2017 at 2:19 AM, Methuku Karthik <kmeth@xxxxxxxxxxxxxx> wrote: > Hi Everyone, > > My name is Karthik. I am a first year graduate student in Embedded > Systems at University of Pennsylvania. I am a avid c, c++ and python > programmer.I have 4 years of work experience as Software developer at > Airbus. > > I have been working as research assistant in PRECISE lab at the > University of Pennsylvania to evaluate the performance of the Xen's > RTDS scheduler. > > Currently, I am doing a course on distributed systems. As a part of > that course ,I am building a small cloud platform using gRPC (Google's > high performance , open-source RPC framework) with the following > features: > > (1)Webmail service (SMTP & POP3) to send, receive and forward mails > (2)A fault-tolerant backend server that employs key-Value store > similar to Google's Bigtable. > (3)The entire Bigtable is distributed across multiple backend servers. > (4)Frontend Http server to process requests from a browser, retrieve > appropriate data from the backend server and construct the http > response for the GUI. > (5)Storage service (Similar to Google Drive) with support for > navigating the directories, folder creation and uploading and > downloading any file type. > (6)This system will be fault tolerant with quorum based causal > replication done across multiple nodes and load balancing done with > dynamic distribution of users among different groups. > > I compiled and hosted a small cluster to observe how ceph works in > storing the data and how the distribution of the data is maintained > while ensuring fault tolerance.With the help of my friend Myna > (cc.ed), I could come to speed and performed few experiments to > observe how data is shuffled after bringing down one osd. > > I am currently doing literature review on crush algorithm and > understanding the Ceph Architecture. > > It would be exciting to work on project "ceph-mgr: Smarter > Reweight-by-Utilization" > > Can you point me to any resources that guide to evaluate performace of > storage system ? > > What kind of factors should one consider to evaluate performace of a > storage system ? > I could think of response time for reading, writing and deleting a > file or how quickly a node is configured into a cluster or how quickly > cluster heals after a node dies. > > Please suggest me some existing simple beginner bug which would give > me a chance to explore the code. > > I'm very much interested in Ceph. I want to become a Ceph contributor > in the near future. > > Thank you very much for your help! > > Best, > Karthik > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Spandan Kumar Sahu IIT Kharagpur -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html