[GSOC] Introduction Mail : ceph-mgr: Smarter Reweight-by-Utilization

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Everyone,

My name is Karthik. I am a first year graduate student in Embedded
Systems at University of Pennsylvania. I am a avid c, c++ and python
programmer.I have 4 years of work experience as  Software developer at
Airbus.

I have been working as research assistant in PRECISE lab at the
University of Pennsylvania to evaluate the performance of the Xen's
RTDS scheduler.

Currently, I am doing a course on distributed systems. As a part of
that course ,I am building a small cloud platform using gRPC (Google's
high performance , open-source RPC framework) with the following
features:

(1)Webmail service (SMTP & POP3) to send, receive and forward mails
(2)A fault-tolerant backend server that employs key-Value store
similar to Google's Bigtable.
(3)The entire Bigtable is distributed across multiple backend servers.
(4)Frontend Http server to process requests from a browser, retrieve
appropriate data from the backend server and construct the http
response for the GUI.
(5)Storage service (Similar to Google Drive) with support for
navigating the directories, folder creation and uploading and
downloading any file type.
(6)This system will be fault tolerant with quorum based causal
replication done across multiple nodes and load balancing done with
dynamic distribution of users among different groups.

I compiled and hosted a small cluster to observe how ceph works in
storing the data and how the  distribution of the data is maintained
while ensuring fault tolerance.With the help of my friend Myna
(cc.ed), I could come to speed and performed few experiments to
observe how data is shuffled after bringing down one osd.

I am currently doing literature review on crush algorithm and
understanding the Ceph Architecture.

It would be exciting to work on project "ceph-mgr: Smarter
Reweight-by-Utilization"

Can you point me to any resources that guide to evaluate performace of
storage system ?

What kind of factors should one consider to evaluate performace of a
storage system ?
I could think of response time for reading, writing and deleting a
file or how quickly a node is configured into a cluster or how quickly
cluster heals after a node dies.

Please suggest me some existing simple beginner bug which would give
me a chance to explore the code.

I'm very much interested in Ceph. I want to become a Ceph contributor
in the near future.

Thank you very much for your help!

Best,
Karthik
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux