Re: CEPH Expansion

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Craig!

Indeed I had reduced the replicated size to 2 instead of 3 while the minimum size is 1.

I hadn't touched the crushmap though.

I would like to keep on going with the replicated size of 2 . Do you think this would be a problem?

Please find below the output of the command:

$ ceph osd dump | grep ^pool
pool 3 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 524 flags hashpspool stripe_width 0 pool 4 'metadata' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 526 flags hashpspool stripe_width 0 pool 5 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 528 flags hashpspool stripe_width 0 pool 6 '.rgw' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 618 flags hashpspool stripe_width 0 pool 7 '.rgw.control' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 616 flags hashpspool stripe_width 0 pool 8 '.rgw.gc' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 614 flags hashpspool stripe_width 0 pool 9 '.log' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 612 flags hashpspool stripe_width 0 pool 10 '.intent-log' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 610 flags hashpspool stripe_width 0 pool 11 '.usage' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 608 flags hashpspool stripe_width 0 pool 12 '.users' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 606 flags hashpspool stripe_width 0 pool 13 '.users.email' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 604 flags hashpspool stripe_width 0 pool 14 '.users.swift' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 602 flags hashpspool stripe_width 0 pool 15 '.users.uid' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 600 flags hashpspool stripe_width 0 pool 16 '.rgw.root' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 598 flags hashpspool stripe_width 0 pool 17 '.rgw.buckets.index' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 596 flags hashpspool stripe_width 0 pool 18 '.rgw.buckets' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 1024 pgp_num 1024 last_change 826 flags hashpspool stripe_width 0 pool 19 '.rgw.buckets.extra' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 722 owner 18446744073709551615 flags hashpspool stripe_width 0



Warmest regards,


George

Youve either modified the crushmap, or changed the pool size to 1. 
The defaults create 3 replicas on different hosts.

What does `ceph osd dump | grep ^pool` output?  If the size param is
1, then you reduced the replica count.  If the size param is > 1, you
mustve adjusted the crushmap.

Either way, after you add the second node would be the ideal time to
change that back to the default.

Given that you only have 40GB of data in the cluster, you shouldnt
have a problem adding the 2nd node.

On Fri, Jan 23, 2015 at 3:58 PM, Georgios Dimitrakakis  wrote:

Hi Craig!

For the moment I have only one node with 10 OSDs.
I want to add a second one with 10 more OSDs.

Each OSD in every node is a 4TB SATA drive. No SSD disks!

The data ara approximately 40GB and I will do my best to have zero
or at least very very low load during the expansion process.

To be honest I havent touched the crushmap. I wasnt aware that I
should have changed it. Therefore, it still is with the default
one.
Is that OK? Where can I read about the host level replication in
CRUSH map in order
to make sure that its applied or how can I find if this is already
enabled?

Any other things that I should be aware of?

All the best,

George

It depends.  There are a lot of variables, like how many nodes
and
disks you currently have.  Are you using journals on SSD.  How
much
data is already in the cluster.  What the client load is on the
cluster.

Since you only have 40 GB in the cluster, it shouldnt take long
to
backfill.  You may find that it finishes backfilling faster than
you
can format the new disks.

Since you only have a single OSD node, you mustve changed the
crushmap
to allow replication over OSDs instead of hosts.  After you get
the
new node in would be the best time to switch back to host level
replication.  The more data you have, the more painful that
change
will become.

On Sun, Jan 18, 2015 at 10:09 AM, Georgios Dimitrakakis  wrote:

Hi Jiri,

thanks for the feedback.

My main concern is if its better to add each OSD one-by-one and
wait for the cluster to rebalance every time or do it
all-together
at once.

Furthermore an estimate of the time to rebalance would be
great!

Regards,

Links:
------
[1] mailto:giorgis@xxxxxxxxxxxx [1]

--



Links:
------
[1] mailto:giorgis@xxxxxxxxxxxx
[2] mailto:giorgis@xxxxxxxxxxxx

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux