Look at 'ceph osd tree'. The weight value in 'ceph osd dump' output is the in/out correction, not the crush weight. s On Mon, 20 May 2013, Alex Bligh wrote: > How do I set the weight for OSDs? I have 4 OSDs I want to create > with very low weight (<1) so they are never used if any other OSDs > are added subsequently (and would like to avoid placement groups). > > These OSDs have been created with default settings using the manual > OSD add procedure as per ceph docs. But (unless I am being stupid > which is quite possible), setting the weight (either to 0.0001 or > to 2) appears to have no effect per a ceph osd dump. > > -- > Alex Bligh > > > > root@kvm:~# ceph osd dump > > epoch 12 > fsid ed0e2e56-bc17-4ef2-a1db-b030c77a8d45 > created 2013-05-20 14:58:02.250461 > modified 2013-05-20 14:59:54.580601 > flags > > pool 0 'data' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 crash_replay_interval 45 > pool 1 'metadata' rep size 2 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 > pool 2 'rbd' rep size 2 min_size 1 crush_ruleset 2 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 > > max_osd 4 > osd.0 up in weight 1 up_from 2 up_thru 10 down_at 0 last_clean_interval [0,0) 10.161.208.1:6800/30687 10.161.208.1:6801/30687 10.161.208.1:6803/30687 exists,up 9cc2a2cf-e79e-404b-9b49-55c8954b0684 > osd.1 up in weight 1 up_from 4 up_thru 11 down_at 0 last_clean_interval [0,0) 10.161.208.1:6804/30800 10.161.208.1:6806/30800 10.161.208.1:6807/30800 exists,up 11628f8d-8234-4329-bf6e-e130d76f18f5 > osd.2 up in weight 1 up_from 3 up_thru 11 down_at 0 last_clean_interval [0,0) 10.161.208.1:6809/30913 10.161.208.1:6810/30913 10.161.208.1:6811/30913 exists,up 050c8955-84aa-4025-961a-f9d9fe60a5b0 > osd.3 up in weight 1 up_from 5 up_thru 11 down_at 0 last_clean_interval [0,0) 10.161.208.1:6812/31024 10.161.208.1:6813/31024 10.161.208.1:6814/31024 exists,up bcd4ad0e-c0e4-4c46-95c2-e68906f8e69a > > > root@kvm:~# ceph osd crush set 0 2 root=default > set item id 0 name 'osd.0' weight 2 at location {root=default} to crush map > root@kvm:~# ceph osd dump > > epoch 14 > fsid ed0e2e56-bc17-4ef2-a1db-b030c77a8d45 > created 2013-05-20 14:58:02.250461 > modified 2013-05-20 15:13:21.009317 > flags > > pool 0 'data' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 crash_replay_interval 45 > pool 1 'metadata' rep size 2 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 > pool 2 'rbd' rep size 2 min_size 1 crush_ruleset 2 object_hash rjenkins pg_num 320 pgp_num 320 last_change 1 owner 0 > > max_osd 4 > osd.0 up in weight 1 up_from 2 up_thru 13 down_at 0 last_clean_interval [0,0) 10.161.208.1:6800/30687 10.161.208.1:6801/30687 10.161.208.1:6803/30687 exists,up 9cc2a2cf-e79e-404b-9b49-55c8954b0684 > osd.1 up in weight 1 up_from 4 up_thru 13 down_at 0 last_clean_interval [0,0) 10.161.208.1:6804/30800 10.161.208.1:6806/30800 10.161.208.1:6807/30800 exists,up 11628f8d-8234-4329-bf6e-e130d76f18f5 > osd.2 up in weight 1 up_from 3 up_thru 13 down_at 0 last_clean_interval [0,0) 10.161.208.1:6809/30913 10.161.208.1:6810/30913 10.161.208.1:6811/30913 exists,up 050c8955-84aa-4025-961a-f9d9fe60a5b0 > osd.3 up in weight 1 up_from 5 up_thru 11 down_at 0 last_clean_interval [0,0) 10.161.208.1:6812/31024 10.161.208.1:6813/31024 10.161.208.1:6814/31024 exists,up bcd4ad0e-c0e4-4c46-95c2-e68906f8e69a > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com