We are doing the latter. What is the best way to force it to host first selection? On Tue, Feb 5, 2013 at 11:34 AM, Sage Weil <sage@xxxxxxxxxxx> wrote: > On Tue, 5 Feb 2013, Mandell Degerness wrote: >> We are using v0.56.2 and it still seems that the default crushmap is >> osd centered. Here is the crushmap as dumped: >> >> [root@node-172-19-0-15 ~]# cat crush.txt >> # begin crush map >> >> # devices >> device 0 osd.0 >> device 1 osd.1 >> device 2 osd.2 >> device 3 osd.3 >> device 4 osd.4 >> device 5 osd.5 >> device 6 osd.6 >> device 7 osd.7 >> device 8 osd.8 >> device 9 osd.9 >> device 10 osd.10 >> device 11 osd.11 >> device 12 osd.12 >> device 13 osd.13 >> device 14 osd.14 >> >> # types >> type 0 osd >> type 1 host >> type 2 rack >> type 3 row >> type 4 room >> type 5 datacenter >> type 6 root >> >> # buckets >> host 172.19.0.14 { >> id -2 # do not change unnecessarily >> # weight 6.000 >> alg straw >> hash 0 # rjenkins1 >> item osd.0 weight 1.000 >> item osd.4 weight 1.000 >> item osd.5 weight 1.000 >> item osd.12 weight 1.000 >> item osd.13 weight 1.000 >> item osd.14 weight 1.000 >> } >> host 172.19.0.13 { >> id -4 # do not change unnecessarily >> # weight 3.000 >> alg straw >> hash 0 # rjenkins1 >> item osd.2 weight 1.000 >> item osd.7 weight 1.000 >> item osd.9 weight 1.000 >> } >> host 172.19.0.16 { >> id -5 # do not change unnecessarily >> # weight 3.000 >> alg straw >> hash 0 # rjenkins1 >> item osd.3 weight 1.000 >> item osd.6 weight 1.000 >> item osd.10 weight 1.000 >> } >> host 172.19.0.15 { >> id -6 # do not change unnecessarily >> # weight 3.000 >> alg straw >> hash 0 # rjenkins1 >> item osd.1 weight 1.000 >> item osd.8 weight 1.000 >> item osd.11 weight 1.000 >> } >> rack 0 { >> id -3 # do not change unnecessarily >> # weight 15.000 >> alg straw >> hash 0 # rjenkins1 >> item 172.19.0.14 weight 6.000 >> item 172.19.0.13 weight 3.000 >> item 172.19.0.16 weight 3.000 >> item 172.19.0.15 weight 3.000 >> } >> root default { >> id -1 # do not change unnecessarily >> # weight 15.000 >> alg straw >> hash 0 # rjenkins1 >> item 0 weight 15.000 >> } >> >> # rules >> rule data { >> ruleset 0 >> type replicated >> min_size 1 >> max_size 10 >> step take default >> step choose firstn 0 type osd >> step emit >> } >> rule metadata { >> ruleset 1 >> type replicated >> min_size 1 >> max_size 10 >> step take default >> step choose firstn 0 type osd >> step emit >> } >> rule rbd { >> ruleset 2 >> type replicated >> min_size 1 >> max_size 10 >> step take default >> step choose firstn 0 type osd >> step emit >> } >> >> # end crush map >> [root@node-172-19-0-15 ~]# ceph --version >> ceph version 0.56.2 (586538e22afba85c59beda49789ec42024e7a061) >> >> We do not run any explicit crushtool commands as part of our start up >> at this time. Should we be? > > Do you run mkcephfs? If you are passing in a ceph.conf to mkcephfs, it is > still dynamically choosing a rule based on whether you have enough osds (3 > i think? i forget). If you are running ceph-mon --mkfs directly (as > ceph-deploy, chef, juju do), it will always default to osds. > > sage > >> >> Regards, >> Mandell Degerness >> >> On Wed, Jan 30, 2013 at 3:46 PM, Sage Weil <sage@xxxxxxxxxxx> wrote: >> > The next bobtail point release is ready, and it's looking pretty good. >> > This is an important update for the 0.56.x backport series that fixes a >> > number of bugs and several performance issues. All v0.56.x users are >> > encouraged to upgrade. >> > >> > Notable changes since v0.56.1: >> > >> > * osd: snapshot trimming fixes >> > * osd: scrub snapshot metadata >> > * osd: fix osdmap trimming >> > * osd: misc peering fixes >> > * osd: stop heartbeating with peers if internal threads are stuck/hung >> > * osd: PG removal is friendlier to other workloads >> > * osd: fix recovery start delay (was causing very slow recovery) >> > * osd: fix scheduling of explicitly requested scrubs >> > * osd: fix scrub interval config options >> > * osd: improve recovery vs client io tuning >> > * osd: improve 'slow request' warning detail for better diagnosis >> > * osd: default CRUSH map now distributes across hosts, not OSDs >> > * osd: fix crash on 32-bit hosts triggered by librbd clients >> > * librbd: fix error handling when talking to older OSDs >> > * mon: fix a few rare crashes >> > * ceph command: ability to easily adjust CRUSH tunables >> > * radosgw: object copy does not copy source ACLs >> > * rados command: fix omap command usage >> > * sysvinit script: set ulimit -n properly on remote hosts >> > * msgr: fix narrow race with message queuing >> > * fixed compilation on some old distros (e.g., RHEL 5.x) >> > >> > There are a small number of interface changes related to the default CRUSH >> > rule and scrub interval configuration options. Please see the full release >> > notes. >> > >> > You can get v0.56.2 in the usual fashion: >> > >> > * Git at git://github.com/ceph/ceph.git >> > * Tarball at http://ceph.com/download/ceph-0.56.2.tar.gz >> > * For Debian/Ubuntu packages, see http://ceph.com/docs/master/install/debian >> > * For RPMs, see http://ceph.com/docs/master/install/rpm >> > >> > -- >> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> > the body of a message to majordomo@xxxxxxxxxxxxxxx >> > More majordomo info at http://vger.kernel.org/majordomo-info.html >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> >> -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html