Re: v0.56.2 released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 5 Feb 2013, Mandell Degerness wrote:
> We are using v0.56.2 and it still seems that the default crushmap is
> osd centered.  Here is the crushmap as dumped:
> 
> [root@node-172-19-0-15 ~]# cat crush.txt
> # begin crush map
> 
> # devices
> device 0 osd.0
> device 1 osd.1
> device 2 osd.2
> device 3 osd.3
> device 4 osd.4
> device 5 osd.5
> device 6 osd.6
> device 7 osd.7
> device 8 osd.8
> device 9 osd.9
> device 10 osd.10
> device 11 osd.11
> device 12 osd.12
> device 13 osd.13
> device 14 osd.14
> 
> # types
> type 0 osd
> type 1 host
> type 2 rack
> type 3 row
> type 4 room
> type 5 datacenter
> type 6 root
> 
> # buckets
> host 172.19.0.14 {
> id -2 # do not change unnecessarily
> # weight 6.000
> alg straw
> hash 0 # rjenkins1
> item osd.0 weight 1.000
> item osd.4 weight 1.000
> item osd.5 weight 1.000
> item osd.12 weight 1.000
> item osd.13 weight 1.000
> item osd.14 weight 1.000
> }
> host 172.19.0.13 {
> id -4 # do not change unnecessarily
> # weight 3.000
> alg straw
> hash 0 # rjenkins1
> item osd.2 weight 1.000
> item osd.7 weight 1.000
> item osd.9 weight 1.000
> }
> host 172.19.0.16 {
> id -5 # do not change unnecessarily
> # weight 3.000
> alg straw
> hash 0 # rjenkins1
> item osd.3 weight 1.000
> item osd.6 weight 1.000
> item osd.10 weight 1.000
> }
> host 172.19.0.15 {
> id -6 # do not change unnecessarily
> # weight 3.000
> alg straw
> hash 0 # rjenkins1
> item osd.1 weight 1.000
> item osd.8 weight 1.000
> item osd.11 weight 1.000
> }
> rack 0 {
> id -3 # do not change unnecessarily
> # weight 15.000
> alg straw
> hash 0 # rjenkins1
> item 172.19.0.14 weight 6.000
> item 172.19.0.13 weight 3.000
> item 172.19.0.16 weight 3.000
> item 172.19.0.15 weight 3.000
> }
> root default {
> id -1 # do not change unnecessarily
> # weight 15.000
> alg straw
> hash 0 # rjenkins1
> item 0 weight 15.000
> }
> 
> # rules
> rule data {
> ruleset 0
> type replicated
> min_size 1
> max_size 10
> step take default
> step choose firstn 0 type osd
> step emit
> }
> rule metadata {
> ruleset 1
> type replicated
> min_size 1
> max_size 10
> step take default
> step choose firstn 0 type osd
> step emit
> }
> rule rbd {
> ruleset 2
> type replicated
> min_size 1
> max_size 10
> step take default
> step choose firstn 0 type osd
> step emit
> }
> 
> # end crush map
> [root@node-172-19-0-15 ~]# ceph --version
> ceph version 0.56.2 (586538e22afba85c59beda49789ec42024e7a061)
> 
> We do not run any explicit crushtool commands as part of our start up
> at this time.  Should we be?

Do you run mkcephfs?  If you are passing in a ceph.conf to mkcephfs, it is 
still dynamically choosing a rule based on whether you have enough osds (3 
i think?  i forget).  If you are running ceph-mon --mkfs directly (as 
ceph-deploy, chef, juju do), it will always default to osds.

sage

> 
> Regards,
> Mandell Degerness
> 
> On Wed, Jan 30, 2013 at 3:46 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> > The next bobtail point release is ready, and it's looking pretty good.
> > This is an important update for the 0.56.x backport series that fixes a
> > number of bugs and several performance issues. All v0.56.x users are
> > encouraged to upgrade.
> >
> > Notable changes since v0.56.1:
> >
> >  * osd: snapshot trimming fixes
> >  * osd: scrub snapshot metadata
> >  * osd: fix osdmap trimming
> >  * osd: misc peering fixes
> >  * osd: stop heartbeating with peers if internal threads are stuck/hung
> >  * osd: PG removal is friendlier to other workloads
> >  * osd: fix recovery start delay (was causing very slow recovery)
> >  * osd: fix scheduling of explicitly requested scrubs
> >  * osd: fix scrub interval config options
> >  * osd: improve recovery vs client io tuning
> >  * osd: improve 'slow request' warning detail for better diagnosis
> >  * osd: default CRUSH map now distributes across hosts, not OSDs
> >  * osd: fix crash on 32-bit hosts triggered by librbd clients
> >  * librbd: fix error handling when talking to older OSDs
> >  * mon: fix a few rare crashes
> >  * ceph command: ability to easily adjust CRUSH tunables
> >  * radosgw: object copy does not copy source ACLs
> >  * rados command: fix omap command usage
> >  * sysvinit script: set ulimit -n properly on remote hosts
> >  * msgr: fix narrow race with message queuing
> >  * fixed compilation on some old distros (e.g., RHEL 5.x)
> >
> > There are a small number of interface changes related to the default CRUSH
> > rule and scrub interval configuration options. Please see the full release
> > notes.
> >
> > You can get v0.56.2 in the usual fashion:
> >
> >  * Git at git://github.com/ceph/ceph.git
> >  * Tarball at http://ceph.com/download/ceph-0.56.2.tar.gz
> >  * For Debian/Ubuntu packages, see http://ceph.com/docs/master/install/debian
> >  * For RPMs, see http://ceph.com/docs/master/install/rpm
> >
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux