Re: v0.56.2 released

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We are doing the latter.  What is the best way to force it to host
first selection?

On Tue, Feb 5, 2013 at 11:34 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
> On Tue, 5 Feb 2013, Mandell Degerness wrote:
>> We are using v0.56.2 and it still seems that the default crushmap is
>> osd centered.  Here is the crushmap as dumped:
>>
>> [root@node-172-19-0-15 ~]# cat crush.txt
>> # begin crush map
>>
>> # devices
>> device 0 osd.0
>> device 1 osd.1
>> device 2 osd.2
>> device 3 osd.3
>> device 4 osd.4
>> device 5 osd.5
>> device 6 osd.6
>> device 7 osd.7
>> device 8 osd.8
>> device 9 osd.9
>> device 10 osd.10
>> device 11 osd.11
>> device 12 osd.12
>> device 13 osd.13
>> device 14 osd.14
>>
>> # types
>> type 0 osd
>> type 1 host
>> type 2 rack
>> type 3 row
>> type 4 room
>> type 5 datacenter
>> type 6 root
>>
>> # buckets
>> host 172.19.0.14 {
>> id -2 # do not change unnecessarily
>> # weight 6.000
>> alg straw
>> hash 0 # rjenkins1
>> item osd.0 weight 1.000
>> item osd.4 weight 1.000
>> item osd.5 weight 1.000
>> item osd.12 weight 1.000
>> item osd.13 weight 1.000
>> item osd.14 weight 1.000
>> }
>> host 172.19.0.13 {
>> id -4 # do not change unnecessarily
>> # weight 3.000
>> alg straw
>> hash 0 # rjenkins1
>> item osd.2 weight 1.000
>> item osd.7 weight 1.000
>> item osd.9 weight 1.000
>> }
>> host 172.19.0.16 {
>> id -5 # do not change unnecessarily
>> # weight 3.000
>> alg straw
>> hash 0 # rjenkins1
>> item osd.3 weight 1.000
>> item osd.6 weight 1.000
>> item osd.10 weight 1.000
>> }
>> host 172.19.0.15 {
>> id -6 # do not change unnecessarily
>> # weight 3.000
>> alg straw
>> hash 0 # rjenkins1
>> item osd.1 weight 1.000
>> item osd.8 weight 1.000
>> item osd.11 weight 1.000
>> }
>> rack 0 {
>> id -3 # do not change unnecessarily
>> # weight 15.000
>> alg straw
>> hash 0 # rjenkins1
>> item 172.19.0.14 weight 6.000
>> item 172.19.0.13 weight 3.000
>> item 172.19.0.16 weight 3.000
>> item 172.19.0.15 weight 3.000
>> }
>> root default {
>> id -1 # do not change unnecessarily
>> # weight 15.000
>> alg straw
>> hash 0 # rjenkins1
>> item 0 weight 15.000
>> }
>>
>> # rules
>> rule data {
>> ruleset 0
>> type replicated
>> min_size 1
>> max_size 10
>> step take default
>> step choose firstn 0 type osd
>> step emit
>> }
>> rule metadata {
>> ruleset 1
>> type replicated
>> min_size 1
>> max_size 10
>> step take default
>> step choose firstn 0 type osd
>> step emit
>> }
>> rule rbd {
>> ruleset 2
>> type replicated
>> min_size 1
>> max_size 10
>> step take default
>> step choose firstn 0 type osd
>> step emit
>> }
>>
>> # end crush map
>> [root@node-172-19-0-15 ~]# ceph --version
>> ceph version 0.56.2 (586538e22afba85c59beda49789ec42024e7a061)
>>
>> We do not run any explicit crushtool commands as part of our start up
>> at this time.  Should we be?
>
> Do you run mkcephfs?  If you are passing in a ceph.conf to mkcephfs, it is
> still dynamically choosing a rule based on whether you have enough osds (3
> i think?  i forget).  If you are running ceph-mon --mkfs directly (as
> ceph-deploy, chef, juju do), it will always default to osds.
>
> sage
>
>>
>> Regards,
>> Mandell Degerness
>>
>> On Wed, Jan 30, 2013 at 3:46 PM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>> > The next bobtail point release is ready, and it's looking pretty good.
>> > This is an important update for the 0.56.x backport series that fixes a
>> > number of bugs and several performance issues. All v0.56.x users are
>> > encouraged to upgrade.
>> >
>> > Notable changes since v0.56.1:
>> >
>> >  * osd: snapshot trimming fixes
>> >  * osd: scrub snapshot metadata
>> >  * osd: fix osdmap trimming
>> >  * osd: misc peering fixes
>> >  * osd: stop heartbeating with peers if internal threads are stuck/hung
>> >  * osd: PG removal is friendlier to other workloads
>> >  * osd: fix recovery start delay (was causing very slow recovery)
>> >  * osd: fix scheduling of explicitly requested scrubs
>> >  * osd: fix scrub interval config options
>> >  * osd: improve recovery vs client io tuning
>> >  * osd: improve 'slow request' warning detail for better diagnosis
>> >  * osd: default CRUSH map now distributes across hosts, not OSDs
>> >  * osd: fix crash on 32-bit hosts triggered by librbd clients
>> >  * librbd: fix error handling when talking to older OSDs
>> >  * mon: fix a few rare crashes
>> >  * ceph command: ability to easily adjust CRUSH tunables
>> >  * radosgw: object copy does not copy source ACLs
>> >  * rados command: fix omap command usage
>> >  * sysvinit script: set ulimit -n properly on remote hosts
>> >  * msgr: fix narrow race with message queuing
>> >  * fixed compilation on some old distros (e.g., RHEL 5.x)
>> >
>> > There are a small number of interface changes related to the default CRUSH
>> > rule and scrub interval configuration options. Please see the full release
>> > notes.
>> >
>> > You can get v0.56.2 in the usual fashion:
>> >
>> >  * Git at git://github.com/ceph/ceph.git
>> >  * Tarball at http://ceph.com/download/ceph-0.56.2.tar.gz
>> >  * For Debian/Ubuntu packages, see http://ceph.com/docs/master/install/debian
>> >  * For RPMs, see http://ceph.com/docs/master/install/rpm
>> >
>> > --
>> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> > the body of a message to majordomo@xxxxxxxxxxxxxxx
>> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux