Hello,
this shows the content of crush-map file, what content should I change
for selecting osd instead of host? thanks in advance.
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable straw_calc_version 1
# devices
device 0 osd.0
device 1 osd.1
device 2 osd.2
# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root
# buckets
host ceph2 {
id -2 # do not change unnecessarily
# weight 0.240
alg straw
hash 0 # rjenkins1
item osd.0 weight 0.080
item osd.1 weight 0.080
item osd.2 weight 0.080
}
root default {
id -1 # do not change unnecessarily
# weight 0.240
alg straw
hash 0 # rjenkins1
item ceph2 weight 0.240
}
# rules
rule replicated_ruleset {
ruleset 0
type replicated
min_size 1
max_size 10
step take default
step chooseleaf firstn 0 type host
step emit
}
# end crush map
On 2015/10/29 星期四 10:03, Robert LeBlanc wrote:
A Google search should have lead you the rest of the way.
Follow this [1] and in the rule section on step choose leaf change host
to osd. You won't need to change the configuration this way, it is saved
in the CRUSH map.
[1]
http://docs.ceph.com/docs/master/rados/operations/crush-map/#editing-a-crush-map
Robert LeBlanc
Sent from a mobile device please excuse any typos.
On Oct 28, 2015 7:46 PM, "Wah Peng" <wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>> wrote:
Is there a ceph sub-command existing instead of changing the config
file? :)
On 2015/10/29 星期四 9:24, Li, Chengyuan wrote:
Try " osd crush chooseleaf type = 0" in /etc/ceph/<clustername>.conf
Regards,
CY.
-----Original Message-----
From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx
<mailto:ceph-users-bounces@xxxxxxxxxxxxxx>] On Behalf Of Wah Peng
Sent: 2015年10月29日 9:14
To: Robert LeBlanc
Cc: Lindsay Mathieson; Gurjar, Unmesh; ceph-users@xxxxxxxxxxxxxx
<mailto:ceph-users@xxxxxxxxxxxxxx>
Subject: Re: creating+incomplete issues
wow this sounds hard to me. can you show the details?
thanks a lot.
On 2015/10/29 星期四 9:01, Robert LeBlanc wrote:
You need to change the CRUSH map to select osd instead of host.
Robert LeBlanc
Sent from a mobile device please excuse any typos.
On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>> wrote:
$ ceph osd tree
# id weight type name up/down reweight
-1 0.24 root default
-2 0.24 host ceph2
0 0.07999 osd.0 up 1
1 0.07999 osd.1 up 1
2 0.07999 osd.2 up 1
On 2015/10/29 星期四 8:55, Robert LeBlanc wrote:
Please paste 'ceph osd tree'.
Robert LeBlanc
Sent from a mobile device please excuse any typos.
On Oct 28, 2015 6:54 PM, "Wah Peng"
<wah_peng@xxxxxxxxxxxx <mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx> <mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>>>
wrote:
Hello,
Just did it, but still no good health. can
you help? thanks.
ceph@ceph:~/my-cluster$ ceph osd stat
osdmap e24: 3 osds: 3 up, 3 in
ceph@ceph:~/my-cluster$ ceph health
HEALTH_WARN 89 pgs degraded; 67 pgs
incomplete; 67 pgs stuck
inactive; 192 pgs stuck unclean
On 2015/10/29 星期四 8:38, Lindsay Mathieson
wrote:
On 29 October 2015 at 10:29, Wah Peng
<wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx> <mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>
<mailto:wah_peng@xxxxxxxxxxxx <mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>>>>>
wrote:
$ ceph osd stat
osdmap e18: 2 osds: 2 up, 2 in
this is what it shows.
does it mean I need to add up to 3
osds? I just
use the
default setup.
If you went with the defaults then your
pool size will
be 3,
meaning it
needs 3 copies of the data (replica 3) to
be valid - as
you only
have
two nodes/osd's that can never happen :)
Your options are:
- Add another node and osd.
or
- reduce the size to 2.(ceph osd set
<poolname> size
2)
--
Lindsay
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
<mailto:ceph-users@xxxxxxxxxxxxxx
<mailto:ceph-users@xxxxxxxxxxxxxx>>
<mailto:ceph-users@xxxxxxxxxxxxxx
<mailto:ceph-users@xxxxxxxxxxxxxx>
<mailto:ceph-users@xxxxxxxxxxxxxx
<mailto:ceph-users@xxxxxxxxxxxxxx>>>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com