You need to change the CRUSH map to select osd instead of host.
Robert LeBlanc
Sent from a mobile device please excuse any typos.
On Oct 28, 2015 7:00 PM, "Wah Peng" <wah_peng@xxxxxxxxxxxx> wrote:
$ ceph osd tree
# id weight type name up/down reweight
-1 0.24 root default
-2 0.24 host ceph2
0 0.07999 osd.0 up 1
1 0.07999 osd.1 up 1
2 0.07999 osd.2 up 1
On 2015/10/29 星期四 8:55, Robert LeBlanc wrote:
Please paste 'ceph osd tree'.
Robert LeBlanc
Sent from a mobile device please excuse any typos.
On Oct 28, 2015 6:54 PM, "Wah Peng" <wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>> wrote:
Hello,
Just did it, but still no good health. can you help? thanks.
ceph@ceph:~/my-cluster$ ceph osd stat
osdmap e24: 3 osds: 3 up, 3 in
ceph@ceph:~/my-cluster$ ceph health
HEALTH_WARN 89 pgs degraded; 67 pgs incomplete; 67 pgs stuck
inactive; 192 pgs stuck unclean
On 2015/10/29 星期四 8:38, Lindsay Mathieson wrote:
On 29 October 2015 at 10:29, Wah Peng <wah_peng@xxxxxxxxxxxx
<mailto:wah_peng@xxxxxxxxxxxx>
<mailto:wah_peng@xxxxxxxxxxxx <mailto:wah_peng@xxxxxxxxxxxx>>>
wrote:
$ ceph osd stat
osdmap e18: 2 osds: 2 up, 2 in
this is what it shows.
does it mean I need to add up to 3 osds? I just use the
default setup.
If you went with the defaults then your pool size will be 3,
meaning it
needs 3 copies of the data (replica 3) to be valid - as you only
have
two nodes/osd's that can never happen :)
Your options are:
- Add another node and osd.
or
- reduce the size to 2.(ceph osd set <poolname> size 2)
--
Lindsay
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com