Hi, all
I'm a newbie to Ceph, and just setup a whole new Ceph cluster (0.87) with two servers. But when its status is always warning: [root@serverA ~]# ceph osd tree # id weight type name up/down reweight -1 62.04 root default -2 36.4 host serverA 0 3.64 osd.0 up 1 2 3.64 osd.2 up 1 1 3.64 osd.1 up 1 3 3.64 osd.3 up 1 4 3.64 osd.4 up 1 5 3.64 osd.5 up 1 6 3.64 osd.6 up 1 7 3.64 osd.7 up 1 8 3.64 osd.8 up 1 9 3.64 osd.9 up 1 -3 25.64 host serverB 10 3.64 osd.10 up 1 11 2 osd.11 up 1 12 2 osd.12 up 1 13 2 osd.13 up 1 14 2 osd.14 up 1 15 2 osd.15 up 1 16 2 osd.16 up 1 17 2 osd.17 up 1 18 2 osd.18 up 1 19 2 osd.19 up 1 20 2 osd.20 up 1 21 2 osd.21 up 1 [root@serverA ~]# ceph -s cluster ???????????????169715 health HEALTH_WARN 493 pgs degraded; 19 pgs peering; 19 pgs stuck inactive; 512 pgs stuck unclean; 493 pgs undersized monmap e1: 2 mons at {serverB=10.??????.78:6789/0,serverA=10.?????.80:6789/0}, election epoch 10, quorum 0,1 mac0090fa6aaf7a,mac0090fa6ab68a osdmap e92634: 22 osds: 22 up, 22 in pgmap v189018: 512 pgs, 1 pools, 0 bytes data, 0 objects 49099 MB used, 63427 GB / 63475 GB avail 493 active+undersized+degraded 19 creating+peering [root@serverA ~]# rados -p test31 ls 2015-03-30 09:57:18.607143 7f5251fcf700 0 -- :/1005913 >> 10.??????.78:6789/0 pipe(0x140a370 sd=3 :0 s=1 pgs=0 cs=0 l=1 c=0x140a600).fault 2015-03-30 09:57:21.610994 7f52484ad700 0 -- 10.????.80:0/1005913 >> 10.????.78:6835/27111 pipe(0x140e010 sd=4 :0 s=1 pgs=0 cs=0 l=1 c=0x140e2a0).fault 2015-03-30 10:02:21.650191 7f52482ab700 0 -- 10.????.80:0/1005913 >> 10.????78:6835/27111 pipe(0x7f5238016c80 sd=5 :0 s=1 pgs=0 cs=0 l=1 c=0x7f5238016f10).fault * serverA is 10.???.80, serverB is 10.????.78 * ntpdate is updated * I tried to remove the pool and re-create it, and clean up all objects inside, but no change at all * firewall are both shutoff Any clue is welcomed, thanks. |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com