New added osd always down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi, cepher

My cluster has a big problem.
ceph version: 0.80.10
1. OSD are full, i can't delete volume, the io seems blocked. when i rm a image, here is the error message:
sudo rbd rm ff3a6870-24cb-427a-979b-6b9b257032c3 -p vol_ssd
2015-11-24 14:14:26.418016 7f9b900a5780 -1 librbd::ImageCtx: error finding header: (2) No such file or directory
2015-11-24 14:14:26.418237 7f9b900a5780  0 client.9237071.objecter  FULL, paused modify 0xcc5870 tid 3

follow is message from ceph -w:
 cluster 19eeb168-7dce-48ae-afb2-b6d1e1e29be4
     health HEALTH_ERR 1164 pgs backfill_toofull; 448 pgs degraded; 12 pgs incomplete; 12 pgs stuck inactive; 1224 pgs stuck unclean; recovery 1039912/5491280 objects degraded (18.938%); 35 full osd(s); 4 near full osd(s)
     monmap e2: 3 mons at {10-180-0-30=10.180.0.30:6789/0,10-180-0-31=10.180.0.31:6789/0,10-180-0-34=10.180.0.34:6789/0}, election epoch 114, quorum 0,1,2 10-180-0-30,10-180-0-31,10-180-0-34
     osdmap e12196: 44 osds: 39 up, 39 in
            flags full
      pgmap v461411: 4096 pgs, 3 pools, 6119 GB data, 1525 kobjects
            12314 GB used, 607 GB / 12921 GB avail
            1039912/5491280 objects degraded (18.938%)
                  38 active+degraded+remapped
                 754 active+remapped+backfill_toofull
                2872 active+clean
                  10 active+remapped
                 410 active+degraded+remapped+backfill_toofull
                  12 remapped+incomplete

2015-11-24 14:17:50.716166 osd.8 [WRN] OSD near full (95%)
2015-11-24 14:18:01.139994 osd.40 [WRN] OSD near full (95%)
2015-11-24 14:17:53.308538 osd.22 [WRN] OSD near full (95%)

2. I try to add some new osd, but it always be a down state.
ceph osd tree|grep down
# id	weight	type name	up/down	reweight
21	0.4			osd.21	down	0	
2	0.36			osd.2	down	0	
4	0.4			osd.4	down	0	

ceph osd dump:
osd.2 down out weight 0 up_from 8751 up_thru 8755 down_at 8766 last_clean_interval [8224,8746) 10.180.0.30:6821/40125 10.180.0.30:6827/40125 10.180.0.30:6828/40125 10.180.0.30:6829/40125 autoout,exists f1dc9181-ed70-48fb-95fa-cc568fee7b98

And here is the log of osd.2:
2015-11-24 14:21:38.547551 7ff48e8cb700 10 osd.2 0 do_waiters -- start 
2015-11-24 14:21:38.547554 7ff48e8cb700 10 osd.2 0 do_waiters -- finish
2015-11-24 14:21:39.386455 7ff47486f700 20 osd.2 0 update_osd_stat osd_stat(33360 kB used, 367 GB avail, 367 GB total, peers []/[] op hist [])
2015-11-24 14:21:39.386473 7ff47486f700  5 osd.2 0 heartbeat: osd_stat(33360 kB used, 367 GB avail, 367 GB total, peers []/[] op hist [])
2015-11-24 14:21:39.547615 7ff48e8cb700  5 osd.2 0 tick

What's wrong with my cluster?
			
--------------
hzwulibin
2015-11-24
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux