Re: ceph stays degraded after crushmap rearrangement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
Am 05.01.2013 18:11, schrieb Stefan Priebe:
Hi,

i just stopped EVERYTHING and have now started ALL osds again. It seems
to recover now. But here is the output.
Just an illusion. Still hangs.

# ceph -s
health HEALTH_WARN 934 pgs degraded; 23 pgs down; 1887 pgs peering; 1330 pgs stale; 670 pgs stuck inactive; 882 pgs stuck stale; 7632 pgs stuck unclean; recovery 4811/79122 degraded (6.080%) monmap e1: 3 mons at {a=10.255.0.100:6789/0,b=10.255.0.101:6789/0,c=10.255.0.102:6789/0}, election epoch 1996, quorum 0,1,2 a,b,c
   osdmap e8393: 24 osds: 24 up, 24 in
pgmap v2212487: 7632 pgs: 475 peering, 4013 active+remapped, 18 down+peering, 490 active+degraded, 798 stale+active+remapped, 1 active+replay+degraded, 1305 remapped+peering, 84 stale+remapped+peering, 5 stale+down+remapped+peering, 364 stale+active+degraded+remapped, 79 stale+active+replay+degraded+remapped; 152 GB data, 314 GB used, 5021 GB / 5336 GB avail; 4811/79122 degraded (6.080%)
   mdsmap e1: 0/0/1 up

Greets,
Stefan
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux