Re: ceph stays degraded after crushmap rearrangement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sat, 5 Jan 2013, Stefan Priebe wrote:
> Hello,
> 
> now i cannot even access an rbd image anymore.
> 
> Hanging status is now:
> 2013-01-05 18:01:21.736298 mon.0 [INF] pgmap v2212193: 7632 pgs: 1 stale, 10
> peering, 14 stale+peering, 1 stale+remapped, 1807 stale+active+remapped, 1
> stale+active+degraded, 2587 remapped+peering, 1767 stale+remapped+peering,
> 1341 stale+active+degraded+remapped, 103
> stale+active+replay+degraded+remapped; 152 GB data, 313 GB used, 5022 GB /
> 5336 GB avail; 7647/79122 degraded (9.665%)

It looks like some of the ceph-osds stopped.

Are all daemons running the testing branch code?  What does 'ceph -s' say?  
Or 'ceph pg <pgid> query' on a random active+remapped pgid?

sage


> 
> 
> Stefan
> Am 05.01.2013 17:40, schrieb Stefan Priebe:
> > Hi list,
> > 
> > i've rearranged my crushmap. Ceph was degraded about 18% and was
> > recovering / rearranging fine.
> > 
> > But now it stays still and degraded status is rising??
> > 
> > 2013-01-05 17:35:40.906587 mon.0 [INF] pgmap v2211269: 7632 pgs: 7632
> > active+remapped; 152 GB data, 312 GB used, 5023 GB / 5336 GB avail;
> > 22/79086 degraded (0.028%)
> > 
> > ...
> > 
> > 2013-01-05 17:37:50.142106 mon.0 [INF] pgmap v2211386: 7632 pgs: 7632
> > active+remapped; 152 GB data, 312 GB used, 5023 GB / 5336 GB avail;
> > 24/79090 degraded (0.030%)
> > 
> > ..
> > 
> > 2013-01-05 17:40:35.292054 mon.0 [INF] pgmap v2211526: 7632 pgs: 7632
> > active+remapped; 152 GB data, 313 GB used, 5023 GB / 5336 GB avail;
> > 32/79106 degraded (0.040%)
> > 
> > I'm on currect testing branch.
> > 
> > Greets,
> > Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux