On Sat, 5 Jan 2013, Stefan Priebe wrote: > Hello, > > now i cannot even access an rbd image anymore. > > Hanging status is now: > 2013-01-05 18:01:21.736298 mon.0 [INF] pgmap v2212193: 7632 pgs: 1 stale, 10 > peering, 14 stale+peering, 1 stale+remapped, 1807 stale+active+remapped, 1 > stale+active+degraded, 2587 remapped+peering, 1767 stale+remapped+peering, > 1341 stale+active+degraded+remapped, 103 > stale+active+replay+degraded+remapped; 152 GB data, 313 GB used, 5022 GB / > 5336 GB avail; 7647/79122 degraded (9.665%) It looks like some of the ceph-osds stopped. Are all daemons running the testing branch code? What does 'ceph -s' say? Or 'ceph pg <pgid> query' on a random active+remapped pgid? sage > > > Stefan > Am 05.01.2013 17:40, schrieb Stefan Priebe: > > Hi list, > > > > i've rearranged my crushmap. Ceph was degraded about 18% and was > > recovering / rearranging fine. > > > > But now it stays still and degraded status is rising?? > > > > 2013-01-05 17:35:40.906587 mon.0 [INF] pgmap v2211269: 7632 pgs: 7632 > > active+remapped; 152 GB data, 312 GB used, 5023 GB / 5336 GB avail; > > 22/79086 degraded (0.028%) > > > > ... > > > > 2013-01-05 17:37:50.142106 mon.0 [INF] pgmap v2211386: 7632 pgs: 7632 > > active+remapped; 152 GB data, 312 GB used, 5023 GB / 5336 GB avail; > > 24/79090 degraded (0.030%) > > > > .. > > > > 2013-01-05 17:40:35.292054 mon.0 [INF] pgmap v2211526: 7632 pgs: 7632 > > active+remapped; 152 GB data, 313 GB used, 5023 GB / 5336 GB avail; > > 32/79106 degraded (0.040%) > > > > I'm on currect testing branch. > > > > Greets, > > Stefan > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html