I got swamped today. I should be able to look tomorrow. Sorry! -Sam On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG <s.priebe@xxxxxxxxxxxx> wrote: > Did you take a look? > > Stefan > > Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@xxxxxxxxxxx>: > >> Great! I'll take a look on Monday. >> -Sam >> >> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@xxxxxxxxxxxx> wrote: >>> Hi Samual, >>> >>> Am 09.08.2013 23:44, schrieb Samuel Just: >>> >>>> I think Stefan's problem is probably distinct from Mike's. >>>> >>>> Stefan: Can you reproduce the problem with >>>> >>>> debug osd = 20 >>>> debug filestore = 20 >>>> debug ms = 1 >>>> debug optracker = 20 >>>> >>>> on a few osds (including the restarted osd), and upload those osd logs >>>> along with the ceph.log from before killing the osd until after the >>>> cluster becomes clean again? >>> >>> >>> done - you'll find the logs at cephdrop folder: >>> slow_requests_recovering_cuttlefish >>> >>> osd.52 was the one recovering >>> >>> Thanks! >>> >>> Greets, >>> Stefan >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html