I don't see the more complete log. -Sam On Mon, Feb 11, 2013 at 11:12 AM, Mandell Degerness <mandell@xxxxxxxxxxxxxxx> wrote: > Anyone have any thoughts on this??? It looks like I may have to wipe > out the OSDs effected and rebuild them, but I'm afraid that may result > in data loss because of the old OSD first crush map in place :(. > > On Fri, Feb 8, 2013 at 1:36 PM, Mandell Degerness > <mandell@xxxxxxxxxxxxxxx> wrote: >> We ran into an error which appears very much like a bug fixed in 0.44. >> >> This cluster is running version: >> >> ceph version 0.48.1argonaut (commit:a7ad701b9bd479f20429f19e6fea7373ca6bba7c) >> >> The error line is: >> >> Feb 8 18:50:07 192.168.8.14 ceph-osd: 2013-02-08 18:50:07.545682 >> 7f40f9f08700 0 filestore(/mnt/osd97) error (17) File exists not >> handled on operation 20 (11279344.0.0, or op 0, counting from 0) >> >> A more complete log is attached. >> >> First question: is this a know bug fixed in more recent versions? >> >> Second question: is there any hope of recovery? > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html