On Wed, May 4, 2016 at 12:00 AM, Nikola Ciprich <nikola.ciprich@xxxxxxxxxxx> wrote: > Hi, > > I was doing some performance tuning on test cluster of just 2 > nodes (each 10 OSDs). I have test pool of 2 replicas (size=2, min_size=2) > > then one of OSD crashed due to failing harddrive. All remaining OSDs were > fine, but health status reported one lost object.. > > here's detail: > > "recovery_state": [ > { > "name": "Started\/Primary\/Active", > "enter_time": "2016-05-04 07:59:10.706866", > "might_have_unfound": [ > { > "osd": "0", > "status": "osd is down" > }, > { > "osd": "10", > "status": "already probed" > } > ], > > > it was no important data, so I just discarded it as I don't need > to recover it, but now I'm wondering what is the cause of all this.. > > I have min_size set to 2 and I though that writes are confirmed after > they reach all target OSD journals, no? Is there something specific I should > check? Maybe I have some bug in configuration? Or how else could this object > be lost? Is OSD 0 the one which had a failing hard drive? And OSD 10 is supposed to be fine? In general what you're saying does make it sound like something under the Ceph code lost objects, but if one of those OSDs has never had a problem I'm not sure what it could be. (The most common failure mode is power loss while the user has barriers turned off, or a RAID card misconfigured, or similar.) -Greg > > I'd be grateful for any info > > br > > nik > > > > > > -- > ------------------------------------- > Ing. Nikola CIPRICH > LinuxBox.cz, s.r.o. > 28.rijna 168, 709 00 Ostrava > > tel.: +420 591 166 214 > fax: +420 596 621 273 > mobil: +420 777 093 799 > www.linuxbox.cz > > mobil servis: +420 737 238 656 > email servis: servis@xxxxxxxxxxx > ------------------------------------- > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com