Ceph inconsistency after deep-scrub

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

During deep-scrub Ceph discovered some inconsistency between OSDs on my
cluster (size 3, min size 2). I have fund broken object and calculated
md5sum of it on each OSD (osd.195 is acting_primary):
 osd.195 - md5sum_aaaa
 osd.40 - md5sum_aaaa
 osd.314 - md5sum_bbbb

I run ceph pg repair and Ceph successfully reported that everything went
OK. I checked md5sum of the objects again:
 osd.195 - md5sum_bbbb
 osd.40 - md5sum_bbbb
 osd.314 - md5sum_bbbb

This is a bit odd. How Ceph decides which copy is the correct one? Based
on last modification time/sequence number (or similar)? If yes, then why
object can be stored on one node only? If not, then why Ceph selected
osd.314 as a correct one? What would happen if osd.314 goes down? Will
ceph return wrong (old?) data, even with three copies and no failure in
the cluster?

For now I'm unable to reproduce this on my test cluster. I'll post here
if I reproduce this.

Thanks for any help,
PS

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux