Le 23/11/2015 19:58, Jose Tavares a
écrit :
One weak data checksum which is not available to the kernel, yes. Filesystems and applications on top of them may use stronger checksums and handle read problems that the drives can't detect themselves.
If it fails AFAIK from past experience it doesn't reset by itself, the kernel driver in charge of the device will receive an IO error and will retry the IO several times. One of those latter attempts might succeed (errors aren't always repeatable) and eventually after a timeout it will try to reset the interface with the drive and the drive itself (the kernel doesn't know where the problem is only that it didn't get the result it was expecting). While this happen I believe the filesystem/md/lvm/... stack can receive an IO error (the timeout at their level might not be the same as the timeout at the device level). So some errors can be masked to md and some can percolate through. In the later case, yes the md array will drop the device.
Then you can move this drive to the trash pile/ask for a replacement. It is basically unusable.
There are 2 weights I'm aware of, the crush weight for an OSD and the temporary OSD weight. The first is the basic weight used by crush to choose how to split your data (an OSD with a weight of 2 is expected to get roughly twice the amount of data of an OSD with a weight of 1 on a normal Ceph cluster), the second is used for temporary adjustments when a OSD gets temporarily overused (during cluster wide rebalancing typically) and is reset when the OSD rejoins the cluster (marked in). Neither of these weights has anything to do with the OSD underlying device health ("being bad"). Lionel |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com