On 11/06/2013 03:33 AM, Sage Weil wrote:
On Wed, 6 Nov 2013, Loic Dachary wrote:
Hi Ceph,
People from Western Digital suggested ways to better take advantage of
the disk error reporting. They gave two examples that struck my
imagination. First there are errors that look like the disk is dying (
read / write failures ) but it's only a transient problem and the driver
should be able to make the difference by properly interpreting the
available information. They said that the prolonged life you get if you
don't decommission a disk that only has a transient error is
This make me think we really need to build or integrate with some generic
SMART reporting infrastructure so that we can identify disks that are
failing or going to fail. What to do with that information is another
question; initially I would lean toward just marking the disk out, but
there may be smarter alternatives to investigate.
significant. The second example is when one head out of ten fails :
disks can keep working with the nine remaining heads. Losing 1/10 of the
disk is likely to result in a full re-install of the Ceph osd. But,
again, the disk could keep going after that, with 9/10 of its original
capacity. And Ceph is good at handling osd failures.
Yeah...but if you lose 1/10 of a block device any existing local file
system is going to blow up. I suspet this is something that newgangled
interfaces like Kinetic will be much better at. Even then, though, it is
challenging for anything sitting above to cope with losing some random
subset of it's data underneath. To a first approximation, for this to be
useful, the fs and disk would need to keep, say, all teh data in a
particular PG confined to a single platter, so that when a head goes the
other PGs are still fully intact and usage. It is probably a long way to
get from here to there...
Putting my sysadmin hat on:
Once I know a drive has had a head failure, do I trust that the rest of
the drive isn't going to go at an inconvenient moment vs just fixing it
right now when it's not 3AM on Christmas morning? (true story) As good
as Ceph is, do I trust that Ceph is smart enough to prevent spreading
corrupt data all over the cluster if I leave bad disks in place and they
start doing terrible things to the data?
Mark
All this is news to me and sounds really cool. But I'm sure there are
people who already know about it and I'm eager to hear their opinion :-)
Cheers
--
Lo?c Dachary, Artisan Logiciel Libre
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com