Re: [ceph-calamari] disk failure prediction

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 19 Feb 2015, John Spray wrote:
> 
> On 18/02/2015 23:20, Sage Weil wrote:
> > We wouldn't see
> > quite the same results since our "raid sets" are effectively entire pools
>
> I think we could do better than pool-wide, e.g. if multiple drives in one
> chassis are at risk (where PG stores at most one copy per chassis), we can
> identify that as less severe than the general case where multiple at-risk
> drives might be in the same PG.  Making it CRUSH-aware like this would be a
> good hook for users to take advantage of the ceph/calamari SMART monitoring
> rather than rolling their own.

Yeah, sounds good.  The big question in my mind is whether we should try 
to pull this into the osd/mon or have calamari do it.  It seems like a 
good fit for calamari...

BTW, a bit more color on the original paper (after talking to Paul): the 
EMC workload in the paper was backup/archival with heavy heavy write, and 
any time there was a read failure it triggered a rewrite and triggered a 
relocated sector.  Other studies have shown some pretty different results.  
For example, one showed that the best preditor was actually not SMART at 
all but (carefully measured) read latency.

In any case, it seems like the bits that are gathering and aggregating 
SMART should be general, and we should make it easy to plug in various 
policies (or delegate to an external agent) for responding to that data. 

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux