Re: Deep scrub, cache pools, replica 1

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Nov 11, 2014 at 2:32 PM, Christian Balzer <chibi@xxxxxxx> wrote:
> On Tue, 11 Nov 2014 10:21:49 -0800 Gregory Farnum wrote:
>
>> On Mon, Nov 10, 2014 at 10:58 PM, Christian Balzer <chibi@xxxxxxx> wrote:
>> >
>> > Hello,
>> >
>> > One of my clusters has become busy enough (I'm looking at you, evil
>> > Window VMs that I shall banish elsewhere soon) to experience client
>> > noticeable performance impacts during deep scrub.
>> > Before this I instructed all OSDs to deep scrub in parallel at Saturday
>> > night and that finished before Sunday morning.
>> > So for now I'll fire them off one by one to reduce the load.
>> >
>> > Looking forward, that cluster doesn't need more space so instead of
>> > adding more hosts and OSDs I was thinking of a cache pool instead.
>> >
>> > I suppose that will keep the clients happy while the slow pool gets
>> > scrubbed.
>> > Is there anybody who tested cache pools with Firefly and compared the
>> > performance to Giant?
>> >
>> > For testing I'm currently playing with a single storage node and 8 SSD
>> > backed OSDs.
>> > Now what very much blew my mind is that a pool with a replication of 1
>> > still does quite the impressive read orgy, clearly reading all the
>> > data in the PGs.
>> > Why? And what is it comparing that data with, the cosmic background
>> > radiation?
>>
>> Yeah, cache pools currently do full-object promotions whenever an
>> object is accessed. There are some ideas and projects to improve this
>> or reduce its effects, but they're mostly just getting started.
> Thanks for confirming that, so probably not much better than Firefly
> _aside_ from the fact that SSD pools should be quite a bit faster in and
> by themselves in Giant.
> Guess there is no other way to find out than to test things, I have a
> feeling that determining the "hot" working set otherwise will be rather
> difficult.
>
>> At least, I assume that's what you mean by a read orgy; perhaps you
>> are seeing something else entirely?
>>
> Indeed I did, this was just an observation that any pool with a replica of
> 1 will still read ALL the data during a deep-scrub. What good would that
> do?

Oh, I see what you're saying; you mean it was reading all the data
during a scrub, not just that it was promoting things.

Anyway, reading all the data during a deep scrub verifies that we
*can* read all the data. That's one of the fundamental tasks of
scrubbing data in a storage system. It's often accompanied by other
checks or recovery behaviors to easily repair issues that are
discovered, but simply maintaining confidence that the data actually
exists is the principle goal. :)
-Greg
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux