Re: RocksDB tuning

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 10 Jun 2016, Igor Fedotov wrote:
> An update:
> 
> I found that my previous results were invalid - SyntheticWorkloadState had an
> odd swap for offset > len case... Made a brief fix.
> 
> Now onode size with csum raises up to 38K, without csum - 28K.
> 
> For csum case there is 350 lextents and about 170 blobs
> 
> For no csum - 343 lextents and about 170 blobs.
> 
> (blobs counting is very inaccurate!)
> 
> Potentially we shouldn't have >64 blobs per 4M thus looks like some issues in
> the write path...

Synthetic randomly twiddles alloc hints, which means some of those 
blobs are probably getting compressed.  I suspect if you set 'bluestore 
compression = none' it'll drop back down to 64.

There is still a problem with compression, though.  I think the write path 
should look at whether we are obscuring an existing blob with more than N 
layers (where N is probably 2?) and if so do a read+write 'compaction' to 
flatten it.  That (or something like it) should get us a ~2x bound on the 
worst case lextent count (in this case ~128)...

sage

> 
> And CSum vs. NoCsum differenct looks pretty consistent - 170 blobs * 4 byte *
> 16 values = 10880
> 
> Branch's @github been updated with corresponding fixes.
> 
> Thanks,
> Igor.
> 
> On 10.06.2016 19:06, Allen Samuels wrote:
> > Let's see, 4MB is 2^22 bytes. If we storage a checksum for each 2^12 bytes
> > that's 2^10 checksums at 2^2 bytes each is 2^12 bytes.
> > 
> > So with optimal encoding, the checksum baggage shouldn't be more than 4KB
> > per oNode.
> > 
> > But you're seeing 13K as the upper bound on the onode size.
> > 
> > In the worst case, you'll need at least another block address (8 bytes
> > currently) and length (another 8 bytes) [though as I point out, the length
> > is something that can be optimized out] So worst case, this encoding would
> > be an addition 16KB per onode.
> > 
> > I suspect you're not at the worst-case yet :)
> > 
> > Allen Samuels
> > SanDisk |a Western Digital brand
> > 2880 Junction Avenue, Milpitas, CA 95134
> > T: +1 408 801 7030| M: +1 408 780 6416
> > allen.samuels@xxxxxxxxxxx
> > 
> > 
> > > -----Original Message-----
> > > From: Igor Fedotov [mailto:ifedotov@xxxxxxxxxxxx]
> > > Sent: Friday, June 10, 2016 8:58 AM
> > > To: Sage Weil <sweil@xxxxxxxxxx>; Somnath Roy
> > > <Somnath.Roy@xxxxxxxxxxx>
> > > Cc: Allen Samuels <Allen.Samuels@xxxxxxxxxxx>; Mark Nelson
> > > <mnelson@xxxxxxxxxx>; Manavalan Krishnan
> > > <Manavalan.Krishnan@xxxxxxxxxxx>; Ceph Development <ceph-
> > > devel@xxxxxxxxxxxxxxx>
> > > Subject: Re: RocksDB tuning
> > > 
> > > Just modified store_test synthetic test case to simulate many random 4K
> > > writes to 4M object.
> > > 
> > > With default settings ( crc32c + 4K block) onode size varies from 2K to
> > > ~13K
> > > 
> > > with disabled crc it's ~500 - 1300 bytes.
> > > 
> > > 
> > > Hence the root cause seems to be in csum array.
> > > 
> > > 
> > > Here is the updated branch:
> > > 
> > > https://github.com/ifed01/ceph/tree/wip-bluestore-test-size
> > > 
> > > 
> > > Thanks,
> > > 
> > > Igor
> > > 
> > > 
> > > On 10.06.2016 18:40, Sage Weil wrote:
> > > > On Fri, 10 Jun 2016, Somnath Roy wrote:
> > > > > Just turning off checksum with the below param is not helping, I
> > > > > still need to see the onode size though by enabling debug..Do I need
> > > > > to mkfs
> > > > > (Sage?) as it is still holding checksum of old data I wrote ?
> > > > Yeah.. you'll need to mkfs to blow away the old onodes and blobs with
> > > > csum data.
> > > > 
> > > > As Allen pointed out, this is only part of the problem.. but I'm
> > > > curious how much!
> > > > 
> > > > >           bluestore_csum = false
> > > > >           bluestore_csum_type = none
> > > > > 
> > > > > Here is the snippet of 'dstat'..
> > > > > 
> > > > > ----total-cpu-usage---- -dsk/total- -net/total- ---paging-->
> > > > > usr sys idl wai hiq siq| read  writ| recv  send|  in   out >
> > > > >    41  14  36   5   0   4| 138M  841M| 212M  145M|   0     0 >
> > > > >    42  14  35   5   0   4| 137M  855M| 213M  147M|   0     0 >
> > > > >    40  14  38   5   0   3| 143M  815M| 209M  144M|   0     0 >
> > > > >    40  14  38   5   0   3| 137M  933M| 194M  134M|   0     0 >
> > > > >    42  15  34   5   0   4| 133M  918M| 220M  151M|   0     0 >
> > > > >    35  13  43   6   0   3| 147M  788M| 194M  134M|   0     0 >
> > > > >    31  11  49   6   0   3| 157M  713M| 151M  104M|   0     0 >
> > > > >    39  14  38   5   0   4| 139M  836M| 246M  169M|   0     0 >
> > > > >    40  14  38   5   0   3| 139M  845M| 204M  140M|   0     0 >
> > > > >    40  14  37   5   0   4| 149M  743M| 210M  144M|   0     0 >
> > > > >    42  14  35   5   0   4| 143M  852M| 216M  150M|   0     0 >
> > > > > For example, what last entry is saying that system (with 8 osds) is
> > > receiving 216M of data over network and in response to that it is writing
> > > total
> > > of 852M of data and reading 143M of data. At this time FIO on client side
> > > is
> > > reporting ~35K 4K RW iops.
> > > > > Now, after a min or so, the throughput goes down to barely 1K from FIO
> > > (and very bumpy) and here is the 'dstat' snippet at that time..
> > > > > ----total-cpu-usage---- -dsk/total- -net/total- ---paging-->
> > > > > usr sys idl wai hiq siq| read  writ| recv  send|  in   out >
> > > > >     2   1  83  14   0   0| 220M   58M|4346k 3002k|   0     0 >
> > > > >     2   1  82  14   0   0| 223M   60M|4050k 2919k|   0     0 >
> > > > >     3   1  82  13   0   0| 217M   63M|6403k 4306k|   0     0 >
> > > > >     2   1  83  14   0   0| 226M   54M|2126k 1497k|   0     0 >
> > > > > 
> > > > > So, system is barely receiving anything (~2M) but still writing ~54M
> > > > > of data
> > > and reading 226M of data from disk.
> > > > > After killing fio script , here is the 'dstat' output..
> > > > > 
> > > > > ----total-cpu-usage---- -dsk/total- -net/total- ---paging-->
> > > > > usr sys idl wai hiq siq| read  writ| recv  send|  in   out >
> > > > >     2   1  86  12   0   0| 186M   66M|  28k   26k|   0     0 >
> > > > >     2   1  86  12   0   0| 201M   78M|  20k   21k|   0     0 >
> > > > >     2   1  85  12   0   0| 230M  100M|  24k   24k|   0     0 >
> > > > >     2   1  85  12   0   0| 206M   78M|  21k   20k|   0     0 >
> > > > > 
> > > > > Not receiving anything from client but still writing 78M of data and
> > > > > 206M
> > > of read.
> > > > > Clearly, it is an effect of rocksdb compaction that stalling IO and
> > > > > even if we
> > > increased compaction thread (and other tuning), compaction is not able to
> > > keep up with incoming IO.
> > > > > Thanks & Regards
> > > > > Somnath
> > > > > 
> > > > > -----Original Message-----
> > > > > From: Allen Samuels
> > > > > Sent: Friday, June 10, 2016 8:06 AM
> > > > > To: Sage Weil
> > > > > Cc: Somnath Roy; Mark Nelson; Manavalan Krishnan; Ceph Development
> > > > > Subject: RE: RocksDB tuning
> > > > > 
> > > > > > -----Original Message-----
> > > > > > From: Sage Weil [mailto:sweil@xxxxxxxxxx]
> > > > > > Sent: Friday, June 10, 2016 7:55 AM
> > > > > > To: Allen Samuels <Allen.Samuels@xxxxxxxxxxx>
> > > > > > Cc: Somnath Roy <Somnath.Roy@xxxxxxxxxxx>; Mark Nelson
> > > > > > <mnelson@xxxxxxxxxx>; Manavalan Krishnan
> > > > > > <Manavalan.Krishnan@xxxxxxxxxxx>; Ceph Development <ceph-
> > > > > > devel@xxxxxxxxxxxxxxx>
> > > > > > Subject: RE: RocksDB tuning
> > > > > > 
> > > > > > On Fri, 10 Jun 2016, Allen Samuels wrote:
> > > > > > > Checksums are definitely a part of the problem, but I suspect the
> > > > > > > smaller part of the problem. This particular use-case (random 4K
> > > > > > > overwrites without the WAL stuff) is the worst-case from an
> > > > > > > encoding perspective and highlights the inefficiency in the
> > > > > > > current
> > > code.
> > > > > > > As has been discussed earlier, a specialized encode/decode
> > > > > > > implementation for these data structures is clearly called for.
> > > > > > > 
> > > > > > > IMO, you'll be able to cut the size of this by AT LEAST a factor
> > > > > > > of
> > > > > > > 3 or
> > > > > > > 4 without a lot of effort. The price will be somewhat increase CPU
> > > > > > > cost for the serialize/deserialize operation.
> > > > > > > 
> > > > > > > If you think of this as an application-specific data compression
> > > > > > > problem, here is a short list of potential compression
> > > > > > > opportunities.
> > > > > > > 
> > > > > > > (1) Encoded sizes and offsets are 8-byte byte values, converting
> > > > > > > these too
> > > > > > block values will drop 9 or 12 bits from each value. Also, the
> > > > > > ranges for these values is usually only 2^22 -- often much less.
> > > > > > Meaning that there's 3-5 bytes of zeros at the top of each word that
> > > > > > can
> > > be dropped.
> > > > > > > (2) Encoded device addresses are often less than 2^32, meaning
> > > > > > > there's 3-4
> > > > > > bytes of zeros at the top of each word that can be dropped.
> > > > > > >    (3) Encoded offsets and sizes are often exactly "1" block,
> > > > > > > clever
> > > > > > > choices of
> > > > > > formatting can eliminate these entirely.
> > > > > > > IMO, an optimized encoded form of the extent table will be around
> > > > > > > 1/4 of the current encoding (for this use-case) and will likely
> > > > > > > result in an Onode that's only 1/3 of the size that Somnath is
> > > > > > > seeing.
> > > > > > That will be true for the lextent and blob extent maps.  I'm
> > > > > > guessing this is a small part of the ~5K somnath saw.  If his
> > > > > > objects are 4MB then 4KB of it
> > > > > > (80%) is the csum_data vector, which is a flat vector of
> > > > > > u32 values that are presumably not very compressible.
> > > > > I don't think that's what Somnath is seeing (obviously some data here
> > > > > will
> > > sharpen up our speculations). But in his use case, I believe that he has a
> > > separate blob and pextent for each 4K write (since it's been subjected to
> > > random 4K overwrites), that means somewhere in the data structures at
> > > least one address and one length for each of the 4K blocks (and likely
> > > much
> > > more in the lextent and blob maps as you alluded to above). The encoding
> > > of
> > > just this information alone is larger than the checksum data.
> > > > > > We could perhaps break these into a separate key or keyspace..
> > > > > > That'll give rocksdb a bit more computation work to do (for a custom
> > > > > > merge operator, probably, to update just a piece of the value) but
> > > > > > for a 4KB value I'm not sure it's big enough to really help.  Also
> > > > > > we'd lose locality, would need a second get to load csum metadata on
> > > read, etc.
> > > > > > :/  I don't really have any good ideas here.
> > > > > > 
> > > > > > sage
> > > > > > 
> > > > > > 
> > > > > > > Allen Samuels
> > > > > > > SanDisk |a Western Digital brand
> > > > > > > 2880 Junction Avenue, Milpitas, CA 95134
> > > > > > > T: +1 408 801 7030| M: +1 408 780 6416 allen.samuels@xxxxxxxxxxx
> > > > > > > 
> > > > > > > 
> > > > > > > > -----Original Message-----
> > > > > > > > From: Sage Weil [mailto:sweil@xxxxxxxxxx]
> > > > > > > > Sent: Friday, June 10, 2016 2:35 AM
> > > > > > > > To: Somnath Roy <Somnath.Roy@xxxxxxxxxxx>
> > > > > > > > Cc: Mark Nelson <mnelson@xxxxxxxxxx>; Allen Samuels
> > > > > > > > <Allen.Samuels@xxxxxxxxxxx>; Manavalan Krishnan
> > > > > > > > <Manavalan.Krishnan@xxxxxxxxxxx>; Ceph Development <ceph-
> > > > > > > > devel@xxxxxxxxxxxxxxx>
> > > > > > > > Subject: RE: RocksDB tuning
> > > > > > > > 
> > > > > > > > On Fri, 10 Jun 2016, Somnath Roy wrote:
> > > > > > > > > Sage/Mark,
> > > > > > > > > I debugged the code and it seems there is no WAL write going
> > > > > > > > > on
> > > > > > > > > and
> > > > > > > > working as expected. But, in the process, I found that onode
> > > > > > > > size
> > > > > > > > it is
> > > > > > writing
> > > > > > > > to my environment ~7K !! See this debug print.
> > > > > > > > > 2016-06-09 15:49:24.710149 7f7732fe3700 20
> > > > > > > > bluestore(/var/lib/ceph/osd/ceph-0)   onode
> > > > > > > > #1:7d3c6423:::rbd_data.10186b8b4567.0000000000070cd4:head# is
> > > 7518
> > > > > > > > > This explains why so much data going to rocksdb I guess. Once
> > > > > > > > > compaction kicks in iops I am getting is *30 times* slower.
> > > > > > > > > 
> > > > > > > > > I have 15 osds on 8TB drives and I have created 4TB rbd image
> > > > > > > > > preconditioned with 1M. I was running 4K RW test.
> > > > > > > > The onode is big because of the csum metdata.  Try setting
> > > > > > > > 'bluestore
> > > > > > csum
> > > > > > > > type = none' and see if that is the entire reason or if
> > > > > > > > something
> > > > > > > > else is
> > > > > > going
> > > > > > > > on.
> > > > > > > > 
> > > > > > > > We may need to reconsider the way this is stored.
> > > > > > > > 
> > > > > > > > s
> > > > > > > > 
> > > > > > > > 
> > > > > > > > 
> > > > > > > > 
> > > > > > > > > Thanks & Regards
> > > > > > > > > Somnath
> > > > > > > > > 
> > > > > > > > > -----Original Message-----
> > > > > > > > > From: ceph-devel-owner@xxxxxxxxxxxxxxx
> > > > > > > > > [mailto:ceph-devel-owner@xxxxxxxxxxxxxxx] On Behalf Of Somnath
> > > > > > Roy
> > > > > > > > > Sent: Thursday, June 09, 2016 8:23 AM
> > > > > > > > > To: Mark Nelson; Allen Samuels; Manavalan Krishnan; Ceph
> > > > > > Development
> > > > > > > > > Subject: RE: RocksDB tuning
> > > > > > > > > 
> > > > > > > > > Mark,
> > > > > > > > > As we discussed, it seems there is ~5X write amp on the system
> > > > > > > > > with 4K
> > > > > > > > RW. Considering the amount of data going into rocksdb (and thus
> > > > > > > > kicking
> > > > > > of
> > > > > > > > compaction so fast and degrading performance drastically) , it
> > > > > > > > seems it is
> > > > > > still
> > > > > > > > writing WAL (?)..I used the following rocksdb option for faster
> > > > > > background
> > > > > > > > compaction as well hoping it can keep up with upcoming writes
> > > > > > > > and
> > > > > > writes
> > > > > > > > won't be stalling. But, eventually, after a min or so, it is
> > > > > > > > stalling io..
> > > > > > > > > bluestore_rocksdb_options =
> > > "compression=kNoCompression,max_write_buffer_number=16,min_write_
> > > buffer_number_to_merge=3,recycle_log_file_num=16,compaction_style=k
> > > > > > CompactionStyleLevel,write_buffer_size=67108864,target_file_size_bas
> > > > > > e=6
> > > > > > 
> > > 7108864,max_background_compactions=31,level0_file_num_compaction_tri
> > > gger=8,level0_slowdown_writes_trigger=32,level0_stop_writes_trigger=
> > > > > > 64,
> > > > > > 
> > > num_levels=4,max_bytes_for_level_base=536870912,max_bytes_for_level
> > > > > > > > _multiplier=8,compaction_threads=32,flusher_threads=8"
> > > > > > > > > I will try to debug what is going on there..
> > > > > > > > > 
> > > > > > > > > Thanks & Regards
> > > > > > > > > Somnath
> > > > > > > > > 
> > > > > > > > > -----Original Message-----
> > > > > > > > > From: ceph-devel-owner@xxxxxxxxxxxxxxx
> > > > > > > > > [mailto:ceph-devel-owner@xxxxxxxxxxxxxxx] On Behalf Of Mark
> > > > > > > > > Nelson
> > > > > > > > > Sent: Thursday, June 09, 2016 6:46 AM
> > > > > > > > > To: Allen Samuels; Manavalan Krishnan; Ceph Development
> > > > > > > > > Subject: Re: RocksDB tuning
> > > > > > > > > 
> > > > > > > > > On 06/09/2016 08:37 AM, Mark Nelson wrote:
> > > > > > > > > > Hi Allen,
> > > > > > > > > > 
> > > > > > > > > > On a somewhat related note, I wanted to mention that I had
> > > > > > forgotten
> > > > > > > > > > that chhabaremesh's min_alloc_size commit for different
> > > > > > > > > > media
> > > > > > > > > > types was committed into master:
> > > > > > > > > > 
> > > > > > > > > > 
> > > https://github.com/ceph/ceph/commit/8185f2d356911274ca679614611dc335
> > > > > > > > > > e3
> > > > > > > > > > efd187
> > > > > > > > > > 
> > > > > > > > > > 
> > > > > > > > > > IE those tests appear to already have been using a 4K min
> > > > > > > > > > alloc
> > > > > > > > > > size due to non-rotational NVMe media.  I went back and
> > > > > > > > > > verified
> > > > > > > > > > that explicitly changing the min_alloc size (in fact all of
> > > > > > > > > > them
> > > > > > > > > > to be
> > > > > > > > > > sure) to 4k does not change the behavior from graphs I
> > > > > > > > > > showed
> > > > > > > > > > yesterday.  The rocksdb compaction stalls due to excessive
> > > > > > > > > > reads
> > > > > > > > > > appear (at least on the
> > > > > > > > > > surface) to be due to metadata traffic during heavy small
> > > > > > > > > > random
> > > > > > writes.
> > > > > > > > > Sorry, this was worded poorly.  Traffic due to compaction of
> > > > > > > > > metadata
> > > > > > (ie
> > > > > > > > not leaked WAL data) during small random writes.
> > > > > > > > > Mark
> > > > > > > > > 
> > > > > > > > > > Mark
> > > > > > > > > > 
> > > > > > > > > > On 06/08/2016 06:52 PM, Allen Samuels wrote:
> > > > > > > > > > > Let's make a patch that creates actual Ceph parameters for
> > > > > > > > > > > these things so that we don't have to edit the source code
> > > > > > > > > > > in the
> > > future.
> > > > > > > > > > > 
> > > > > > > > > > > Allen Samuels
> > > > > > > > > > > SanDisk |a Western Digital brand
> > > > > > > > > > > 2880 Junction Avenue, San Jose, CA 95134
> > > > > > > > > > > T: +1 408 801 7030| M: +1 408 780 6416
> > > > > > > > > > > allen.samuels@xxxxxxxxxxx
> > > > > > > > > > > 
> > > > > > > > > > > 
> > > > > > > > > > > > -----Original Message-----
> > > > > > > > > > > > From: ceph-devel-owner@xxxxxxxxxxxxxxx
> > > > > > > > > > > > [mailto:ceph-devel-
> > > > > > > > > > > > owner@xxxxxxxxxxxxxxx] On Behalf Of Manavalan Krishnan
> > > > > > > > > > > > Sent: Wednesday, June 08, 2016 3:10 PM
> > > > > > > > > > > > To: Mark Nelson <mnelson@xxxxxxxxxx>; Ceph Development
> > > > > > <ceph-
> > > > > > > > > > > > devel@xxxxxxxxxxxxxxx>
> > > > > > > > > > > > Subject: RocksDB tuning
> > > > > > > > > > > > 
> > > > > > > > > > > > Hi Mark
> > > > > > > > > > > > 
> > > > > > > > > > > > Here are the tunings that we used to avoid the IOPs
> > > > > > > > > > > > choppiness
> > > > > > > > > > > > caused by rocksdb compaction.
> > > > > > > > > > > > 
> > > > > > > > > > > > We need to add the following options in
> > > > > > > > > > > > src/kv/RocksDBStore.cc
> > > > > > > > > > > > before rocksdb::DB::Open in RocksDBStore::do_open
> > > > > > > > opt.IncreaseParallelism(16);
> > > > > > > > > > > >     opt.OptimizeLevelStyleCompaction(512 * 1024 * 1024);
> > > > > > > > > > > > 
> > > > > > > > > > > > 
> > > > > > > > > > > > 
> > > > > > > > > > > > Thanks
> > > > > > > > > > > > Mana
> > > > > > > > > > > > 
> > > > > > > > > > > > 
> > > > > > > > > > > > PLEASE NOTE: The information contained in this
> > > > > > > > > > > > electronic mail
> > > > > > > > > > > > message is intended only for the use of the designated
> > > > > > > > > > > > recipient(s) named above.
> > > > > > > > > > > > If the
> > > > > > > > > > > > reader of this message is not the intended recipient,
> > > > > > > > > > > > you are
> > > > > > > > > > > > hereby notified that you have received this message in
> > > > > > > > > > > > error
> > > > > > > > > > > > and that any review, dissemination, distribution, or
> > > > > > > > > > > > copying
> > > > > > > > > > > > of this message is strictly prohibited. If you have
> > > > > > > > > > > > received
> > > > > > > > > > > > this communication in error, please notify the sender by
> > > > > > > > > > > > telephone or e-mail (as shown
> > > > > > > > > > > > above) immediately and destroy any and all copies of
> > > > > > > > > > > > this
> > > > > > > > > > > > message in your possession (whether hard copies or
> > > > > > > > > > > > electronically stored copies).
> > > > > > > > > > > > --
> > > > > > > > > > > > To unsubscribe from this list: send the line
> > > > > > > > > > > > "unsubscribe
> > > > > > > > > > > > ceph-
> > > > > > devel"
> > > > > > > > > > > > in the
> > > > > > > > > > > > body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > majordomo
> > > > > > > > info
> > > > > > > > > > > > at http://vger.kernel.org/majordomo-info.html
> > > > > > > > > > > --
> > > > > > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > > > > > ceph-
> > > devel"
> > > > > > > > > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > > > > > > majordomo info at
> > > > > > > > > > > http://vger.kernel.org/majordomo-info.html
> > > > > > > > > > > 
> > > > > > > > > > --
> > > > > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > > > > ceph-
> > > devel"
> > > > > > > > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > > > majordomo
> > > > > > > > > > info at  http://vger.kernel.org/majordomo-info.html
> > > > > > > > > --
> > > > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > > > ceph-devel"
> > > > > > > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > > > majordomo
> > > > > > > > > info at  http://vger.kernel.org/majordomo-info.html
> > > > > > > > > PLEASE NOTE: The information contained in this electronic mail
> > > > > > > > > message
> > > > > > is
> > > > > > > > intended only for the use of the designated recipient(s) named
> > > > > > > > above. If
> > > > > > the
> > > > > > > > reader of this message is not the intended recipient, you are
> > > > > > > > hereby
> > > > > > notified
> > > > > > > > that you have received this message in error and that any
> > > > > > > > review,
> > > > > > > > dissemination, distribution, or copying of this message is
> > > > > > > > strictly
> > > > > > prohibited. If
> > > > > > > > you have received this communication in error, please notify the
> > > > > > > > sender
> > > > > > by
> > > > > > > > telephone or e-mail (as shown above) immediately and destroy any
> > > > > > > > and
> > > > > > all
> > > > > > > > copies of this message in your possession (whether hard copies
> > > > > > > > or
> > > > > > > > electronically stored copies).
> > > > > > > > > --
> > > > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > > > ceph-devel"
> > > > > > > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > > > majordomo
> > > > > > > > > info at  http://vger.kernel.org/majordomo-info.html
> > > > > > > > > --
> > > > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > > > ceph-devel"
> > > > > > > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > > > > > > majordomo
> > > > > > > > > info at  http://vger.kernel.org/majordomo-info.html
> > > > > > > > > 
> > > > > > > > > 
> > > > > > > --
> > > > > > > To unsubscribe from this list: send the line "unsubscribe
> > > > > > > ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx
> > > > > > > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> > > > > > > 
> > > > > > > 
> > > > > PLEASE NOTE: The information contained in this electronic mail message
> > > > > is
> > > intended only for the use of the designated recipient(s) named above. If
> > > the
> > > reader of this message is not the intended recipient, you are hereby
> > > notified
> > > that you have received this message in error and that any review,
> > > dissemination, distribution, or copying of this message is strictly
> > > prohibited. If
> > > you have received this communication in error, please notify the sender by
> > > telephone or e-mail (as shown above) immediately and destroy any and all
> > > copies of this message in your possession (whether hard copies or
> > > electronically stored copies).
> > > > > 
> > > > --
> > > > To unsubscribe from this list: send the line "unsubscribe ceph-devel"
> > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More
> > > majordomo
> > > > info at  http://vger.kernel.org/majordomo-info.html
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux