Re: Ceph cluster NO read / write performance :: Ops are blocked

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We have CephFS utilizing a cache tier + EC backend. The cache tier and ec pool sit on the same spinners — no SSDs. Our cache tier has a target_max_bytes of 5TB and the total storage is about 1PB. 

I do have a separate test pool with 3x replication and no cache tier, and I still see significant performance drops and blocked ops with no/minimal client I/O from CephFS. Right now I have 530 blocked ops with 20MB/s of client write I/O and no active scrubs. The rados bench on my test pool looks like this:

  sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
    0       0         0         0         0         0         -         0
    1      31        94        63   251.934       252   0.31017  0.217719
    2      31       103        72   143.969        36  0.978544  0.260631
    3      31       103        72   95.9815         0         -  0.260631
    4      31       111        80   79.9856        16   2.29218  0.476458
    5      31       112        81   64.7886         4    2.5559   0.50213
    6      31       112        81   53.9905         0         -   0.50213
    7      31       115        84   47.9917         6   3.71826  0.615882
    8      31       115        84   41.9928         0         -  0.615882
    9      31       115        84    37.327         0         -  0.615882
   10      31       117        86   34.3942   2.66667   6.73678  0.794532

I’m really leaning more toward it being a weird controller/disk problem. 

As a test, I suppose I could double the target_max_bytes, just so the cache tier stops evicting while client I/O is writing?

—Lincoln

> On Sep 17, 2015, at 11:59 AM, Nick Fisk <nick@xxxxxxxxxx> wrote:
> 
> Ah right....this is where it gets interesting.
> 
> You are probably hitting a cache full on a PG somewhere which is either making everything wait until it flushes or something like that. 
> 
> What cache settings have you got set?
> 
> I assume you have SSD's for the cache tier? Can you share the size of the pool.
> 
> If possible could you also create a non tiered test pool and do some benchmarks on that to rule out any issue with the hardware and OSD's.
> 
>> -----Original Message-----
>> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of
>> Lincoln Bryant
>> Sent: 17 September 2015 17:54
>> To: Nick Fisk <nick@xxxxxxxxxx>
>> Cc: ceph-users@xxxxxxxxxxxxxx
>> Subject: Re:  Ceph cluster NO read / write performance :: Ops
>> are blocked
>> 
>> Hi Nick,
>> 
>> Thanks for responding. Yes, I am.
>> 
>> —Lincoln
>> 
>>> On Sep 17, 2015, at 11:53 AM, Nick Fisk <nick@xxxxxxxxxx> wrote:
>>> 
>>> You are getting a fair amount of reads on the disks whilst doing these
>> writes. You're not using cache tiering are you?
>>> 
>>>> -----Original Message-----
>>>> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf
>>>> Of Lincoln Bryant
>>>> Sent: 17 September 2015 17:42
>>>> To: ceph-users@xxxxxxxxxxxxxx
>>>> Subject: Re:  Ceph cluster NO read / write performance ::
>>>> Ops are blocked
>>>> 
>>>> Hello again,
>>>> 
>>>> Well, I disabled offloads on the NIC -- didn’t work for me. I also
>>>> tried setting net.ipv4.tcp_moderate_rcvbuf = 0 as suggested elsewhere
>>>> in the thread to no avail.
>>>> 
>>>> Today I was watching iostat on an OSD box ('iostat -xm 5') when the
>>>> cluster got into “slow” state:
>>>> 
>>>> Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s avgrq-sz avgqu-
>> sz
>>>> await  svctm  %util
>>>> sdb               0.00    13.57   84.23  167.47     0.45     2.78    26.26     2.06    8.18
>> 3.85
>>>> 96.93
>>>> sdc               0.00    46.71    5.59  289.22     0.03     2.54    17.85     3.18   10.77
>> 0.97
>>>> 28.72
>>>> sdd               0.00    16.57   45.11   91.62     0.25     0.55    12.01     0.75    5.51
>> 2.45
>>>> 33.47
>>>> sde               0.00    13.57    6.99  143.31     0.03     2.53    34.97     1.99   13.27
>> 2.12
>>>> 31.86
>>>> sdf               0.00    18.76    4.99  158.48     0.10     1.09    14.88     1.26    7.69   1.24
>>>> 20.26
>>>> sdg               0.00    25.55   81.64  237.52     0.44     2.89    21.36     4.14   12.99
>> 2.58
>>>> 82.22
>>>> sdh               0.00    89.42   16.17  492.42     0.09     3.81    15.69    17.12   33.66
>> 0.73
>>>> 36.95
>>>> sdi               0.00    20.16   17.76  189.62     0.10     1.67    17.46     3.45   16.63
>> 1.57
>>>> 32.55
>>>> sdj               0.00    31.54    0.00  185.23     0.00     1.91    21.15     3.33   18.00
>> 0.03
>>>> 0.62
>>>> sdk               0.00    26.15    2.40  133.33     0.01     0.84    12.79     1.07    7.87
>> 0.85
>>>> 11.58
>>>> sdl               0.00    25.55    9.38  123.95     0.05     1.15    18.44     0.50    3.74   1.58
>>>> 21.10
>>>> sdm               0.00     6.39   92.61   47.11     0.47     0.26    10.65     1.27    9.07
>> 6.92
>>>> 96.73
>>>> 
>>>> The %util is rather high on some disks, but I’m not an expert at
>>>> looking at iostat so I’m not sure how worrisome this is. Does
>>>> anything here stand out to anyone?
>>>> 
>>>> At the time of that iostat, Ceph was reporting a lot of blocked ops
>>>> on the OSD associated with sde (as well as about 30 other OSDs), but
>>>> it doesn’t look all that busy. Some simple ‘dd’ tests seem to indicate the
>> disk is fine.
>>>> 
>>>> Similarly, iotop seems OK on this host:
>>>> 
>>>> TID  PRIO  USER     DISK READ  DISK WRITE  SWAPIN     IO>    COMMAND
>>>> 472477 be/4 root        0.00 B/s    5.59 M/s  0.00 %  0.57 % ceph-osd -i 111 --
>> pid-
>>>> file /var/run/ceph/osd.111.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 470621 be/4 root        0.00 B/s   10.09 M/s  0.00 %  0.40 % ceph-osd -i 111 --
>> pid-
>>>> file /var/run/ceph/osd.111.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3495447 be/4 root        0.00 B/s  272.19 K/s  0.00 %  0.36 % ceph-osd -i 114 --
>>>> pid-file /var/run/ceph/osd.114.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488389 be/4 root	 0.00 B/s  596.80 K/s  0.00 %  0.16 % ceph-osd -i 109 --
>>>> pid-file /var/run/ceph/osd.109.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488060 be/4 root        0.00 B/s  600.83 K/s  0.00 %  0.15 % ceph-osd -i 108 --
>>>> pid-file /var/run/ceph/osd.108.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3505573 be/4 root        0.00 B/s  528.25 K/s  0.00 %  0.10 % ceph-osd -i 119 --
>>>> pid-file /var/run/ceph/osd.119.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3495434 be/4 root        0.00 B/s    2.02 K/s  0.00 %  0.10 % ceph-osd -i 114 --
>> pid-
>>>> file /var/run/ceph/osd.114.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3502327 be/4 root        0.00 B/s  506.07 K/s  0.00 %  0.09 % ceph-osd -i 118 --
>>>> pid-file /var/run/ceph/osd.118.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3489100 be/4 root        0.00 B/s  106.86 K/s  0.00 %  0.09 % ceph-osd -i 110 --
>>>> pid-file /var/run/ceph/osd.110.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3496631 be/4 root        0.00 B/s  229.85 K/s  0.00 %  0.05 % ceph-osd -i 115 --
>>>> pid-file /var/run/ceph/osd.115.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3505561 be/4 root	 0.00 B/s    2.02 K/s  0.00 %  0.03 % ceph-osd -i 119 --
>>>> pid-file /var/run/ceph/osd.119.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488059 be/4 root        0.00 B/s    2.02 K/s  0.00 %  0.03 % ceph-osd -i 108 --
>> pid-
>>>> file /var/run/ceph/osd.108.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488391 be/4 root       46.37 K/s  431.47 K/s  0.00 %  0.02 % ceph-osd -i 109 -
>> -
>>>> pid-file /var/run/ceph/osd.109.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3500639 be/4 root        0.00 B/s  221.78 K/s  0.00 %  0.02 % ceph-osd -i 117 --
>>>> pid-file /var/run/ceph/osd.117.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488392 be/4 root       34.28 K/s  185.49 K/s  0.00 %  0.02 % ceph-osd -i 109 -
>> -
>>>> pid-file /var/run/ceph/osd.109.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 3488062 be/4 root        4.03 K/s   66.54 K/s  0.00 %  0.02 % ceph-osd -i 108 --
>> pid-
>>>> file /var/run/ceph/osd.108.pid -c /etc/ceph/ceph.conf --cluster ceph
>>>> 
>>>> These are all 6TB seagates in single-disk RAID 0 on a PERC H730 Mini
>>>> controller.
>>>> 
>>>> I did try removing the disk with 20k non-medium errors, but that
>>>> didn’t seem to help.
>>>> 
>>>> Thanks for any insight!
>>>> 
>>>> Cheers,
>>>> Lincoln Bryant
>>>> 
>>>>> On Sep 9, 2015, at 1:09 PM, Lincoln Bryant <lincolnb@xxxxxxxxxxxx>
>> wrote:
>>>>> 
>>>>> Hi Jan,
>>>>> 
>>>>> I’ll take a look at all of those things and report back (hopefully
>>>>> :))
>>>>> 
>>>>> I did try setting all of my OSDs to writethrough instead of
>>>>> writeback on the
>>>> controller, which was significantly more consistent in performance
>>>> (from 1100MB/s down to 300MB/s, but still occasionally dropping to
>>>> 0MB/s). Still plenty of blocked ops.
>>>>> 
>>>>> I was wondering if not-so-nicely failing OSD(s) might be the cause.
>>>>> My
>>>> controller (PERC H730 Mini) seems frustratingly terse with SMART
>>>> information, but at least one disk has a “Non-medium error count” of
>>>> over 20,000..
>>>>> 
>>>>> I’ll try disabling offloads as well.
>>>>> 
>>>>> Thanks much for the suggestions!
>>>>> 
>>>>> Cheers,
>>>>> Lincoln
>>>>> 
>>>>>> On Sep 9, 2015, at 3:59 AM, Jan Schermer <jan@xxxxxxxxxxx> wrote:
>>>>>> 
>>>>>> Just to recapitulate - the nodes are doing "nothing" when it drops to
>> zero?
>>>> Not flushing something to drives (iostat)? Not cleaning pagecache
>>>> (kswapd and similiar)? Not out of any type of memory (slab,
>>>> min_free_kbytes)? Not network link errors, no bad checksums (those are
>> hard to spot, though)?
>>>>>> 
>>>>>> Unless you find something I suggest you try disabling offloads on
>>>>>> the NICs
>>>> and see if the problem goes away.
>>>>>> 
>>>>>> Jan
>>>>>> 
>>>>>>> On 08 Sep 2015, at 18:26, Lincoln Bryant <lincolnb@xxxxxxxxxxxx>
>> wrote:
>>>>>>> 
>>>>>>> For whatever it’s worth, my problem has returned and is very
>>>>>>> similar to
>>>> yours. Still trying to figure out what’s going on over here.
>>>>>>> 
>>>>>>> Performance is nice for a few seconds, then goes to 0. This is a
>>>>>>> similar setup to yours (12 OSDs per box, Scientific Linux 6, Ceph
>>>>>>> 0.94.3, etc)
>>>>>>> 
>>>>>>> 384      16     29520     29504   307.287      1188 0.0492006  0.208259
>>>>>>> 385      16     29813     29797   309.532      1172 0.0469708  0.206731
>>>>>>> 386      16     30105     30089   311.756      1168 0.0375764  0.205189
>>>>>>> 387      16     30401     30385   314.009      1184  0.036142  0.203791
>>>>>>> 388      16     30695     30679   316.231      1176 0.0372316  0.202355
>>>>>>> 389      16     30987     30971    318.42      1168 0.0660476  0.200962
>>>>>>> 390      16     31282     31266   320.628      1180 0.0358611  0.199548
>>>>>>> 391      16     31568     31552   322.734      1144 0.0405166  0.198132
>>>>>>> 392      16     31857     31841   324.859      1156 0.0360826  0.196679
>>>>>>> 393      16     32090     32074   326.404       932 0.0416869   0.19549
>>>>>>> 394      16     32205     32189   326.743       460 0.0251877  0.194896
>>>>>>> 395      16     32302     32286   326.897       388 0.0280574  0.194395
>>>>>>> 396      16     32348     32332   326.537       184 0.0256821  0.194157
>>>>>>> 397      16     32385     32369   326.087       148 0.0254342  0.193965
>>>>>>> 398      16     32424     32408   325.659       156 0.0263006  0.193763
>>>>>>> 399      16     32445     32429   325.054        84 0.0233839  0.193655
>>>>>>> 2015-09-08 11:22:31.940164 min lat: 0.0165045 max lat: 67.6184 avg lat:
>>>> 0.193655
>>>>>>> sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>>>>>>> 400      16     32445     32429   324.241         0         -  0.193655
>>>>>>> 401      16     32445     32429   323.433         0         -  0.193655
>>>>>>> 402      16     32445     32429   322.628         0         -  0.193655
>>>>>>> 403      16     32445     32429   321.828         0         -  0.193655
>>>>>>> 404      16     32445     32429   321.031         0         -  0.193655
>>>>>>> 405      16     32445     32429   320.238         0         -  0.193655
>>>>>>> 406      16     32445     32429    319.45         0         -  0.193655
>>>>>>> 407      16     32445     32429   318.665         0         -  0.193655
>>>>>>> 
>>>>>>> needless to say, very strange.
>>>>>>> 
>>>>>>> —Lincoln
>>>>>>> 
>>>>>>> 
>>>>>>>> On Sep 7, 2015, at 3:35 PM, Vickey Singh
>>>> <vickey.singh22693@xxxxxxxxx> wrote:
>>>>>>>> 
>>>>>>>> Adding ceph-users.
>>>>>>>> 
>>>>>>>> On Mon, Sep 7, 2015 at 11:31 PM, Vickey Singh
>>>> <vickey.singh22693@xxxxxxxxx> wrote:
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Mon, Sep 7, 2015 at 10:04 PM, Udo Lembke
>>>> <ulembke@xxxxxxxxxxxx> wrote:
>>>>>>>> Hi Vickey,
>>>>>>>> Thanks for your time in replying to my problem.
>>>>>>>> 
>>>>>>>> I had the same rados bench output after changing the motherboard
>>>>>>>> of
>>>> the monitor node with the lowest IP...
>>>>>>>> Due to the new mainboard, I assume the hw-clock was wrong during
>>>> startup. Ceph health show no errors, but all VMs aren't able to do IO
>>>> (very high load on the VMs - but no traffic).
>>>>>>>> I stopped the mon, but this don't changed anything. I had to
>>>>>>>> restart all
>>>> other mons to get IO again. After that I started the first mon also
>>>> (with the right time now) and all worked fine again...
>>>>>>>> 
>>>>>>>> Thanks i will try to restart all OSD / MONS and report back , if
>>>>>>>> it solves my problem
>>>>>>>> 
>>>>>>>> Another posibility:
>>>>>>>> Do you use journal on SSDs? Perhaps the SSDs can't write to
>>>>>>>> garbage
>>>> collection?
>>>>>>>> 
>>>>>>>> No i don't have journals on SSD , they are on the same OSD disk.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Udo
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On 07.09.2015 16:36, Vickey Singh wrote:
>>>>>>>>> Dear Experts
>>>>>>>>> 
>>>>>>>>> Can someone please help me , why my cluster is not able write
>> data.
>>>>>>>>> 
>>>>>>>>> See the below output  cur MB/S  is 0  and Avg MB/s is decreasing.
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Ceph Hammer  0.94.2
>>>>>>>>> CentOS 6 (3.10.69-1)
>>>>>>>>> 
>>>>>>>>> The Ceph status says OPS are blocked , i have tried checking ,
>>>>>>>>> what all i know
>>>>>>>>> 
>>>>>>>>> - System resources ( CPU , net, disk , memory )    -- All normal
>>>>>>>>> - 10G network for public and cluster network  -- no saturation
>>>>>>>>> - Add disks are physically healthy
>>>>>>>>> - No messages in /var/log/messages OR dmesg
>>>>>>>>> - Tried restarting OSD which are blocking operation , but no
>>>>>>>>> luck
>>>>>>>>> - Tried writing through RBD  and Rados bench , both are giving
>>>>>>>>> same problemm
>>>>>>>>> 
>>>>>>>>> Please help me to fix this problem.
>>>>>>>>> 
>>>>>>>>> #  rados bench -p rbd 60 write
>>>>>>>>> Maintaining 16 concurrent writes of 4194304 bytes for up to 60
>>>>>>>>> seconds or 0 objects Object prefix:
>> benchmark_data_stor1_1791844
>>>>>>>>> sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>>>>>>>>> 0       0         0         0         0         0         -         0
>>>>>>>>> 1      16       125       109   435.873       436  0.022076 0.0697864
>>>>>>>>> 2      16       139       123   245.948        56  0.246578 0.0674407
>>>>>>>>> 3      16       139       123   163.969         0         - 0.0674407
>>>>>>>>> 4      16       139       123   122.978         0         - 0.0674407
>>>>>>>>> 5      16       139       123    98.383         0         - 0.0674407
>>>>>>>>> 6      16       139       123   81.9865         0         - 0.0674407
>>>>>>>>> 7      16       139       123   70.2747         0         - 0.0674407
>>>>>>>>> 8      16       139       123   61.4903         0         - 0.0674407
>>>>>>>>> 9      16       139       123   54.6582         0         - 0.0674407
>>>>>>>>> 10      16       139       123   49.1924         0         - 0.0674407
>>>>>>>>> 11      16       139       123   44.7201         0         - 0.0674407
>>>>>>>>> 12      16       139       123   40.9934         0         - 0.0674407
>>>>>>>>> 13      16       139       123   37.8401         0         - 0.0674407
>>>>>>>>> 14      16       139       123   35.1373         0         - 0.0674407
>>>>>>>>> 15      16       139       123   32.7949         0         - 0.0674407
>>>>>>>>> 16      16       139       123   30.7451         0         - 0.0674407
>>>>>>>>> 17      16       139       123   28.9364         0         - 0.0674407
>>>>>>>>> 18      16       139       123   27.3289         0         - 0.0674407
>>>>>>>>> 19      16       139       123   25.8905         0         - 0.0674407
>>>>>>>>> 2015-09-07 15:54:52.694071min lat: 0.022076 max lat: 0.46117 avg
>> lat:
>>>> 0.0674407
>>>>>>>>> sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>>>>>>>>> 20      16       139       123    24.596         0         - 0.0674407
>>>>>>>>> 21      16       139       123   23.4247         0         - 0.0674407
>>>>>>>>> 22      16       139       123     22.36         0         - 0.0674407
>>>>>>>>> 23      16       139       123   21.3878         0         - 0.0674407
>>>>>>>>> 24      16       139       123   20.4966         0         - 0.0674407
>>>>>>>>> 25      16       139       123   19.6768         0         - 0.0674407
>>>>>>>>> 26      16       139       123     18.92         0         - 0.0674407
>>>>>>>>> 27      16       139       123   18.2192         0         - 0.0674407
>>>>>>>>> 28      16       139       123   17.5686         0         - 0.0674407
>>>>>>>>> 29      16       139       123   16.9628         0         - 0.0674407
>>>>>>>>> 30      16       139       123   16.3973         0         - 0.0674407
>>>>>>>>> 31      16       139       123   15.8684         0         - 0.0674407
>>>>>>>>> 32      16       139       123   15.3725         0         - 0.0674407
>>>>>>>>> 33      16       139       123   14.9067         0         - 0.0674407
>>>>>>>>> 34      16       139       123   14.4683         0         - 0.0674407
>>>>>>>>> 35      16       139       123   14.0549         0         - 0.0674407
>>>>>>>>> 36      16       139       123   13.6645         0         - 0.0674407
>>>>>>>>> 37      16       139       123   13.2952         0         - 0.0674407
>>>>>>>>> 38      16       139       123   12.9453         0         - 0.0674407
>>>>>>>>> 39      16       139       123   12.6134         0         - 0.0674407
>>>>>>>>> 2015-09-07 15:55:12.697124min lat: 0.022076 max lat: 0.46117 avg
>> lat:
>>>> 0.0674407
>>>>>>>>> sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>>>>>>>>> 40      16       139       123   12.2981         0         - 0.0674407
>>>>>>>>> 41      16       139       123   11.9981         0         - 0.0674407
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> cluster 86edf8b8-b353-49f1-ab0a-a4827a9ea5e8
>>>>>>>>> health HEALTH_WARN
>>>>>>>>>        1 requests are blocked > 32 sec  monmap e3: 3 mons at
>>>>>>>>> {stor0111=10.100.1.111:6789/0,stor0113=10.100.1.113:6789/0,stor0
>>>>>>>>> 11
>>>>>>>>> 5=10.100.1.115:6789/0}
>>>>>>>>>        election epoch 32, quorum 0,1,2
>>>>>>>>> stor0111,stor0113,stor0115  osdmap e19536: 50 osds: 50 up, 50 in
>>>>>>>>>  pgmap v928610: 2752 pgs, 9 pools, 30476 GB data, 4183 kobjects
>>>>>>>>>        91513 GB used, 47642 GB / 135 TB avail
>>>>>>>>>            2752 active+clean
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Tried using RBD
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> # dd if=/dev/zero of=file1 bs=4K count=10000 oflag=direct
>>>>>>>>> 10000+0 records in
>>>>>>>>> 10000+0 records out
>>>>>>>>> 40960000 bytes (41 MB) copied, 24.5529 s, 1.7 MB/s
>>>>>>>>> 
>>>>>>>>> # dd if=/dev/zero of=file1 bs=1M count=100 oflag=direct
>>>>>>>>> 100+0 records in
>>>>>>>>> 100+0 records out
>>>>>>>>> 104857600 bytes (105 MB) copied, 1.05602 s, 9.3 MB/s
>>>>>>>>> 
>>>>>>>>> # dd if=/dev/zero of=file1 bs=1G count=1 oflag=direct
>>>>>>>>> 1+0 records in
>>>>>>>>> 1+0 records out
>>>>>>>>> 1073741824 bytes (1.1 GB) copied, 293.551 s, 3.7 MB/s ]#
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> ceph-users mailing list
>>>>>>>>> 
>>>>>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> ceph-users mailing list
>>>>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> ceph-users mailing list
>>>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>> 
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> 
>>> 
>>> 
>>> 
>> 
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux