Re: multiple journals on SSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Just to add if you really want to go with lots of HDD's to Journals then go
NVME. They are not a lot more expensive than the equivalent SATA based
3700's, but the latency is low low low. Here is an example of a node I have
just commissioned with 12 HDD's to one P3700

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz
avgqu-sz   await r_await w_await  svctm  %util
sdb               0.00     0.00   68.00    0.00  8210.00     0.00   241.47
0.26    3.85    3.85    0.00   2.09  14.20
sdd               2.50     0.00  198.50   22.00 24938.00  9422.00   311.66
4.34   27.80    6.21  222.64   2.45  54.00
sdc               0.00     0.00   63.00    0.00  7760.00     0.00   246.35
0.15    2.16    2.16    0.00   1.56   9.80
sda               0.00     0.00   61.50   47.00  7600.00 22424.00   553.44
2.77   25.57    2.63   55.57   3.82  41.40
nvme0n1           0.00    22.50    2.00 2605.00     8.00 139638.00   107.13
0.14    0.05    0.00    0.05   0.03   6.60
sdg               0.00     0.00   61.00   28.00  6230.00 12696.00   425.30
3.66   74.79    5.84  225.00   3.87  34.40
sdf               0.00     0.00   34.50   47.00  4108.00 21702.00   633.37
3.56   43.75    1.51   74.77   2.85  23.20
sdh               0.00     0.00   75.00   15.50  9180.00  4984.00   313.02
0.45   12.55    3.28   57.42   3.51  31.80
sdi               1.50     0.50  142.00   48.50 18102.00 21924.00   420.22
3.60   18.92    4.99   59.71   2.70  51.40
sdj               0.50     0.00   74.50    5.00  9362.00  1832.00   281.61
0.33    4.10    3.33   15.60   2.44  19.40
sdk               0.00     0.00   54.00    0.00  6420.00     0.00   237.78
0.12    2.30    2.30    0.00   1.70   9.20
sdl               0.00     0.00   21.00    1.50  2286.00    16.00   204.62
0.32   18.13   13.81   78.67   6.67  15.00
sde               0.00     0.00   98.00    0.00 12304.00     0.00   251.10
0.30    3.10    3.10    0.00   2.08  20.40

50us latency at 2605 iops!!!

Compared to one of the other nodes with 2 100GB S3700's, 6 disks each

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz
avgqu-sz   await r_await w_await  svctm  %util
sda               0.00    30.50    0.00  894.50     0.00 50082.00   111.98
0.36    0.41    0.00    0.41   0.20  17.80
sdb               0.00     9.00    0.00  551.00     0.00 32044.00   116.31
0.23    0.42    0.00    0.42   0.19  10.40
sdc               0.00     2.00    6.50   17.50   278.00  8422.00   725.00
1.08   44.92   18.46   54.74   8.08  19.40
sdd               0.00     0.00    0.00    0.00     0.00     0.00     0.00
0.00    0.00    0.00    0.00   0.00   0.00
sde               0.00     2.50   27.50   21.50  2112.00  9866.00   488.90
0.59   12.04    6.91   18.60   6.53  32.00
sdf               0.50     0.00   50.50    0.00  6170.00     0.00   244.36
0.18    4.63    4.63    0.00   2.10  10.60
md1               0.00     0.00    0.00    0.00     0.00     0.00     0.00
0.00    0.00    0.00    0.00   0.00   0.00
md0               0.00     0.00    0.00    0.00     0.00     0.00     0.00
0.00    0.00    0.00    0.00   0.00   0.00
sdg               0.00     1.50   32.00  386.50  3970.00 12188.00    77.22
0.15    0.35    0.50    0.34   0.15   6.40
sdh               0.00     0.00    6.00    0.00    34.00     0.00    11.33
0.07   12.67   12.67    0.00  11.00   6.60
sdi               0.00     0.50    1.50   19.50     6.00  8862.00   844.57
0.96   45.71   33.33   46.67   6.57  13.80
sdj               0.00     0.00   67.00    0.00  8214.00     0.00   245.19
0.17    2.51    2.51    0.00   1.88  12.60
sdk               1.50     2.50   61.00   48.00  6216.00 21020.00   499.74
2.01   18.46   11.41   27.42   5.05  55.00
sdm               0.00     0.00   30.50    0.00  3576.00     0.00   234.49
0.07    2.43    2.43    0.00   1.90   5.80
sdl               0.00     4.50   25.00   23.50  2092.00 12648.00   607.84
1.36   19.42    5.60   34.13   4.04  19.60
sdn               0.50     0.00   23.00    0.00  2670.00     0.00   232.17
0.07    2.96    2.96    0.00   2.43   5.60

Pretty much 10x the latency. I'm seriously impressed with these NVME things.


> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of
> Christian Balzer
> Sent: 07 July 2016 03:23
> To: ceph-users@xxxxxxxxxxxxxx
> Subject: Re:  multiple journals on SSD
> 
> 
> Hello,
> 
> I have a multitude of of problems with the benchmarks and conclusions
here,
> more below.
> 
> But firstly to address the question of the OP, definitely not filesystem
based
> journals.
> Another layer of overhead and delays, something I'd be willing to ignore
if
> we're talking about a full SSD as OSD with an inline journal, but not with
> journal SSDs.
> Similar with LVM, though with a lower impact.
> 
> Partitions really are your best bet.
> 
> On Wed, 6 Jul 2016 18:20:43 +0300 George Shuklin wrote:
> 
> > Yes.
> >
> > On my lab (not production yet) with 9 7200 SATA (OSD) and one INTEL
> > SSDSC2BB800G4 (800G, 9 journals)
> 
> First and foremost, a DC 3510 with 1 DWPD endurance is not my idea of good
> journal device, even if it had the performance.
> If you search in the ML archives there is at least one case where somebody
> lost a full storage node precisely because their DC S3500s were worn out:
> https://www.mail-archive.com/ceph-users@xxxxxxxxxxxxxx/msg28083.html
> 
> Unless you have a read-mostly cluster, a 400GB DC S3610 (same or lower
> price) would be a better deal, at 50% more endurance and only slightly
lower
> sequential write speed.
> 
> And depending on your expected write volume (which you should
> know/estimate as close as possible before buying HW), a 400GB DC S3710
> might be the best deal when it comes to TBW/$.
> It's 30% more expensive than your 3510, but has the same speed and an
> endurance that's 5 times greater.
> 
> > during random write I got ~90%
> > utilization of 9 HDD with ~5% utilization of SSD (2.4k IOPS). With
> > linear writing it somehow worse: I got 250Mb/s on SSD, which
> > translated to 240Mb of all OSD combined.
> >
> This test shows us a lot of things, mostly the failings of filestore.
> But only partially if a SSD is a good fit for journals or not.
> 
> How are you measuring these things on the storage node, iostat, atop?
> At 250MB/s (Mb would be mega-bit) your 800 GB DC S3500 should register
> about/over 50% utilization, given that its top speed is 460MB/s.
> 
> With Intel DC SSDs you can pretty much take the sequential write speed
from
> their specifications page and roughly expect that to be the speed of your
> journal.
> 
> For example a 100GB DC S3700 (200MB/s) doing journaling for 2 plain SATA
> HDDs will give us this when running "ceph tell osd.nn bench" in parallel
> against 2 OSDs that share a journal SSD:
> ---
> Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz
avgqu-sz
> await r_await w_await  svctm  %util
> sdd               0.00     2.00    0.00  409.50     0.00 191370.75
934.66   146.52  356.46
> 0.00  356.46   2.44 100.00
> sdl               0.00    85.50    0.50  120.50     2.00 49614.00   820.10
2.25   18.51
> 0.00   18.59   8.20  99.20
> sdk               0.00    89.50    1.50  119.00     6.00 49348.00   819.15
2.04   16.91
> 0.00   17.13   8.23  99.20
> ---
> 
> Where sdd is the journal SSD and sdl/sdk are the OSD HDDs.
> And the SSD is nearly at 200MB/s (and 100%).
> For the record, that bench command is good for testing, but the result:
> ---
> # ceph tell osd.30 bench
> {
>     "bytes_written": 1073741824,
>     "blocksize": 4194304,
>     "bytes_per_sec": 100960114.000000
> }
> ---
> should be taken with a grain of salt, realistically those OSDs can do
about
> 50MB/s sustained.
> 
> On another cluster I have 200GB DC S3700s (360MB/s), holding 3 journals
for
> 4 disk RAID10 (4GB HW cache Areca controller) OSDs.
> Thus the results are more impressive:
> ---
> Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz
avgqu-sz
> await r_await w_await  svctm  %util
> sda               0.00   381.00    0.00  485.00     0.00 200374.00
826.28     3.16    6.49
> 0.00    6.49   1.53  74.20
> sdb               0.00   350.50    1.00  429.00     4.00 177692.00
826.49     2.78    6.46
> 4.00    6.46   1.53  65.60
> sdg               0.00     1.00    0.00  795.00     0.00 375514.50
944.69   143.68  180.43
> 0.00  180.43   1.26 100.00
> ---
> 
> Where sda/sdb are the OSD RAIDs and sdg is the journal SSD.
> Again, a near perfect match to the Intel specifications and also an
example
> where the journal is the bottleneck (never mind that his cluster is all
about
> IOPS, not throughput).
> 
> As for the endurance mentioned above, these 200GB DC 3700s are/were
> overkill:
> ---
> 233 Media_Wearout_Indicator 0x0032   098   098   000    Old_age   Always
-
> 0
> 241 Host_Writes_32MiB       0x0032   100   100   000    Old_age   Always
-
> 4818100
> 242 Host_Reads_32MiB        0x0032   100   100   000    Old_age   Always
-
> 84403
> ---
> 
> Again, this cluster is all about (small) IOPS, it only sees about 5MB/s
sustained
> I/O.
> So a 3610 might been a better fit, but not only didn't they exist back
then, it
> would have to be the 400GB model to match the speed, which is more
> expensive.
> A DC S3510 would be down 20% in terms of wearout (assuming same size)
> and of course significantly slower.
> With a 480GB 3510 (similar speed) it would still be about 10% worn out and
> thus still no match for the expected life time of this cluster.
> 
> The numbers above do correlate nicely with dd or fio tests (with 4MB
> blocks) from VMs against the same clusters.
> 
> > Obviously, it sucked with cold randread too (as expected).
> >
> Reads never touch the journal SSDs.
> 
> > Just for comparacment, my baseline benchmark (fio/librbd, 4k,
> > iodepth=32, randwrite) for single OSD in the pool with size=1:
> >
> > Intel 53x and Pro 2500 Series SSDs - 600 IOPS Intel 730
> Consumer models, avoid.
> 
> > and DC S35x0/3610/3700 Series SSDs - 6605 IOPS
> Again, what you're comparing here is only part of the picture.
> With tests as shown above you'd see significant differences.
> 
> > Samsung SSD 840 Series - 739 IOPS
> Also consumer model, with impressive and unpredicted deaths reported.
> 
> Christian
> 
> > EDGE Boost Pro Plus 7mm - 1000 IOPS
> >
> > (so 3500 is clear winner)
> >
> > On 07/06/2016 03:22 PM, Alwin Antreich wrote:
> > > Hi George,
> > >
> > > interesting result for your benchmark. May you please supply some
> > > more numbers? As we didn't get that good of a result on our tests.
> > >
> > > Thanks.
> > >
> > > Cheers,
> > > Alwin
> > >
> > >
> > > On 07/06/2016 02:03 PM, George Shuklin wrote:
> > >> Hello.
> > >>
> > >> I've been testing Intel 3500 as journal store for few HDD-based OSD.
> > >> I stumble on issues with multiple partitions (>4) and UDEV (sda5,
> > >> sda6,etc sometime do not appear after partition creation). And I'm
> > >> thinking that partition is not that useful for OSD management,
> > >> because linux do no allow partition rereading with it contains used
> > >> volumes.
> > >>
> > >> So my question: How you store many journals on SSD? My initial
> > >> thoughts:
> > >>
> > >> 1)  filesystem with filebased journals
> > >> 2) LVM with volumes
> > >>
> > >> Anything else? Best practice?
> > >>
> > >> P.S. I've done benchmarking: 3500 can support up to 16 10k-RPM HDD.
> > >> _______________________________________________
> > >> ceph-users mailing list
> > >> ceph-users@xxxxxxxxxxxxxx
> > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@xxxxxxxxxxxxxx
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> 
> 
> --
> Christian Balzer        Network/Systems Engineer
> chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
> http://www.gol.com/
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux