Antw: Re: SSD Journal

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




>>> Christian Balzer <chibi@xxxxxxx> schrieb am Donnerstag, 14. Juli
2016 um
17:06:

> Hello,
> 
> On Thu, 14 Jul 2016 13:37:54 +0200 Steffen Weißgerber wrote:
> 
>> 
>> 
>> >>> Christian Balzer <chibi@xxxxxxx> schrieb am Donnerstag, 14. Juli
2016 um
>> 05:05:
>> 
>> Hello,
>> 
>> > Hello,
>> > 
>> > On Wed, 13 Jul 2016 09:34:35 +0000 Ashley Merrick wrote:
>> > 
>> >> Hello,
>> >> 
>> >> Looking at using 2 x 960GB SSD's (SM863)
>> >>
>> > Massive overkill.
>> >  
>> >> Reason for larger is I was thinking would be better off with them
in Raid 1 
> 
>> > so enough space for OS and all Journals.
>> >>
>> > As I pointed out several times in this ML, Ceph journal usage
rarely
>> > exceeds hundreds of MB, let alone several GB with default
parameters.
>> > So 10GB per journal is plenty, unless you're doing something very
special
>> > (and you aren't with normal HDDs as OSDs).
>> >  
>> >> Instead am I better off using 2 x 200GB S3700's instead, with 5
disks per a 
> 
>> > SSD?
>> >>
>> > S3700s are unfortunately EOL'ed, the 200GB ones were great at
375MB/s.
>> > 200GB S3710s are about on par for 5 HDDs at 300MB/s, but if you
can afford
>> > it and have a 10Gb/s network, the 400GB ones at 470MB/s would be
optimal.
>> > 
>> > As for sharing the SSDs with OS, I do that all the time, the
minute
>> > logging of a storage node really has next to no impact.
>> > 
>> > I prefer this over using DoMs for reasons of:
>> > 1. Redundancy
>> > 2. hot-swapability  
>> > 
>> > If you go the DoM route, make sure it's size AND endurance are a
match for
>> > what you need. 
>> > This is especially important if you were to run a MON on those
machines as
>> > well.
>> > 
>> 
>> Cause we had to change some DoM's due to heavy MON logging, how do
you
>> configure MON logging? On that redundant SSD's or remote?
>>  
> 
> What model/maker DoM where those?

DeLOCK 54655 16GB

> 
> Anyway, everything that runs a MON in my clusters has SSDs with
sufficient
> endurance for the OS.
> Heck, even 180GB Intel 530s (aka consumer SSD) doing the OS for a
> dedicated MON in a busy (but standard level logging) cluster are only
at
> 98% wear-out (as in 2% down) after a year.
> Though that's a HW RAID1 and the controller has 512MB cache, so
writes do
> get nicely coalesced. 
> All my other MONs (shared with on OSD storage nodes) are on S37x0
SSDs.
> 
> OTOH, the Supermicro DoMs look nice enough on paper with 1 DWPD:
> https://www.supermicro.com/products/nfo/SATADOM.cfm 
> 
> The 64GB model should do the trick in most scenarios.
> 

Thank you for the suggestion.

We sometimes run in massive logging caused by issue

http://tracker.ceph.com/issues/4282

what may lead to faster storage aging. Therefore we switched to remote
log.


Steffen

> Christian
> 
>> Steffen
>> 
>> > Christian
>> > 
>> >> Thanks,
>> >> Ashley
>> >> 
>> >> -----Original Message-----
>> >> From: Christian Balzer [mailto:chibi@xxxxxxx] 
>> >> Sent: 13 July 2016 01:12
>> >> To: ceph-users@xxxxxxxxxxxxxx 
>> >> Cc: Wido den Hollander <wido@xxxxxxxx>; Ashley Merrick
<ashley@xxxxxxxxxxxxxx>
>> >> Subject: Re:  SSD Journal
>> >> 
>> >> 
>> >> Hello,
>> >> 
>> >> On Tue, 12 Jul 2016 19:14:14 +0200 (CEST) Wido den Hollander
wrote:
>> >> 
>> >> > 
>> >> > > Op 12 juli 2016 om 15:31 schreef Ashley Merrick
<ashley@xxxxxxxxxxxxxx>:
>> >> > > 
>> >> > > 
>> >> > > Hello,
>> >> > > 
>> >> > > Looking at final stages of planning / setup for a CEPH
Cluster.
>> >> > > 
>> >> > > Per a Storage node looking @
>> >> > > 
>> >> > > 2 x SSD OS / Journal
>> >> > > 10 x SATA Disk
>> >> > > 
>> >> > > Will have a small Raid 1 Partition for the OS, however not
sure if best 
> to 
>> > do:
>> >> > > 
>> >> > > 5 x Journal Per a SSD
>> >> > 
>> >> > Best solution. Will give you the most performance for the OSDs.
RAID-1 
> will 
>> > just burn through cycles on the SSDs.
>> >> > 
>> >> > SSDs do
n't fail that often.
>> >> >
>> >> What Wido wrote, but let us know what SSDs you're planning to
use.
>> >> 
>> >> Because the detailed version of that sentence should read: 
>> >> "Well known and tested DC level SSDs whose size/endurance levels
are 
> matched 
>> > to the workload rarely fail, especially unexpected."
>> >>  
>> >> > Wido
>> >> > 
>> >> > > 10 x Journal on Raid 1 of two SSD's
>> >> > > 
>> >> > > Is the "Performance" increase from splitting 5 Journal's on
each SSD 
> worth 
>> > the "issue" caused when one SSD goes down?
>> >> > > 
>> >> As always, assume at least a node being the failure domain you
need to be 
>> > able to handle.
>> >> 
>> >> Christian
>> >> 
>> >> > > Thanks,
>> >> > > Ashley
>> >> > > _______________________________________________
>> >> > > ceph-users mailing list
>> >> > > ceph-users@xxxxxxxxxxxxxx 
>> >> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>> >> > _______________________________________________
>> >> > ceph-users mailing list
>> >> > ceph-users@xxxxxxxxxxxxxx 
>> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>> >> > 
>> >> 
>> >> 
>> > 
>> > 
>> > -- 
>> > Christian Balzer        Network/Systems Engineer                
>> > chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
>> > http://www.gol.com/ 
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx 
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>> 
>> 
> 
> 
> -- 
> Christian Balzer        Network/Systems Engineer                
> chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
> http://www.gol.com/

-- 
Klinik-Service Neubrandenburg GmbH
Allendestr. 30, 17036 Neubrandenburg
Amtsgericht Neubrandenburg, HRB 2457
Geschaeftsfuehrerin: Gudrun Kappich
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux