Re: List of SSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We replaced 32 S3500s with 48 Micron M600s in our production cluster. The S3500s were only doing journals because they were too small and we still ate 3-4% of their life in a couple of months. We started having high wait times on the M600s so we got 6 S3610s, 6 M500dcs, and 6 500 GB M600s (they have the SLC to MLC conversion that we thought might work better). And we swapped out 18 of the M600s throughout our cluster with these test drives. We have graphite gathering stats on the admin sockets for Ceph and the standard system stats. We weighted the drives so they had the same byte usage and let them run for a week or so, then made them the same percentage of used space, let them run a couple of weeks, then set them to 80% full and let them run a couple of weeks. We compared IOPS and IO time of the drives to get our comparison. This was done on live production clusters and not synthetic benchmarks. Some of the data about the S3500s is from my test cluster that has them.

Sent from a mobile device, please excuse any typos.

On Feb 25, 2016 9:20 PM, "Christian Balzer" <chibi@xxxxxxx> wrote:

Hello,

On Wed, 24 Feb 2016 22:56:15 -0700 Robert LeBlanc wrote:

> We are moving to the Intel S3610, from our testing it is a good balance
> between price, performance and longevity. But as with all things, do your
> testing ahead of time. This will be our third model of SSDs for our
> cluster. The S3500s didn't have enough life and performance tapers off
> add it gets full. The Micron M600s looked good with the Sebastian journal
> tests, but once in use for a while go downhill pretty bad. We also tested
> Micron M500dc drives and they were on par with the S3610s and are more
> expensive and are closer to EoL. The S3700s didn't have quite the same
> performance as the S3610s, but they will last forever and are very stable
> in terms of performance and have the best power loss protection.
>
That's interesting, how did you come to that conclusion and how did test
it?
Also which models did you compare?


> Short answer is test them for yourself to make sure they will work. You
> are pretty safe with the Intel S3xxx drives. The Micron M500dc is also
> pretty safe based on my experience. It had also been mentioned that
> someone has had good experience with a Samsung DC Pro (has to have both
> DC and Pro in the name), but we weren't able to get any quick enough to
> test so I can't vouch for them.
>
I have some Samsung DC Pro EVOs in production (non-Ceph, see that
non-barrier thread).
They do have issues with LSI occasionally, haven't gotten around to make
that FS non-barrier to see if it fixes things.

The EVOs are also similar to the Intel DC S3500s, meaning that they are
not really suitable for Ceph due to their endurance.

Never tested the "real" DC Pro ones, but they are likely to be OK.

Christian

> Sent from a mobile device, please excuse any typos.
> On Feb 24, 2016 6:37 PM, "Shinobu Kinjo" <skinjo@xxxxxxxxxx> wrote:
>
> > Hello,
> >
> > There has been a bunch of discussion about using SSD.
> > Does anyone have any list of SSDs describing which SSD is highly
> > recommended, which SSD is not.
> >
> > Rgds,
> > Shinobu
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >


--
Christian Balzer        Network/Systems Engineer
chibi@xxxxxxx           Global OnLine Japan/Rakuten Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux