Re: SSD OSDs - more Cores or more GHz

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



See this benchmark I did last year

http://www.sys-pro.co.uk/ceph-storage-fast-cpus-ssd-performance/


> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of
> Oliver Dzombic
> Sent: 20 January 2016 13:33
> To: ceph-users@xxxxxxxx
> Subject: Re:  SSD OSDs - more Cores or more GHz
> 
> Hi,
> 
> to be honest, i never made real benchmarks about that.
> 
> But to me, i doubt that the higher frequency of a cpu will have a "real"
> impact on ceph's performance.
> 
> I mean, yes, mathematically, just like Wade pointed out, its true.
> > frequency = < latency
> 
> But when we compare CPU's of the same model, with different frequencies.
> 
> How much time ( in nano seconds ), do we save ?
> I mean i have really no numbers here.
> 
> But the difference between a 2,1 GHz and a 2,9 GHz ( Low End Xeon E5 / High
> End Xeon E5 ) ( when it comes to delay in "memory/what ever" allocation ),
> will be, inside an Linux OS, quiet small. And i mean nano seconds tiny/non
> existing small.
> But again, thats just my guess. Of course, if we talk about complete different
> CPU Models ( E5 vs. I7 vs. AMD vs. what ever ) we will have different 1st/2nd
> level Caches in CPU, different Architecture/RAM/everything.
> 
> But we are talking here about pure frequency issues. So we compare
> identical CPU Models, just with different frequencies.
> 
> And there, the difference, especially inside an OS and inside a productive
> environment must be nearly not existing.
> 
> I can not imagine how much an OSD / HDD needs to be hammered, that a
> server is in general not totally overloaded and that the higher frequency will
> make a measureable difference.
> 
> ----
> 
> But again, i have here no numbers/benchmarks that could proove this pure
> theory of mine.
> 
> In the very end, more cores will usually mean more GHz frequency in sum.
> 
> So maybe the whole discussion is very theoretically, because usually we
> wont run in a situation where we have to choose frequency vs. cores.
> 
> Simply because more cores always means more frequency in sum.
> 
> Except you compare totally different cpu models and generations, and this is
> even more worst theoretically and maybe pointless since the different cpu
> generations have totally different inner architecture which has a great impact
> in overall performance ( aside from numbers of frequency and cores ).
> 
> --
> Mit freundlichen Gruessen / Best regards
> 
> Oliver Dzombic
> IP-Interactive
> 
> mailto:info@xxxxxxxxxxxxxxxxx
> 
> Anschrift:
> 
> IP Interactive UG ( haftungsbeschraenkt ) Zum Sonnenberg 1-3
> 63571 Gelnhausen
> 
> HRB 93402 beim Amtsgericht Hanau
> Geschäftsführung: Oliver Dzombic
> 
> Steuer Nr.: 35 236 3622 1
> UST ID: DE274086107
> 
> 
> Am 20.01.2016 um 14:14 schrieb Wade Holler:
> > Great commentary.
> >
> > While it is fundamentally true that higher clock speed equals lower
> > latency, I'm my practical experience we are more often interested in
> > latency at the concurrency profile of the applications.
> >
> > So in this regard I favor more cores when I have to choose, such that
> > we can support more concurrent operations at a queue depth of 0.
> >
> > Cheers
> > Wade
> > On Wed, Jan 20, 2016 at 7:58 AM Jan Schermer <jan@xxxxxxxxxxx
> > <mailto:jan@xxxxxxxxxxx>> wrote:
> >
> >     I'm using Ceph with all SSDs, I doubt you have to worry about speed that
> >     much with HDD (it will be abysmall either way).
> >     With SSDs you need to start worrying about processor caches and
> memory
> >     colocation in NUMA systems, linux scheduler is not really that smart
> >     right now.
> >     Yes, the process will get its own core, but it might be a different
> >     core every
> >     time it spins up, this increases latencies considerably if you start
> >     hammering
> >     the OSDs on the same host.
> >
> >     But as always, YMMV ;-)
> >
> >     Jan
> >
> >
> >     > On 20 Jan 2016, at 13:28, Oliver Dzombic <info@xxxxxxxxxxxxxxxxx
> >     <mailto:info@xxxxxxxxxxxxxxxxx>> wrote:
> >     >
> >     > Hi Jan,
> >     >
> >     > actually the linux kernel does this automatically anyway ( sending new
> >     > processes to "empty/low used" cores ).
> >     >
> >     > A single scrubbing/recovery or what ever process wont take more than
> >     > 100% CPU ( one core ) because technically this processes are not
> >     able to
> >     > run multi thread.
> >     >
> >     > Of course, if you configure your ceph to have ( up to ) 8 backfill
> >     > processes, then 8 processes will start, which can utilize ( up to ) 8
> >     > CPU cores.
> >     >
> >     > But still, the single process wont be able to use more than one
> >     cpu core.
> >     >
> >     > ---
> >     >
> >     > In a situation where you have 2x E5-2620v3 for example, you have 2x 6
> >     > Cores x 2 HT Units = 24 Threads ( vCores ).
> >     >
> >     > So if you use inside such a system 24 OSD's every OSD will have (
> >     > mathematically ) its "own" CPU Core automatically.
> >     >
> >     > Such a combination will perform better compared if you are using 1x E5
> >     > CPU with a much higher frequency ( but still the same amout of
> >     cores ).
> >     >
> >     > This kind of CPU's are so fast, that the physical HDD ( no matter if
> >     > SAS/SSD/ATA ) will not be able to overload the cpu ( no matter
> >     which cpu
> >     > you use of this kind ).
> >     >
> >     > Its like if you are playing games. If the game is running smooth, it
> >     > does not matter if its running on a 4 GHz machine on 40%
> >     utilization or
> >     > on a 2 GHz machine with 80% utilization. Is running smooth, it can not
> >     > do better :-)
> >     >
> >     > So if your data is coming as fast as the HDD can physical deliver it,
> >     > its not important if the cpu runs with 2, 3, 4, 200 Ghz. Its
> >     already the
> >     > max of what the HDD can deliver.
> >     >
> >     > So as long as the HDD's dont get faster, the CPU's does not need to be
> >     > faster.
> >     >
> >     > The Ceph storage is usually just delivering data, not running a
> >     > commercial webserver/what ever beside that.
> >     >
> >     > So if you are deciding what CPU you have to choose, you only have to
> >     > think about how fast your HDD devices are. So that the CPU does not
> >     > become the bottleneck.
> >     >
> >     > And the more cores you have, the lower is the chance, that different
> >     > requests will block each other.
> >     >
> >     > ----
> >     >
> >     > So all in all, Core > Frequency, always. ( As long as you use
> >     fast/up to
> >     > date CPUs ). If you are using old cpu's, of course you have to
> >     make sure
> >     > that the performance of the cpu ( which does by the way not only
> >     depend
> >     > on the frequency ) is sufficient that its not breaking the HDD
> >     data output.
> >     >
> >     >
> >     >
> >     > --
> >     > Mit freundlichen Gruessen / Best regards
> >     >
> >     > Oliver Dzombic
> >     > IP-Interactive
> >     >
> >     > mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
> >     >
> >     > Anschrift:
> >     >
> >     > IP Interactive UG ( haftungsbeschraenkt )
> >     > Zum Sonnenberg 1-3
> >     > 63571 Gelnhausen
> >     >
> >     > HRB 93402 beim Amtsgericht Hanau
> >     > Geschäftsführung: Oliver Dzombic
> >     >
> >     > Steuer Nr.: 35 236 3622 1
> >     > UST ID: DE274086107
> >     >
> >     >
> >     > Am 20.01.2016 um 13:10 schrieb Jan Schermer:
> >     >> This is very true, but do you actually exclusively pin the cores
> >     to the OSD daemons so they don't interfere?
> >     >> I don't think may people do that, it wouldn't work with more than
> >     a handful of OSDs.
> >     >> The OSD might typicaly only need <100% of one core, but during
> >     startup or some reshuffling it's beneficial
> >     >> to allow it to get more (>400%), and that will interfere with
> >     whatever else was pinned there...
> >     >>
> >     >> Jan
> >     >>
> >     >>> On 20 Jan 2016, at 13:07, Oliver Dzombic <info@xxxxxxxxxxxxxxxxx
> >     <mailto:info@xxxxxxxxxxxxxxxxx>> wrote:
> >     >>>
> >     >>> Hi,
> >     >>>
> >     >>> Cores > Frequency
> >     >>>
> >     >>> If you think about recovery / scrubbing tasks its better when a
> >     cpu core
> >     >>> can be assigned to do this.
> >     >>>
> >     >>> Compared to a situation where the same cpu core needs to
> >     recovery/scrub
> >     >>> and still deliver the productive content at the same time.
> >     >>>
> >     >>> The more you can create a situation where an osd has its "own"
> >     cpu core,
> >     >>> the better it is. Modern CPU's are anyway so fast, that even
> >     SSDs cant
> >     >>> run the CPU's to their limit.
> >     >>>
> >     >>> --
> >     >>> Mit freundlichen Gruessen / Best regards
> >     >>>
> >     >>> Oliver Dzombic
> >     >>> IP-Interactive
> >     >>>
> >     >>> mailto:info@xxxxxxxxxxxxxxxxx <mailto:info@xxxxxxxxxxxxxxxxx>
> >     >>>
> >     >>> Anschrift:
> >     >>>
> >     >>> IP Interactive UG ( haftungsbeschraenkt )
> >     >>> Zum Sonnenberg 1-3
> >     >>> 63571 Gelnhausen
> >     >>>
> >     >>> HRB 93402 beim Amtsgericht Hanau
> >     >>> Geschäftsführung: Oliver Dzombic
> >     >>>
> >     >>> Steuer Nr.: 35 236 3622 1
> >     >>> UST ID: DE274086107
> >     >>>
> >     >>>
> >     >>> Am 20.01.2016 um 10:01 schrieb Götz Reinicke - IT Koordinator:
> >     >>>> Hi folks,
> >     >>>>
> >     >>>> we plan to use more ssd OSDs in our first cluster layout
> >     instead of SAS
> >     >>>> osds. (more IO is needed than space)
> >     >>>>
> >     >>>> short question: What would influence the performance more?
> more
> >     Cores or
> >     >>>> more GHz/Core.
> >     >>>>
> >     >>>> Or is it as always: Depeds on the total of
> >     OSDs/nodes/repl-level/etc ... :)
> >     >>>>
> >     >>>> If needed, I can give some more detailed information on the layout.
> >     >>>>
> >     >>>>    Thansk for feedback . Götz
> >     >>>>
> >     >>>>
> >     >>>>
> >     >>>> _______________________________________________
> >     >>>> ceph-users mailing list
> >     >>>> ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
> >     >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >     >>>>
> >     >>> _______________________________________________
> >     >>> ceph-users mailing list
> >     >>> ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
> >     >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >     > _______________________________________________
> >     > ceph-users mailing list
> >     > ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
> >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >     _______________________________________________
> >     ceph-users mailing list
> >     ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
> >     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux