How did you obtain those numbers? Were the 8k and 16k numbers per osd, or the raw throughput of 1 client? -Sam On Fri, Nov 9, 2012 at 1:34 PM, Stefan Priebe - Profihost AG <s.priebe@xxxxxxxxxxxx> wrote: > Am 09.11.2012 um 22:21 schrieb Samuel Just <sam.just@xxxxxxxxxxx>: > >> Can you describe the osd and client set up (number of nodes, number of >> osds per node, journal disks, replication level, and osd disks)? >> Looks like a lot of time is spent looking up objects in the filestore >> (lfn_open, etc). > > Sure. I have 5 nodes each with 4 ssds one per osd. The graph is from one osd process. Replication level was set to two. Journal is on tmpfs. > > Anything else you need to know? > > Stefan > >> -Sam >> >> On Fri, Nov 9, 2012 at 2:21 AM, Stefan Priebe - Profihost AG >> <s.priebe@xxxxxxxxxxxx> wrote: >>> New graph from today. fsetxattr seems to take a lot of CPU too. >>> >>> Am 09.11.2012 11:09, schrieb Stefan Priebe - Profihost AG: >>> >>>> >>>> Disabling the logging with: >>>> debug lockdep = 0/0 >>>> debug context = 0/0 >>>> debug crush = 0/0 >>>> debug buffer = 0/0 >>>> debug timer = 0/0 >>>> debug journaler = 0/0 >>>> debug osd = 0/0 >>>> debug optracker = 0/0 >>>> debug objclass = 0/0 >>>> debug filestore = 0/0 >>>> debug journal = 0/0 >>>> debug ms = 0/0 >>>> debug monc = 0/0 >>>> debug tp = 0/0 >>>> debug auth = 0/0 >>>> debug finisher = 0/0 >>>> debug heartbeatmap = 0/0 >>>> debug perfcounter = 0/0 >>>> debug asok = 0/0 >>>> debug throttle = 0/0 >>>> >>>> reduced the CPU load about 50% ! So each OSD process now takes only one >>>> whole 3.6Ghz core instead of two. >>>> >>>> Have you looked at my latest profile graph with disabled debug options? >>>> >>>> Greets, >>>> Stefan >>>> >>>> >>>> Am 08.11.2012 17:06, schrieb Mark Nelson: >>>>> >>>>> On 11/08/2012 09:45 AM, Stefan Priebe - Profihost AG wrote: >>>>>> >>>>>> Am 08.11.2012 16:01, schrieb Sage Weil: >>>>>>> >>>>>>> On Thu, 8 Nov 2012, Stefan Priebe - Profihost AG wrote: >>>>>>>> >>>>>>>> Is there any way to find out why a ceph-osd process takes around 10 >>>>>>>> times more >>>>>>>> load on rand 4k writes than on 4k reads? >>>>>>> >>>>>>> >>>>>>> Something like perf or oprofile is probably your best bet. perf can be >>>>>>> tedious to deploy, depending on where your kernel is coming from. >>>>>>> oprofile seems to be deprecated, although I've had good results with >>>>>>> it in >>>>>>> the past. >>>>>> >>>>>> >>>>>> I've recorded 10s with perf - it is now a 300MB perf.data file. Sadly >>>>>> i've no idea what todo with it next. >>>>> >>>>> >>>>> Pour yourself a stiff drink! (haha!) >>>>> >>>>> Try just doing a "perf report" in the directory where you've got the >>>>> data file. Here's a nice tutorial: >>>>> >>>>> https://perf.wiki.kernel.org/index.php/Tutorial >>>>> >>>>> Also, if you see missing symbols you might benefit by chowning the file >>>>> to root and running perf report as root. If you still see missing >>>>> symbols, you may want to just give up and try sysprof. >>>>> >>>>>> >>>>>>> would love to see where the CPU is spending most of it's time. >>>>>>> This is >>>>>>> on current master? >>>>>> >>>>>> Yes >>>>>> >>>>>>> I expect there are still some low-hanging fruit that >>>>>>> can bring CPU utilization down (or even boost iops). >>>>>> >>>>>> Would be great to find them. >>>>>> >>>>>> Stefan >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>> >>>>> >>>>> -- >>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>> >>>> -- >>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html