Manish On 01/06/2015 03:19 PM, Manish Awasthi wrote:
Hi Neil, Any findings on some of the logs I shared earlier?Thanks in advance for reply. I'm having trouble booting 3.12 kernel, should probably sort it out soon and come back with results.Manish On 12/10/2014 01:29 PM, Manish Awasthi wrote:Here is the perf report for the tests run on 3.6-11 and 3.18. Compating both the results, it just appears that raid in older version is busier than it is with the latest version. I will also monitor the system activity via `perf top` now. Also, I should be back with results on 3.12 by the weekendManish On 12/09/2014 01:56 PM, Manish Awasthi wrote:this time with attachment: manish On 12/09/2014 01:54 PM, Manish Awasthi wrote:resending: dirty_ratio same for both the kernels.vm.dirty_background_bytes = 0 vm.dirty_background_ratio = 10 vm.dirty_bytes = 0 vm.dirty_expire_centisecs = 3000 vm.dirty_ratio = 20 vm.dirty_writeback_centisecs = 500I re-ran the tests with the same set of kernel without enabling multithread support on 3.18 and measured a few things with perf.perf-stat-<kernel>.txt: test ran for some time and measured various parameters.Meanwhile I'm also running complete test under perf record. I'll share the results soon.Manish On 12/03/2014 11:51 AM, NeilBrown wrote:On Wed, 26 Nov 2014 13:41:39 +0530 Manish Awasthi <manish.awasthi@xxxxxxxxxxxxxxxxxx> wrote:Whatever data I have on comparison is attached, I have consolidated thisraid_3_18_performance.xls shows read throughput to be consistently 20% downfrom log files to excel. See if this helps.on 3.18 compared to 3.6.11.Writes are a few percent better for 4G/8G files, 20% better for 16G/32G files.unchanged above that.Given that you have 8G of RAM, that seems like it could be some change incaching behaviour, and not necessarily a change in RAID behaviour.The CPU utilization roughly follows the throughput: 40% higher when writethroughput is 20% better.Could you check if the value of /proc/sys/vm/dirty_ratio is the same for both tests. That number has changed occasionally and could affect these tests.The second file, 3SSDs-perf-2-Cores-3.18-rc1 has the "change" numbersnegative where I expected positive.. i.e. negative mean an increase. Writes consistently have higher CPU utilisation. Reads consistently have much lower CPU utilization. I don't know what that means ... it might not mean anything.Could you please run the tests between the two kernels *with* RAID. i.e. directly on an SSD. That will give us a baseline for what changes are caused by other parts of the kernel (filesystem, block layer, MM, etc). Then we cansee how much change RAID5 is contributing.The third file, 3SSDs-perf-4Core.xls seems to show significantly reducedthroughput across the board.CPU utilization is less (better) for writes, but worse for reads. That isthe reverse of what the second file shows.I might try running some tests across a set of kernel versions and see what Ican come up with. NeilBrown
Attachment:
3.6.11-vs-3.12.xls
Description: MS-Excel spreadsheet