On Mon, Jun 05, 2017 at 04:52:27PM +0700, CoolCold wrote: > Hello! > The data i was able to find with lsscsi: > Enclosure: > [0:0:24:0] enclosu LSI SAS3x40 0601 - > state=running queue_depth=254 scsi_level=6 type=13 device_blocked=0 timeout=0 > > Server https://www.supermicro.com/products/system/2u/2028/ssg-2028r-e1cr24l.cfm > OK. So you have a single expander SAS3 backplane. And a Broadcom/LSI SAS 3008 IT-mode HBA, which should be good. Well assuming you're running up-to-date firmware with it. > Drives are: > === START OF INFORMATION SECTION === > Vendor: TOSHIBA > Product: AL14SEB18EQ > Revision: 0101 > User Capacity: 1,800,360,124,416 bytes [1.80 TB] > Logical block size: 512 bytes > Physical block size: 4096 bytes > Lowest aligned LBA: 0 > Rotation Rate: 10500 rpm > Form Factor: 2.5 inches > Logical Unit id: 0x500003975840f759 > Serial number: X6K0A0D5FZRC > Device type: disk > Transport protocol: SAS > Local Time is: Mon Jun 5 09:51:56 2017 UTC > SMART support is: Available - device has SMART capability. > SMART support is: Enabled > Temperature Warning: Enabled > > (don't think it matters though) > Have you tried for example reading from all the disks simultaneously? Just to verify the link(s) between the HBA and the expander backplane is/are working properly, and you get good throughput with simultaneous reads from all disks? -- Pasi > On Thu, Jun 1, 2017 at 6:31 PM, Nikhil Kshirsagar <nkshirsa@xxxxxxxxxx> wrote: > > Did you check max_sectors_kb ? We found once, in some setups with raid, in > > particular with HP raid controllers, larger number of drives seemed to > > reduce the value of max_sectors_kb. Sorry if its already mentioned, I > > haven't read the entire thread in detail. > > > > On Thu, Jun 1, 2017 at 4:50 PM, Roger Heflin <rogerheflin@xxxxxxxxx> wrote: > >> > >> Also supply an "iostat -x 1 5" since that will show each disks usage. > >> > >> vmstat in my experience does not appear to show internal MD disk traffic. > >> > >> On Thu, Jun 1, 2017 at 1:33 AM, Pasi Kärkkäinen <pasik@xxxxxx> wrote: > >> > On Thu, Jun 01, 2017 at 12:59:01PM +0700, CoolCold wrote: > >> >> Hello! > >> >> Roman, i've updated the kernel to 4.11 and started "check" action, > >> >> results are basically the same, output on github > >> >> https://gist.github.com/CoolCold/663de7c006490d7fd0ac7cc98b7a6844 > >> >> 1 cpu is overloaded, not more than 1.3 - 1.4GB/sec > >> >> > >> > > >> > You need to provide more details about the actual storage setup. > >> > > >> > Like already said/asked for: > >> > > >> > - Which HBA are you using? > >> > - Which PCIe link speed are you using for the HBA? > >> > - Which driver version for the HBA? > >> > - Which HBA firmware version? > >> > > >> > - How are the disks connected to the HBA ? Direct-connect, or via an > >> > Expander? > >> > - If you have an expander, what's the (SAS) link speed/count between the > >> > HBA(s) and the Expander? > >> > > >> > > >> > -- Pasi > >> > > >> >> On Wed, May 31, 2017 at 9:14 PM, Roman Mamedov <rm@xxxxxxxxxxx> wrote: > >> >> > On Wed, 31 May 2017 19:20:10 +0700 > >> >> > CoolCold <coolthecold@xxxxxxxxx> wrote: > >> >> > > >> >> >> Creation (disable write intent bitmap, with bitmap all is much > >> >> >> worse): > >> >> >> mdadm --create -c 64 -b none -n 20 -l 10 /dev/md1 /dev/sde /dev/sdf > >> >> >> /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl /dev/sdm > >> >> >> /dev/sdn /dev/sdo /dev/sdp /dev/sdq /dev/sdr /dev/sds /dev/sdt > >> >> >> /dev/sdu /dev/sdv /dev/sdw /dev/sdx > >> >> >> > >> >> >> kernel: > >> >> >> [root@spare-a17484327407661 rovchinnikov]# cat /proc/version > >> >> >> Linux version 3.10.0-327.el7.x86_64 > >> >> >> (builder@xxxxxxxxxxxxxxxxxxxxxxx) > >> >> >> (gcc version 4.8.3 20140911 (Red Hat 4.8.3-9) (GCC) ) #1 SMP Thu Nov > >> >> >> 19 22:10:57 UTC 2015 > >> >> >> > >> >> >> So, the question is - why cpu usage is so high and I suppose is a > >> >> >> limit here? > >> >> > > >> >> > Definitely try a newer kernel, 4.4 at the very least; if no changes > >> >> > then 4.11. > >> >> > > >> >> > Also I would suggest to try out larger chunk sizes, such as 512 and > >> >> > 1024 KB. > >> >> > > >> >> > If you plan to use this long-term in production, also read up on the > >> >> > various > >> >> > RAID10 data layouts and their benefits and downsides (man md, search > >> >> > for > >> >> > "layout"; and search the Internet for benchmarks of all three). > >> >> > > >> >> > -- > >> >> > With respect, > >> >> > Roman > >> >> > >> >> > >> >> > >> >> -- > >> >> Best regards, > >> >> [COOLCOLD-RIPN] > >> >> -- > >> >> To unsubscribe from this list: send the line "unsubscribe linux-raid" > >> >> in > >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx > >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html > >> > -- > >> > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > >> > the body of a message to majordomo@xxxxxxxxxxxxxxx > >> > More majordomo info at http://vger.kernel.org/majordomo-info.html > >> -- > >> To unsubscribe from this list: send the line "unsubscribe linux-raid" in > >> the body of a message to majordomo@xxxxxxxxxxxxxxx > >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > > > > > > > -- > Best regards, > [COOLCOLD-RIPN] -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html