On 3/5/2013 9:53 AM, Adam Goryachev wrote: > On 05/03/13 20:30, Stan Hoeppner wrote: > BCP = Best Computing Practise ? Typically "best current practice". > Thanks to the tip about running fio on windows, I think I've now come > full circle.... Today I had numerous complaints from users that their > outlook froze/etc, and some cases were the TS couldn't copy a file from > the DC to it's local C: (iSCSI). The cause was the DC was logging events > with event ID 2020 which is "The server was unable to allocate from the > system paged pool because the pool was empty". Supposedly the solution > to this is tuning two random numbers in the registry, not much is said > what the consequences of this are, nor about how to calculate the > correct value. ... > Running the same fio test on the same TS (win2003) against a SMB share > from the DC (SMB -> Win2000 -> Xen -> iSCSI -> etc) >> READ: io=16384MB, aggrb=14818KB/s, minb=14818KB/s, maxb=14818KB/s, mint=1132181msec, maxt=0msec >> WRITE: io=16384MB, aggrb=8039KB/s, minb=8039KB/s, maxb=8039KB/s, mint=2086815msec, maxt=0msec Run FIO on the DC itself and see what your NTFS throughput is to this 300GB filesystem. Use a small file, say 2GB, since the FS is nearly full. Post results. Fire up the Windows CLI FTP client in a TS session DOS box and do a GET and PUT into this filesystem share on the DC. This will tell us if the TS to DC problem is TCP in general or limited to SMB. Post transfer rate results for GET and PUT. > This is pretty shockingly slow, and seems to clearly indicate why the > users are so upset... 14MB/s read and 8MB/s write, it's a wonder they > haven't formed a mob and lynched me yet! I've never used FIO on Windows against a Windows SMB share. And according to Google nobody else does. So before we assume these numbers paint an accurate picture of your DC SMB performance, and that the CPU burn isn't due to an anomaly of FIO, you should run some simple Windows file copy tests in Explorer and use Netmeter to measure the speed. If they're in the same ballpark then you know you can somewhat trust FIO for SMB testing. If they're wildly apart, probably not. > However, the truly useful information is that during the read portion of > the test, the DC has a CPU load of 100% (no variation, just pegged at > 100%), during the write portion, it fluctuates between 80% to 100%. That 100% CPU is bothersome. Turn off NTFS compression on any/all NTFS volumes residing on SAN LUNs on the SSD array. You're burning 100% DC CPU at ~12MB/s data rate on the DC, so I can only assume it's turned on for this 300GB volume. These SSDs do on the fly compression, and very quickly as you've seen. Doing NTFS compression on top simply wastes cycles. This should drop the CPU burn for new writes to the filesystem. It probably won't for reads, since NTFS must still decompress the existing 250GB+ of files. If CPU drops considerably for writes but reads still eat 100%, the only fix for this is to backup the filesystem, reformat the device, then restore. On the off chance that NTFS compression is more efficient than the SandForce controller, you probably want to increase the size of the volume before formatting. And in fact, Sysadmin 101 tells us never to run a production filesystem at more than ~70% capacity, so it would be smart to bump it up to 400GB to cover your bases. My second recommendation is to turn off the indexing service for all these NTFS volumes as well as this will conserve CPU cycle as well. > Extended the data drive from 279GB to 300GB (it was 90% full, now 84% full) Growing filesystem in small chunks like this is a recipe for disaster. Your free space map is always heavily fragmented and is very large. The more entries the filesystem driver must walk the more CPU you burn. Recall we just discussed the table walking overhead of the md/RAID stripe cache? Filesystem maps/tables/B+ trees are much, much larger structures. When they don't fit it cache we read memory, and when they don't fit in memory (remember you "pool" problem) we must read from disk. If you've been expanding this NTFS this way for a while it would also explain some of your CPU burn at the DC. FYI, XFS is a MUCH higher performance and much more efficient filesystem than NTFS ever dreamed of becoming, but even XFS suffers slow IO and CPU burn due to heavily fragmented free space. -- Stan -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html