I just untar'ed several(8) copies of of a 3.4GB test dir into separate test directories, The untarring is in background, but I wait for each "tar" to return, so I can list out the times I untar the same image 8 times in 8 different dirs. The initial tars go very quickly, consuming 100% cpu as the first tars basically extract to buffer memory, but after the 4th, times begin increasing almost geometrically: 1...5.03sec 0.12usr 4.90sys (99.84% cpu) 2...5.17sec 0.12usr 5.03sys (99.71% cpu) 3...5.36sec 0.13usr 5.21sys (99.68% cpu) 4...5.35sec 0.15usr 5.17sys (99.66% cpu) 5...7.36sec 0.12usr 5.69sys (79.14% cpu) 6...27.81sec 0.16usr 6.76sys (24.93% cpu) 7...85.54sec 0.21usr 7.33sys (8.83% cpu) 8...101.64sec 0.25usr 7.88sys (8.00% cpu) 2nd run: 1...5.23sec 0.12usr 5.10sys (99.73% cpu) 2...5.25sec 0.15usr 5.08sys (99.71% cpu) 3...6.08sec 0.13usr 5.09sys (85.86% cpu) 4...5.31sec 0.14usr 5.15sys (99.71% cpu) 5...14.02sec 0.18usr 6.28sys (46.11% cpu) 6...23.32sec 0.17usr 6.47sys (28.50% cpu) 7...31.14sec 0.21usr 6.74sys (22.32% cpu) 8...82.36sec 0.22usr 7.23sys (9.05% cpu) Now -- and for 3-4 minutes after this point, I see 7 kworker processes -- 6 of them on the even cpu's (2-node numa) and 1 on an odd cpu. The 7 consume about 12-18% cpu each and the odd one is "matched" by a "flush-254:odd" process that has the same cpu as the odd kworker. I added a "time sync" after the script finished (to show, approximately, how long disk activity continues) 222.95sec 0.00usr 0.23sys (0.10% cpu) So what are all the kworkers doing and does having 6 of them do things at the same time really help disk-throughput? Seems like they would conflict w/each other, cause disk contention, and extra fragmentation as they do things? If they were all writing to separate disks, that would make sense, but do that many kworker threads need to be finishing out disk I/O on 1 disk? FWIW, I remove and create a skeleton dir-struct before I untar to the dirs. The "rm -fr" on those dirs happens in parallel with a "wait" at the end to wait for all of them. That takes: 1.82sec 0.07usr 12.91sys (711.34% cpu) Creating the dirnames + empty filenames) takes: (for ~1-2 dirs, and 5000+ filenames) 6.85sec 0.38usr 11.68sys (176.03% cpu) So is it efficient to use that many writers on 1 disk? Note the disks' max-write speed in writing a large contiguous multi gig file, is about 1GB/s and is mounted (showing mount output): /dev/mapper/HnS-Home on /home type xfs \ (rw,nodiratime,relatime,swalloc,attr2,largeio,inode64,allocsize=128k,\ logbsize=256k,sunit=128,swidth=1536,noquota) _______________________________________________ xfs mailing list xfs@xxxxxxxxxxx http://oss.sgi.com/mailman/listinfo/xfs