On Thu, Oct 20, 2016 at 12:50:32PM +0200, Johannes Schindelin wrote: > That reflects my findings, too. I want to add that I found preciously > little difference between running slow-to-fast and running in numeric > order, so I gave up on optimizing on that front. Interesting. It makes a 10-15% difference here. I also point "--root" at a ram disk. The tests are very I/O heavy and sometimes fsync; even on a system with an SSD, this saves another ~10%. I know that's small potatoes compared to the Windows vs Linux times, but it might be worth exploring. > Further, I found that the Subversion tests (which run at the end) are so > close in their running time that running the tests in parallel with -j5 > does not result in any noticeable improvement when reordered. I normally don't run the Subversion tests at all. Installing cvs, cvsps, subversion, and libsvn-perl nearly doubles the runtime of the test suite for me (I imagine adding p4 to the mix would bump it further). While it's certainly possible to break them with a change in core git, it doesn't seem like a good tradeoff if I'm not touching them often. As the GfW maintainer, you probably should be running them, at least before a release. But cutting them might be a good way to speed up your day-to-day runs. I also use -j16 on a quad-core (+hyperthreads) machine, which I arrived at experimentally. At least on Linux, it's definitely worth having more threads than processors, to keep the processors busy. > I guess I will have to bite into the sour apple and try to profile, say, > t3404 somehow, including all the shell scripting stuff, to identify where > exactly all that time is lost. My guess is that it boils down to > gazillions of calls to programs like expr.exe or merely subshells. I'm not so sure it isn't gazillions of calls to git. It is testing rebase, after all, which is itself a shell script. GIT_TRACE_PERFORMANCE gives sort of a crude measure; it reports only builtins (so it will underestimate the total time spent in git), but it also doesn't make clear which programs call which, so some times are double-counted (if a builtin shells out to another builtin). But: $ export GIT_TRACE_PERFORMANCE=/tmp/foo.out $ rm /tmp/foo.out $ time ./t3404-rebase-interactive.sh real 0m29.755s user 0m1.444s sys 0m2.268s $ perl -lne ' /performance: ([0-9.]+)/ and $total += $1; END { print $total } ' /tmp/foo.out 32.851352624 Clearly that's not 100% accurate, as it claims we spent longer in git than the script actually took to run. Given the caveats above, I'm not even sure if it is in the right ballpark. But there are 11,000 git builtins run as part of that script. Even at 2ms each, that's still most of the time going to git. And obviously the fix involves converting git-rebase, which you're already working on. But it's not clear to me that the test infrastructure or shell scripts are the primary cause of the slowness in this particular case. -Peff