On Sun, 13 May 2007, Martin Koegler wrote: > --- > On Fri, 04 May 2007, Nicolas Pitre wrote: > > On Fri, 4 May 2007, Martin Koegler wrote: > > > On 2007-05-01 16:05:24, Nicolas Pitre wrote: > > > In the long term, I think, that the delta generation code needs to get > > > tunable. > > > > No. It should be self-tunable certainly, but there are way too many > > config options already, and another one for the inner working of the > > delta algorithm would be a bit too esoteric for most users and they > > won't get advantage of it. This thing really has to self tune itself. > > I would like that too, but that will probably not the case for > everybody. > > Why does git has options to control the zlib compression? > > Why are patches like "Custom compression levels for objects and packs" > (http://www.spinics.net/lists/git/msg30244.html) sent to the mailing > list? Because there is no nice ways to auto-tune them given the current zlib interface. > > > > > I tried to speed up the delta generation by searching for a common > > > > > prefix, as my blobs are mostly append only. I tested it with about > > > > > less than 1000 big blobs. The time for finding the deltas decreased > > > > > from 17 to 14 minutes cpu time. > > > > > > > > I'm surprised that your patch makes so much of a difference. Normally > > > > the first window should always match in the case you're trying to > > > > optimize and the current code should already perform more or less the > > > > same as your common prefix match does. > > > > > > A block is limited to 64k. If the file has some hundred MBs, it has to > > > match many blocks. > > > > Only if the first match is smaller than 64K. If the first match is 64K > > in size then the rest of the file is not considered at all. > > If I read the code correctly, the currect code does a new hash table > search after writing a block. Right. > As my original patch is not considering other and possibly better > matches, I rewrote the patch. The new version matches block of > unlimited length (stopping after finding a match >=64kB). If the block > it bigger than 64kB, it writes block of 64kB in a loop. I like this idea a lot. > The patch recomputes the hash of 16 bytes after writing a block, as I > haven't had time to correct the hash update code for blocks < 16 > bytes. (By the way, is this code path necessary?) It certainly helps in those cases where the block match is smaller than a Rabin window size, although I don't know how much. Sliding the window over 8 bytes might be approximately the same cost as computing the hash from scratch over 16 bytes since in the sliding case there are twice the amount of pointer dereferences. So the test should probably be "if (msize >= RABIN_WINDOW/2) ..." for the full reinitialization. > I did some tests on differenent machines: > > Repacking a test repository with big blobs: > > - original version: > Total 6452 (delta 4582), reused 1522 (delta 0) > 11752.26user 4256.21system 4:26:52elapsed 99%CPU (0avgtext+0avgdata 0maxresident)k > 0inputs+0outputs (576major+1042499851minor)pagefaults 0swaps > =>92 MB pack size > - your patch (stop at 4096 bytes block size): > Total 6452 (delta 4582), reused 1522 (delta 0) > 11587.41user 4064.73system 4:20:54elapsed 99%CPU (0avgtext+0avgdata 0maxresident)k > 0inputs+0outputs (0major+1042500427minor)pagefaults 0swaps > =>92 MB pack size > - my first patch > Total 6452 (delta 4706), reused 1450 (delta 0) > 10316.93user 4220.67system 4:02:20elapsed 99%CPU (0avgtext+0avgdata 0maxresident)k > 0inputs+0outputs (0major+1045003394minor)pagefaults 0swaps > =>92 MB pack size > - attached patch > Total 6452 (delta 4581), reused 1522 (delta 0) > 11354.38user 5451.60system 4:40:09elapsed 99%CPU (0avgtext+0avgdata 0maxresident)k > 0inputs+0outputs (0major+1371504762minor)pagefaults 0swaps > =>75 MB pack size This is quite weird. I wonder what might cause such a large difference in pack size. Your first patch is probably faster due to the use of memcmp() which is certainly highly optimized, more than the comparison loop we have. It is unfortunate that there is no library function to find the number of identical bytes between two buffers. Or is there some? But the size difference? That has certainly something to do with your data set since your patch makes no significant difference on the git.git nor the Linux kernel repos. Would it be possible for me to have a copy of your repo for further analysis? Nicolas - To unsubscribe from this list: send the line "unsubscribe git" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html