On Tue, 24 Apr 2007, Andreas Ericsson wrote: > Using a more efficient compression algorithm for the objects > themselves (bzip2, anyone?) will most likely reduce storage size an > order of magnitude more than reducing the size of the hash, although > at the expense of CPU-efficiency. An order of magnitude I really doubt it. Maybe 20% could be a really optimistic prediction. But if bzip2 could reduce the repo by 20%, it will slow runtime usage of that repo by maybe 100%. That is not worth it. This is also the reason why we changed the default zlib compression level from "best" to "default". Nicolas - To unsubscribe from this list: send the line "unsubscribe git" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html