Nicolas Pitre wrote: > On Tue, 24 Apr 2007, Andreas Ericsson wrote: > >> Using a more efficient compression algorithm for the objects >> themselves (bzip2, anyone?) will most likely reduce storage size an >> order of magnitude more than reducing the size of the hash, although >> at the expense of CPU-efficiency. > > An order of magnitude I really doubt it. Maybe 20% could be a really > optimistic prediction. But if bzip2 could reduce the repo by 20%, it > will slow runtime usage of that repo by maybe 100%. That is not worth > it. > > This is also the reason why we changed the default zlib compression > level from "best" to "default". > ... order of magnitude *more than reducing the size of the hash*. -- Andreas Ericsson andreas.ericsson@xxxxxx OP5 AB www.op5.se Tel: +46 8-230225 Fax: +46 8-230231 - To unsubscribe from this list: send the line "unsubscribe git" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html