On Thu, Apr 20, 2017 at 04:41:18PM -0400, David Turner wrote: > When core.packedGitLimit is exceeded, git will close packs. If there > is a repack operation going on in parallel with a fetch, the fetch > might open a pack, and then be forced to close it due to > packedGitLimit being hit. The repack could then delete the pack > out from under the fetch, causing the fetch to fail. > > Increase core.packedGitLimit's default value to prevent > this. > > On current 64-bit x86_64 machines, 48 bits of address space are > available. It appears that 64-bit ARM machines have no standard > amount of address space (that is, it varies by manufacturer), and IA64 > and POWER machines have the full 64 bits. So 48 bits is the only > limit that we can reasonably care about. We reserve a few bits of the > 48-bit address space for the kernel's use (this is not strictly > necessary, but it's better to be safe), and use up to the remaining > 45. No git repository will be anywhere near this large any time soon, > so this should prevent the failure. I happened to run into another case of this which failed fairly reliably, and confirmed that bumping packedGitLimit fixed it. Having forgotten completely that we did this bump here, I wrote a similar patch. :-/ The content of the patch is basically the same, but for the sake of the list archive, here's the bit from my commit message about the reproduction recipe (in case we should ever need to revisit it again, though I'm pretty sure that your patch should fix the problem for good). -- >8 -- You can see this situation in practice with strace: # two objects, one big and one small; for the sake of # speed, we'll make our big object only 1.1MB and drop # core.packedGitLimit to 1MB. Note that we care about # on-disk compressed sizes here, so we need uncompressible # data. small=$(echo small | git hash-object -w --stdin) big=$(dd if=/dev/urandom bs=1k count=1100 | git hash-object -w --stdin) # Now put each in a pack, which will get mmap'd # separately. echo $small | git pack-objects objects/pack/pack echo $big | git pack-objects objects/pack/pack # Now we look in the packs alternately, and observe the # access patterns. { echo $small echo $big echo $small } | strace -e open,close,mmap,munmap \ git -c core.packedgitlimit=1M \ -c core.packedGitWindowSize=512K \ cat-file --batch >/dev/null The trace will look something like this (I've omitted the uninteresting .idx bits): open("./objects/pack/pack-b2a8452664fe8848fd0a78be8b6f69ce65af3c57.pack", O_RDONLY|O_CLOEXEC) = 3 mmap(NULL, 47, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7fb5890af000 close(3) = 0 There we opened the small pack, mapped the whole thing, and closed it. open("./objects/pack/pack-acfa5f8886b391d0b2475ef3f19bcc387ce19271.pack", O_RDONLY|O_CLOEXEC) = 3 mmap(NULL, 524288, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7fb589011000 mmap(NULL, 1130496, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) = 0x7fb588efd000 munmap(0x7fb5890af000, 47) = 0 mmap(NULL, 524288, PROT_READ, MAP_PRIVATE, 3, 0x80000) = 0x7fb5881c0000 munmap(0x7fb589011000, 524288) = 0 mmap(NULL, 78211, PROT_READ, MAP_PRIVATE, 3, 0x100000) = 0x7fb58909a000 munmap(0x7fb588efd000, 1130496) = 0 And here we open the larger pack. Note that we don't close the descriptor because the file is bigger than our 512k window. And in fact we end up creating several maps as we progress through the file. But before making the second one, which would push us over our packedGitLimit, we close the original small 47-byte map. open("./objects/pack/pack-b2a8452664fe8848fd0a78be8b6f69ce65af3c57.pack", O_RDONLY|O_CLOEXEC) = 4 mmap(NULL, 47, PROT_READ, MAP_PRIVATE, 4, 0) = 0x7fb5890af000 close(4) = 0 And here we access the small pack again, but we have to re-open and map it from scratch. If there had been a simultaneous repack and if this had been pack-objects, we would now fail the operation. These numbers are shrunk to make the test more tractable, but you can see the same thing with the defaults when fetching from a large repository. The real-world situation that would trigger this is something like: 0. Repo has 8GB or larger packfile (this is the default core.packedGitLimit on 64-bit machines). 1. Somebody pushes a "small" pack (smaller than the 1GB default window). 2. Somebody tries to clone/fetch the result. The pack-objects process on the server starts by generating the list of objects to write, committing to the use of those packs. After accessing the recent objects in the small pack, it opens the big pack and closes the small one. 3. Meanwhile, git-repack runs and folds the small pack into a larger one. 4. The pack-objects process hits the write phase, but it can no longer access the recent objects from the small pack, and dies.