I did something stupid like trying to push a copy of WebKit[1] into my GitHub account. This is ~5.2 GiB of data, which GitHub prefers not to accept. Ok ... $ git push --all git@xxxxxxxxxx:spearce/wk.git Counting objects: 2752427, done. Delta compression using up to 12 threads. Compressing objects: 100% (442684/442684), done. remote: fatal: pack exceeds maximum allowed size Writing objects: 100% (2752427/2752427), 5.28 GiB | 8.86 MiB/s, done. Total 2752427 (delta 2225007), reused 2752427 (delta 2225007) fatal: The remote end hung up unexpectedly fatal: The remote end hung up unexpectedly Notice GitHub prints "remote: fatal: pack exceeds maximum allowed size". That interrupted my "Writing objects" progress meter, and then git push just kept going and wrote really really fast (170 MiB/s!) until the entire pack was sent. A similar thing happens on https:// if the remote HTTP server goes away in the middle of sending the pack. Except its slower to send the remainder of the pack before git push finally terminates with an error. Shouldn't git push realize its stream is broken and stop writing when the peer is all like "uh, no, I'm not going to do that, but thanks for trying"? [1] https://webkit.googlesource.com/WebKit/ -- To unsubscribe from this list: send the line "unsubscribe git" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html