On Thu, Jul 20, 2017 at 12:41 AM, Volodymyr Sendetskyi <volodymyrse@xxxxxxxxxx> wrote: > It is known, that git handles badly storing binary files in its > repositories at all. > This is especially about large files: even without any changes to > these files, their copies are snapshotted on each commit. So even > repositories with a small amount of code can grove very fast in size > if they contain some great binary files. Alongside this, the SVN is > much better about that, because it make changes to the server version > of file only if some changes were done. > > So the question is: why not implementing some feature, that would > somehow handle this problem? Like Git LFS or git annex? Features have been implemented to better handle large files; they're just not necessarily part of core Git. Have you checked whether one of those solutions might work for your use case? Best regards, Bryan Turner