Re: recommendations for copying large filesystems

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]




Am 21.06.2008 um 23:44 schrieb Matt Morgan:

O
Then if you get the network sorted out, the fastest & most reliable way I know to copy lots of files is

star --copy

You can get star with

yum install star



Now that I know the details - I don' think this is going to work. Not with 100 TB of data. It kind-of-works with 1 TB.
Can anybody comment on the feasibility of rsync on 1 million files?
Maybe DRBD would be a solution.
If you can retrofit DRDB to an existing setup...

If not it's faster to move the drives physically - believe me, this will create far less problems. In a SAN, you would have the possibility of synching the data outside of the filesystem, during normal operations.


100 TB is a lot of data.
How do you back that up, BTW?
What is your estimated time to restore it from the medium you back it up to?



cheers,
Rainer
--
Rainer Duffner
CISSP, LPI, MCSE
rainer@xxxxxxxxxxxxxxx


_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos

[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux