Am 21.06.2008 um 23:44 schrieb Matt Morgan:
O
Then if you get the network sorted out, the fastest & most reliable
way I know to copy lots of files is
star --copy
You can get star with
yum install star
Now that I know the details - I don' think this is going to work. Not
with 100 TB of data. It kind-of-works with 1 TB.
Can anybody comment on the feasibility of rsync on 1 million files?
Maybe DRBD would be a solution.
If you can retrofit DRDB to an existing setup...
If not it's faster to move the drives physically - believe me, this
will create far less problems.
In a SAN, you would have the possibility of synching the data outside
of the filesystem, during normal operations.
100 TB is a lot of data.
How do you back that up, BTW?
What is your estimated time to restore it from the medium you back
it up to?
cheers,
Rainer
--
Rainer Duffner
CISSP, LPI, MCSE
rainer@xxxxxxxxxxxxxxx
_______________________________________________
CentOS mailing list
CentOS@xxxxxxxxxx
http://lists.centos.org/mailman/listinfo/centos