Any logs from GlusterFS? That would help a lot. 2008/7/12 skimber <nabble@xxxxxxxxxxxxxxxxx>: > > Hi Everyone, > > I have just set up glusterfs for the first time using two server machines > and two clients. > > I'm trying to rsync a large amount of data (approx 1.2m files totalling > 14Gb) from another server using the following command (real names changed!) > > rsync -v --progress --stats --links --safe-links --recursive > www.domain.com::mydir /mydir > > Rsync connects, gets the file list and then as soon as it starts trying to > copy files it dies, like this: > > receiving file list ... > 1195208 files to consider > .bash_history > rsync: connection unexpectedly closed (24714877 bytes received so far) > [generator] > rsync error: error in rsync protocol data stream (code 12) at io.c(453) > [generator=2.6.9] > > As soon as that happens the following lines are added, twice, to syslog: > > Jul 12 13:23:03 server01 kernel: hdb: dma_intr: status=0x51 { DriveReady > SeekComplete Error } > Jul 12 13:23:03 server01 kernel: hdb: dma_intr: error=0x40 { > UncorrectableError }, LBAsect=345341, sector=345335 > Jul 12 13:23:03 server01 kernel: ide: failed opcode was: unknown > Jul 12 13:23:03 server01 kernel: end_request: I/O error, dev hdb, sector > 345335 > > We were previously looking at a DRBD based solution with which rsync worked > fine. Nothing has changed at the remote server, the only difference is we > are now using Gluster instead of DRBD. > > We are using glusterfs-1.3.9 and fuse-2.7.3glfs10 on Debian Etch > > The server config looks like this: > > volume posix > type storage/posix > option directory /data/export > end-volume > > volume plocks > type features/posix-locks > subvolumes posix > end-volume > > volume brick > type performance/io-threads > option thread-count 4 > subvolumes plocks > end-volume > > volume brick-ns > type storage/posix > option directory /data/export-ns > end-volume > > volume server > type protocol/server > option transport-type tcp/server > option auth.ip.brick.allow * > option auth.ip.brick-ns.allow * > subvolumes brick brick-ns > end-volume > > > > And the client config looks like this: > > > volume brick1 > type protocol/client > option transport-type tcp/client # for TCP/IP transport > option remote-host data01 # IP address of the remote brick > option remote-subvolume brick # name of the remote volume > end-volume > > volume brick2 > type protocol/client > option transport-type tcp/client > option remote-host data02 > option remote-subvolume brick > end-volume > > volume brick-ns1 > type protocol/client > option transport-type tcp/client > option remote-host data01 > option remote-subvolume brick-ns # Note the different remote volume name. > end-volume > > volume brick-ns2 > type protocol/client > option transport-type tcp/client > option remote-host data02 > option remote-subvolume brick-ns # Note the different remote volume name. > end-volume > > volume afr1 > type cluster/afr > subvolumes brick1 brick2 > end-volume > > volume afr-ns > type cluster/afr > subvolumes brick-ns1 brick-ns2 > end-volume > > volume unify > type cluster/unify > option namespace afr-ns > option scheduler rr > subvolumes afr1 > end-volume > > > Any advice or suggestions would be greatly appreciated! > > Many thanks > > Simon > -- > View this message in context: > http://www.nabble.com/Rsync-failure-problem-tp18420195p18420195.html > Sent from the gluster-devel mailing list archive at Nabble.com. > > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel > -- Amar Tumballi Gluster/GlusterFS Hacker [bulde on #gluster/irc.gnu.org] http://www.zresearch.com - Commoditizing Super Storage!