On Fri, 2009-03-27 at 10:59 -0400, Ross S. W. Walker wrote: > John A. Sullivan III wrote: > > > > On Fri, 2009-03-27 at 10:09 -0400, Ross S. W. Walker wrote: > > > > > > I'm sorry you lost me here on this test. NFS file transfers are > > > completely different then iSCSI as NFS uses the OS' (for better or > > > worse) implementation of file io and not block io, so using NFS as a > > > test isn't the best. > > > > Sorry if I wasn't clear. These were iSCSI tests but using file > > services, e.g., cp, mv. > > > > > My whole idea of mentioning the multiple ios was basically if one > > > disk's maximum transfer speed is 40MB/s and you put 4 in a RAID0 then > > > if you tested that RAID0 with an app that did 1 io at a time you would > > > still only see 40MB/s and not the 160MB/s it is capable of. Obviously > > > that isn't what's happening here. > > > > Yes, that's what I was hoping for :) > > > > > One would hope throughput scales up with number of simultaneous ios or > > > there is something broken. Maybe it is dm-raid that is broken, or > > > maybe it's just dm-raid+iSCSI that's broken, or maybe it's just > > > dm-raid+iSCSI+ZFS that's broken? Who knows? Iit's obvious that this > > > setup won't work though. > > > > <snip> > > > > > > I would take a step back at this point and re-evaluate if running > > > software RAID over iSCSI is the best. It seems like a highly complex > > > system to get the performance you need. > > > > Yes, that's what we've decided to do which has forced a bit of redesign > > on our part. > > > > > The best, most reliably performing results I have seen with iSCSI have > > > been those where the target is performing the RAID (software or > > > hardware) and the initiator treats it as a plain disk, using interface > > > aggregates or multi-paths to gain throughput. > > > > Strangely, multipath didn't seem to help as I thought it would. Thanks > > for all your help - John > > Not a problem, in fact I have a vested interest in you succeeding because > we are looking at moving our iSCSI over from Linux to Solaris to get > that warm fuzzy ZFS integrity assurance feeling. > > It is just (and we are running Solaris 10u6 and not OpenSolaris here), we > find the user space iscsitgt daemon leaves something to be desired over > the kernel based iSCSI Enterprise Target on Linux. Also we have noticed > strange performance abnormalities in our tests of iSCSI to a ZFS ZVOL > as opposed to iSCSI to a raw disk. > > Maybe it's just the price you pay doing iSCSI to a COW file system instead > of a raw disk? I dunno, I always prefer to get it all then compromise, > but sometimes one doesn't have a choice. > > Oh BTW ZFS as an NFS server works brilliantly. We have moved 50% of our > VMs over from iSCSI to NFS and performance is great as each VM has > a level of guaranteed throughput so no one VM can starve out the others. > > We are going to stick with our Linux iSCSI target for a wee bit longer > though to see if the ZFS/ZVOL/iscsitgt situation improves in 10u7. Gives > me time to learn the Solaris LiveUpgrade feature. <snip> Hi Ross. I hope you don't mind that I copied the mailing list in case it is of interest to anyone. We are using COMSTAR which has given a substantial improvement over iscsitgt. I was concerned about the overhead of writing to the file system although I'm not sure that's really the issue. When we run bonnie++ tests on the unit itself, we see it able to read at over 500 MB/s and write at almost 250. That's in a unit with four mirror pairs. The killer issue seems to be latency. As seen in a previous post, the latency on the opensolaris unit is dramatically higher than on the Linux unit and more erratic. That brings me back to a question about NFS. I could use NFS for our scenario as much of our disk I/O is file system I/O (hence the 4KB block size problem and impact of latency). We've thought about switching to NFS and I've just not benchmarked it because we are so behind on our project (not a good reason). However, everything I read says NFS is slower iSCSI especially for random file I/O as we expect although I wonder if it would help us on sequential I/O since we can change the I/O size to something much larger than the file system block size. So, what are the downsides of using NFS? Will I see the same problem because of latency (since we get plenty of raw throughput on disk) or is it a solution to our latency problem because we can tell it to use large transfer sizes? Thanks - John > -- John A. Sullivan III Open Source Development Corporation +1 207-985-7880 jsullivan@xxxxxxxxxxxxxxxxxxx http://www.spiritualoutreach.com Making Christianity intelligible to secular society -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel