> It's pretty simple to set up. You just need to be familliar with iSCSI > tools and software RAID tools, all of which are almost certainly in your > distro's apt/yum repositories. Figured I would ask. Never know, might be some cool management tools that help keep an eye on things. The setup sounds simple enough as you say. >> I need a machine which will become the aggregator, plenty of memory, >> multi-port Ethernet card and of course an FC HBA. >> FC storage will be attached to this machine. Then, iSCSI storage targets >> will also export to this machine. > Not quite sure I follow this - you want to use FC storage and combine it > with iSCSI storage into a bigger iSCSI storage pool? No reasib why not, I > suppose. I need relatively small central GFS for the shared data between the servers but the rest is for media and such. I'll need to have FC HBA's in every LAMP server since it needs access to GFS but media servers, those who only offload, don't need access to GFS so why install an FC HBA in those? Rather, I could export the FCC storage as part of the aggregate volume so that any server can gain access over iSCSI. Seems that would give me more options. Am I thinking incorrectly on this? > Note that software RAID only goes up to RAID 6 (i.e. n+2). So you cannot > lose more than 2 nodes (FC or iSCSI), otherwise you lose your data. So basically, can't lose more than one storage chassis. Since they are all RAID with hot swap, I should be ok so long as I keep a close eye on it all which one needs to anyhow. That's why I wondered about any software tools that might help but I'm sure there's a ton out there which will work. > yum install iscsi-target No?! I saw this a long time ago as a new concept, never looked at it since. Wonderful :). > When a node needs to take over, it fences the other node, connects > the iSCSI shares, starts up the RAID on them, assumes the floating IP and > exports the iSCSI/NFS shares. So this sounds like the complex part then because being able to fail over or switch over seems terribly important to me. If one machine is handling all this I/O and something happens to it, everything is down until that one machine is fixed. This, I would need to find a solution for first. I need to better understand how I would do this fencing. Mike -- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster