best practices? 500+ win/mac computers, gluster, samba, new SAN, new hardware

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 1/28/12 6:02 PM, Brian Candler wrote:
> On Sat, Jan 28, 2012 at 05:31:28PM -0600, D. Dante Lorenso wrote:
>> Thinking about buying 8 servers with 4 x 2TB 7200 rpm SATA drives
>> (expandable to 8 drives).  Each server will have 8 network ports and
>> will be connected to a SAN switch using 4 ports link aggregated and
>> connected to a LAN switch using the other 4 ports aggregated.  The
>> servers will run CentOS 6.2 Linux.  The LAN side will run Samba and
>> export the network shares, and the SAN side will run Gluster daemon.
>
> Just a terminology issue, but Gluster isn't really a SAN, it's a distributed
> NAS.
>
> A SAN uses a block-level protocol (e.g. iSCSI), into which the client runs a
> regular filesystem like ext4 or xfs or whatever.  A NAS is a file-sharing
> protocol (e.g.  NFS).  Gluster is the latter.

I need a word to describe the switch that I'll plug all my storage 
machines into.  Distributed NAS sounds good.  Might have a few iSCSI on 
there too, however.

>> With 8 machines and 4 ports for SAN each, I need 32 ports total.
>> I'm thinking a 48 port switch would work well as a SAN back-end
>> switch giving me left over space to add iSCSI devices and backup
>> servers which need to hook into the SAN.
>
> Out of interest, why are you considering two different network fabrics? Are
> there one set of clients which are talking CIFS and a different set of
> clients using the Gluster native client?

Most of my clients (95%) are all Windows 7 workstations.  The only way I 
think I can get GlusterFS to work with Win7 is through Samba.  I was 
planning to use SMB/CIFS on the Win7 side of the network (using 2 bonded 
ports) and use Gluster native client on the storage side (using another 
2 bonded ports).

>> 4) Performance tuning.  So far, I've discovered using dd and iperf
>> to debug my transfer rates.  I use dd to test raw speed of the
>> underlying disks (should I use RAID 0, RAID 1, RAID 5 ?)
>
> Try some dd measurements onto a RAID 5 volume, especially for writing, and
> you'll find it sucks.
>
> I also suggest something like bonnie++ to get a more realistic performance
> measurement than just the dd throughput, as it will include seeks and
> filesystem operations (e.g.  file creations/deletions)

Good advice, I'll check into bonnie++.

>> Perhaps if my drives on each of the 8
>> servers are RAID 0, then I can use "replicate 2" through gluster and
>> get the "RAID 1" equivalent.  I think using replicate 2 in gluster
>> will 1/2 my network write/read speed, though.
>
> In theory Gluster replication ought to improve your read speed, since some
> clients can access one copy spindle while other clients access the other.
> But I'm not sure how much it will impact the write speed.
>
> I would however suggest that building a local RAID 0 array is probably a bad
> idea, because if one disk of the set fails, that whole filesystem is toast.
>
> Gluster does give you the option of a "distributed replicated" volume, so
> you can get both the "RAID 0" and "RAID 1" functionality.

If you have 8 drives connected to a single machine, how do you introduce 
those drives to Gluster?  I was thinking I'd combine them into a single 
volume using RAID 0 and mount that volume on a box and turn it into a 
brick.  Otherwise you have to add 8 separate bricks, right?  That's not 
better is it?

-- Dante

D. Dante Lorenso
dante at lorenso.com
972-333-4139


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux