Re: mkfs.ext3 on a 9TB volume -- [Practices] Striping across intelligent RAID card

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]



Francois Caen <frcaen@xxxxxxxxx> wrote:
> Wow! Odd! RH says 8TB but ext3 FAQ says 4TB.

Any filesystem originally designed for 32-bit x86 is full of
signed 32-bit structures.  The 2^31 * 512 = 1.1TB (1TiB)
limit comes from those structures using a 512 sector size.

Ext3 has used a couple of different techniques to allow
larger and larger support.  Depending on the hardware, kernel
(especially 2.4), etc..., there can be limits at 1, 2, 4, 8
and 16TiB.

Which is why the "common denominator" is 1.1TB (1TiB).  It
was rather enfuriating in front of a client when I attempted
to mount one 2TB Ext3 volume over a SAN created by one Red
Hat Enterprise Linux from another.  For the "heck of it" -- I
created a 1TB and tried again ... it worked.

ReiserFS 3 has the same issue, it grew up as a PC LBA32
filesystem.  ReiserFS 4 is supposed 64-bit clean.  Although
JFS for Linux came from OS/2 (32-bit PC) and not AIX/Power
(true 64-bit), it was designed to be largely "64-bit clean"
too.  XFS came from Irix/MIPS4000+ (true 64-bit).

Both JFS and XFS would _not_ work on 32-bit Linux until
patched with complete POSIX32 Large File Support (LFS).  LFS
became standard in the x86 target Linux kernel 2.4 and GLibC
2.2 (Red Hat Linux 7.x / Red Hat Enterprise Linux 2.1).

> Joshua, thanks for the reply on this.
> There's something kludgy about having to do softraid across
> 2 partitions before formatting.

RAID-0 is an _ideal_ software RAID.  Striping is best handled
by the OS, which can schedule over multiple I/O options.  In
2x and 4x S940 Opteron systems with at least one AMD8131
(dual PCI-X channels), I put a 3Ware card on each PCI-X
channel connected to the same CPU and stripe with LVM.  The
CPU interlaces writes directly over two (2) PCI-X channels to
two (2) 3Ware cards.  Ultimate I/O affinity, no bus
arbitration overhead, etc..., as well as the added
performance of striping.

The only negative is if one 3Ware card dies.  But that's why
I keep a spare per N servers (typically 1 for every 4
servers, 8 cards total).

> It adds a layer of complexity and reduces reliability.

That varies.  Yes, various kernel storage approaches --
especially LVM2/Device Manager (DM) at this point -- have
race conditions if you use more than one operation.  E.g.,
resizing and snapshots, RAID-1 (DM) atop of RAID-0, etc... 
But I _only_ use LVM/LVM2 with its native RAID-0 stripe, and
across two (2) 3Ware cards.

I've yet to have an issue.  But that's probably because LVM2
doesn't require DM for RAID-0.  DM is required for RAID-1,
snapshots, FRAID meta-data, etc...


Joshua Baker-LePain <jlb17@xxxxxxxx> wrote:
> I wouldn't call it that odd.  RH patches their kernels to a
> fair extent, both for stability and features.

Yep.  They are _very_ well trusted.  Now if they'd put that
into XFS too, I'd be a happy camper.

> > > mke2fs -b 4096 -j -m 0 -R stride=1024 -T largefile4
> > > /dev/md0

BTW, aren't you worried about running out of inodes?

At the same time, have you benchmarked how much faster a full
fsck takes using 1 inode per 4MiB versus the standard
16-64KiB?

That would be an interesting test IMHO.

> Err, it's not a kludge and it's not a trick.  Those 2
> "disks" are hardware RAID5 arrays from 2 12 port 3ware 9500
> cards.  I like 3ware's hardware RAID, and those are the
> biggest (in terms of ports) cards 3ware makes.  
> So, I hook 12 disks up to each card, and the OS sees those
> as 2 SCSI disks.  I then do the software RAID to get 1)
> speed and 2) one partition to present to the users.  Folks
> (myself included) have been doing this for years.

I am in total agreeance with you, with one exception.  I
always make 2 volumes (one System, one Data) per card (yes,
I'm aware of the 9.2 firmware bug, hence why I have avoided
the 9500S largely, although 9.2.1.1 seems promising now that
it's officially released).  So in my case, I'd have two
RAID-0 stripes.

BTW, supposedly 3Ware supports volumes across up to 4 cards. 
Have you tried this?  I have not myself.

> The one gotcha in this setup (other than not being able to
> boot from the big RAID5 arrays, since each is >2TiB)

Another reason to create a "System" volume and a "Data"
volume.  My "System" volume is typically 2/4 drives in
RAID-1/10.  My "Data" volume is typically RAID-5, or if I
really need performance, RAID-10.

> is that the version of mdadm shipped with RHEL4 does not
> support array members bigger than 2TiB.  I had 
> to upgrade to an upstream release to get that support.

Which is why I use LVM (and now LVM2) for RAID-0.  I know
there are claims it is slower than MD (at least LVM2), but I
just like the management of LVM.  I guess I'm typical of a
commercial UNIX wennie.


Chris Mauritz <chrism@xxxxxxxxx> wrote:
> For what it's worth, I have also done RAID0 stripes of 2
> raid arrays to get *really* fast read/write performance
> when used for storing  uncompressed video.  Recently, when
> I was at Apple for a meeting, that was their engineer's
> preferred method for getting huge RAIDs....running
> software RAID volumes across multiple Xserve-RAID devices

Software RAID-0 at the OS level (and not some FRAID driver)
is _always_ going to be the _ultimate_ because you can span
peripheral interconnects and cards.

> Perhaps I'm just extremely lucky, but I've not run into
> this magic 1TB barrier that I see bandied about here.

As I said, I've just ran into it on kernel 2.4 distributions.
  
Any filesystem that grows up on a POSIX32 implementation
(especially pre-kernel 2.4 / GLibC 2.2 before LFS was
standard) is going to have signed 32-bit int structures.

I'm sure Tweedie and the gang have gotten around all of them
in kernel 2.6 now.  But at the same time, I don't trust how
they are doing.

> Unfortunately, a lot of the documentation and FAQs are
> quite out of date which can lead to some confusion.

Yeah.  LVM2 and Device Mapper (DM) are a real PITA if you
start playing with newer developments, and race conditions
seem to be never-ending.

But when it comes to using intelligent 3Ware RAID with just
LVM2 for RAID-0, it has worked flawlessly for myself on
kernel 2.6.



-- 
Bryan J. Smith                | Sent from Yahoo Mail
mailto:b.j.smith@xxxxxxxx     |  (please excuse any
http://thebs413.blogspot.com/ |   missing headers)

[Index of Archives]     [CentOS]     [CentOS Announce]     [CentOS Development]     [CentOS ARM Devel]     [CentOS Docs]     [CentOS Virtualization]     [Carrier Grade Linux]     [Linux Media]     [Asterisk]     [DCCP]     [Netdev]     [Xorg]     [Linux USB]
  Powered by Linux