On 01/11/2012 02:54, Curtis J Blank wrote:
On 10/31/12 15:04, David Brown wrote:
On 31/10/12 18:34, Curtis J Blank wrote:
On 10/31/12 03:32, David Brown wrote:
I was planning, all the partitions i.e. mount points will be below 50%
used, most way below that and I don't see them filling up. That is on
purpose, theses SSD's are for the OS to gain performance and not a lot
of data storage with the exception of mysql.
So, if I have unused space at the end of the SSD, say 60G out of the
256G don't use it, don't partition it the SSD will use it for what ever?
It will know that it can use it when in a RAID1 set? Or make the raidset
only using cylinders to 196G and partition that leaving the rest unused?
If you want to leave extra space to improve the over-provisioning (it is
typically not necessary with more high-end SSDs, but you might want to
do it anyway), then it is important that the extra space is never
written. The easiest way to ensure that is to leave extra space during
partitioning. But be careful with raid - you have to use the
partition(s) for your raid devices, not the disk, or else you will write
to the entire SSD during the initial raid1 sync.
A typical arrangement would be to make a 1 GB partition at the start of
each SSD, then perhaps a 4 GB partition, then a big partition of about
200 GB in this case. Make a raid1 with metadata 1.0 from the first
partition of each disk for /boot, to make life easier for the
bootloader. Use the second partition of each disk for swap (no need for
raid here unless you are really concerned about uptime in the face of
disk failure and you actually expect to use swap significantly - in
which case go for raid1 or raid10 if you have more than 2 disks). Use
the third partition for your main raid (such as raid1, or perhaps
something else if you have more than two disks).
David, first off I want to say thanks for all the advice and your time.
This was what I was looking for to make informed decisions and I see I
came to the right place.
No problem. I learn a lot by making suggestions her, and having other
people correct me! So if my advice had been badly wrong, I expect
someone else would have said by now.
Yep, that's the way I do it, partition the disk then use the partitions
in the raid, not the whole disk. Although I do make more partitions and
more mount points only so that one thing can't use up all the space and
break other things. But still any one won't be over 50% utilization.
If you make your big raid1 pair an LVM physical volume, you can split it
into logical volumes as and when you want, and re-size them whenever
necessary. Note, however, that the unpartitioned space within the LVM
physical volume is still "used" as far as the SSD is concerned, since
the initial raid1 synchronisation has written to it. So only space
outside the raid1 partition acts as extra over-provisioning. (Not that
you will need much extra, if any.)
Of course, you can always start with a 50% size partition for your raid1
pair, leaving (almost) 50% of the SSD completely unused. And if you
want more space, you can just add another partition of say 30% on each
disk, match them up as a raid1 pair, put a new LVM physical volume onto
it, then add that physical volume to the volume group. You end up with
the same data in the same place, with only a tiny overhead for the LVM
indirection.
Once no-sync tracking is in place for md raid, it will be easier, as
there is no initial sync for raid1 (everything is marked no-sync). In
that case, space that is not partitioned within the LVM physical volume
will not be written to at all, and will therefore act as extra
over-provisioning until you actually need it.
If your SSDs do transparent compression, then another trick is to write
blocks of zero to unused space (you can do this across the whole disk
before partitioning). Blocks of zero compress rather well, so take tiny
amounts of physical space on the disk - and the freed space is then
extra recyclable blocks.
Oh and I do raid swap, not because it's used a lot, it's not, but to
raid everything else and leave a single point of failure kind of defeats
the purpose unless the goal is only to protect the data. Mine is that
and uptime.
That makes lots of sense.
I find swap useful even on machines with lots of ram - I put /tmp and
/var/tmp on tmpfs mounts, and sometimes use tmpfs mounts in other
places. tmpfs is always the fastest filesystem, as it has no overheads
for safety or to match sector layouts on disk. And with plenty of swap,
you don't have to worry about the space it takes - anything beyond
memory will automatically spill out to disk (making it slower, but still
faster than putting those same files on a disk filesystem).
<snip>
Put the DB's on the SSD.
As with all database applications, if you can get enough memory to have
most work done without reading from disks, it will go faster.
With decent SSD's (and since you have quite big ones, I assume they are
good quality), there is no harm in writing lots. You can probably write
at 30 MB/s continuously for years before causing any wearout on the disk.
Memory is currently at 16G, when I get around to it which won't be in
the too distant future it will be 32G. I'm fully aware and try to have
everything running in memory
The SSD's are OCZ Vertex 4 VTX4-25SAT3-256G. I hope they're good ones.
I'm trying to get their PEC just because I want to know. I'm also going
to try and get the over provisioned number, again just so I know.
I still haven't decided whether to connect the SSD's to the motherboard
which is SATA III and use Linux raid or connect them to my Areca 1882i
battery backed up caching raid controller which is also SATA III. Kind
of hinges on whether or not the controller passes discard. It's their
second generation card PCIe 2.0 not the new third generation PCIe 3.0
card. Trying to find that out too.
One thing to be very careful about with raid cards is that they can add
a lot of latency to SSDs. You can end up dropping your IOPs by a factor
of 20 or more. So check if the card works well with SSDs before using it.
For two disks, I'd connect them directly to the motherboard SATA (and
use an external UPS). But that depends on how much you value the
battery on the raid card, and how likely you see the risk of a system
crash (there is slightly lower chance of data loss via a raid card with
battery cache in such circumstances).
Like to hear your thoughts on this. My thinking is the performance would
really scream on the 1882i. And it just dawned on me if I use the
motherboard I might not be able to use the noop scheduler which is what
I currently use with my ARC-1220 because it has all the disks.
I would be very surprised if it ran faster on the raid card than
connected directly to the motherboard SATA. Raid cards can, sometimes,
give you higher speeds for raid5/6 compared to direct connections. In
particular, they help if you have a large number of disks (though with
the latest md raid multithreading for raid5/6, that will probably
change). But generally speaking, a raid card is not for speed -
especially not for SSDs where the extra layer will add noticeable
latency. Your CPU, motherboard and memory are more than capable of
saturating two fast SSDs - how could a raid card go any faster?
Ok but what about making a change to a page in a block whose other pages
are valid? The whole block gets moved then the old block is later
erased? That's what I'm understanding which sounds ok.
No, the changed page will get re-mapped to a different page somewhere
else - the unchanged data will remain where it was. That data will only
get moved if it makes sense for "defragmenting" to free up erase blocks,
or as part of wear-levelling routines.
Got it.
I think I was over thinking this. If a page changes the only way to do
that is read-modify-write of the block to where ever. So it might as
well be to an already erased block. I was getting hung up on having
erased pages in the blocks that can be immediately and just written.
Period. But that only occurs when appending data to a file. Let the
filesystem and SSD's do there thing...
I'm really thinking I don't need TRIM now. And when it is finally in the
kernel I can maybe try it. I was worried that if I don't do it from the
start it be too late later after the SSD's had been used for a while to
get the full benefit of it.
I think what you really want to use is "fstrim" - this walks through a
filesystem metadata, identifies free blocks, and sends TRIM commands for
each of them. Obviously this can take a bit of time, and will slow down
the disks while working, but you typically do it with a cron job in the
middle of the night.
<http://www.vdmeulen.net/cgi-bin/man/man2html?fstrim+8>
Yep, this sounds like the ticket. I was aware of it but didn't pursue it.
I haven't tried fstrim myself. Some day I must upgrade my ageing Fedora
14 system so that I can play with these new toys instead of just reading
about them...
I don't think the patches for passing TRIM through the md layer have yet
made it to mainstream distro kernels, but once they do you can run
fstrim.
Neil Brown told me probably 3.7, so we'll see I guess. It's becoming
less important to me though, but maybe nice when they do. I haven't
totally ruled out building a kernel with the patches but leaning towards
not doing it.
Incidentally, have a look at the figures in this:
<https://patrick-nagel.net/blog/archives/337>
A sample size of 1 web page is not great statistically evidence, but the
difference in the times for "sync" are quite large...
That says pretty much what I learned so far and the numbers are
interesting. Sort of says not to use trim real time continuously.
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html