Is there any experience with Postgresql and really huge tables? I'm
talking about terabytes (plural) here in a single table. Obviously the
table will be partitioned, and probably spread among several different
file systems. Any other tricks I should know about?
We have a problem of that form here. When I asked why postgres wasn't
being used, the opinion that postgres would "just <explicitive> die" was
given. Personally, I'd bet money postgres could handle the problem (and
better than the ad-hoc solution we're currently using). But I'd like a
couple of replies of the form "yeah, we do that here- no problem" to
wave around.
I've done a project using 8.1 on solaris that had a table that was
closed to 2TB. The funny thing is that it just worked fine even without
partitioning.
But, then again: the size of a single record was huge too: ~ 50K.
So there were not insanly many records: "just" something
in the order of 10ths of millions.
The queries just were done on some int fields, so the index of the
whole thing fit into RAM.
A lot of data, but not a lot of records... I don't know if that's
valid. I guess the people at Greenplum and/or Sun have more exciting
stories ;)
Bye, Chris.