You could look at contrib/dblink but if you plan to join between your four
servers a lot it seems like you might want cleaner solution, like, say,
slony-I or pgpool, and maybe some data partitioning on that main table as
well.
On Wed, 21 Feb 2007, Michelle Konzack wrote:
Hello,
My current PostgreSQL Database has arround 560 GByte without indices.
If not more then 3-5 user query it it works fast like Google... :-)
But I have over 600 Users now which all query it over the Net and hit
a heavy performance problem.
The biggest problem is the "main" table which is arround 80% of the DB.
Now I have made a statistic and know, which Datas are most quried and
like to split the Database (exactly THE MONSTER-TABLE) over at least 4
Physicaly PostgreSQL servers.
The Database is generaly STATIC an Data will only be added and NEVER
removed or changed. (If DATE or a ROW change, it will get a FULL NEW
ROW with version number)
Since I access the PostgreSQL from a pysicaly seperated dedicated
apache server using php5 I like to know from some Experts how to
query 4 or more Databases ar once and join the data.
I thinking, putting the main database with most requested data on the
first Server which will have arround 180 GByte and then the rest of
"THE" table on the other servers.
Note: "THE" table is a historical timeline...
and the datasets are not ordered in the table.
Greetings
Michelle Konzack
Systemadministrator
Tamay Dogan Network
Debian GNU/Linux Consultant
--
Linux-User #280138 with the Linux Counter, http://counter.li.org/
##################### Debian GNU/Linux Consultant #####################
Michelle Konzack Apt. 917 ICQ #328449886
50, rue de Soultz MSM LinuxMichi
0033/6/61925193 67100 Strasbourg/France IRC #Debian (irc.icq.com)