> Orlando Giovanny Solarte Delgado wrote: >> I am designing a system that it takes information of several databases >> distributed in Interbase (RDBMS). It is a system web and each user can >> to do out near 50 consultations for session. I can have simultaneously >> around 100 users. Therefore I can have 5000 consultations >> simultaneously. Each consultation goes join to a space component in >> Postgis, therefore I need to store each consultation in PostgreSQL to >> be able to use all the capacity of PostGIS. The question is if for >> each consultation in execution time build a table in PostGRESQL I use >> it and then I erase it. Is a system efficient this way? Is it possible >> to have 5000 tables in PostGRESQL? How much performance? >> > Use TEMP tables. Hmm. To what degree do temp tables leave dead tuples lying around in pg_class, pg_attribute, and such? I expect that each one of these connections will leave a bunch of dead tuples lying around in the system tables. The system tables will need more vacuuming than if the data was placed in some set of more-persistent tables... None of this seems forcibly bad; you just need to be sure that you vacuum the right things :-). It is a big drag if system tables get filled with vast quantities of dead tuples; you can't do things like reindexing them without shutting down the postmaster. -- (reverse (concatenate 'string "moc.liamg" "@" "enworbbc")) http://linuxdatabases.info/info/x.html "Listen, strange women, lyin' in ponds, distributin' swords, is no basis for a system of government. Supreme executive power derives itself from a mandate from the masses, not from some farcical aquatic ceremony." -- Monty Python and the Holy Grail