Christopher Browne wrote: >> Orlando Giovanny Solarte Delgado wrote: >>> It is a system web and each user can >>> to do out near 50 consultations for session. I can have simultaneously >>> around 100 users. Therefore I can have 5000 consultations >>> simultaneously. Each consultation goes join to a space component in >>> Postgis, therefore I need to store each consultation in PostgreSQL to >>> be able to use all the capacity of PostGIS. The question is if for >>> each consultation in execution time build a table in PostGRESQL I use >>> it and then I erase it. Is a system efficient this way? Is it possible >>> to have 5000 tables in PostGRESQL? How much performance? >> Use TEMP tables. > Hmm. To what degree do temp tables leave dead tuples lying around in > pg_class, pg_attribute, and such? > I expect that each one of these connections will leave a bunch of dead > tuples lying around in the system tables. The system tables will need > more vacuuming than if the data was placed in some set of > more-persistent tables... > None of this seems forcibly bad; you just need to be sure that you > vacuum the right things :-). Since there is pg_autovacuum you don't need to think about it. -- Wbr, Sergey Moiseev