Hi yall, although I've worked with databases for more than 7 years now, I'm petty new to PostgreSQL. I have an application using SQLite3 as an embedded SQL solution because it's simple and it can handle the load that *most* of my clients have. Because of that '*most*' part, because of the client/server way and because of the license, I'm think about start using PostgreSQL. My app uses only three tables: one has low read and really high write rates, a second has high read and low write and the third one is equally high on both. I need a db that can handle something like 500 operations/sec continuously. It's something like 250 writes/sec and 250 reads/sec. My databases uses indexes. Each table would have to handle 5 million rows/day. So I'm thinking about creating different tables (clusters?) to different days to make queries return faster. Am I right or there is no problem in having a 150 million (one month) rows on a table? All my data is e-mail traffic: user's quarentine, inbond traffic, outbond traffic, sender, recipients, subjects, attachments, etc... What do you people say, is it possible with PostgreSQL? What kind of hardware would I need to handle that kind of traffic? On a first test, at a badly tunned AMD Athlon XP 1800+ (ergh!) I could do 1400 writes/sec locally after I disabled fsync. We have UPSs, in the last year we only had 1 power failure. Thank you all for your tips. Best regards, Daniel Colchete