I haven't had issues about the max_fsm_pages. The problem I had was caused because I was not committing (or rolling back) after JDBC select calls. I didn't think that was necessary because the select calls didnt modify database. But, once I modified my selects to rollback after the data is read from the result set, autovacuum worked fine. I was hoping that with autovacumming working well, I don't need the cron tasks etc to do the vacuum full periodically.. But I read a few days back that reindexdb is recommended if there is a lot of inserts/deletes. I have to experiment with that yet. I'll check what I have set for max_fsm_pages tomorrow. --- Francisco Reyes <lists@xxxxxxxxxxxxxxxx> wrote: > Matthew T. O'Connor writes: > > > In many instances the default thresholds for autovacuum are too > > conservative. You can try making it more aggressive and see what > > happens, you can do this generally, or on a table specific basis. > > I think something is very wrong with the estimates I am getting back. After > 3 times increasing the max_fsm_pages, every time I get the same error.. > about needing to increase max_fsm_pages.. and always the error is > recommending exactly the same number of additional fsm_pages. > > I see other threads with other people having simmilar problems. Going to see > if I can find a resolution in the archives. __________________________________________________ Do You Yahoo!? Tired of spam? Yahoo! Mail has the best spam protection around http://mail.yahoo.com