On Mon, Feb 9, 2009 at 2:17 AM, John R Pierce <pierce@xxxxxxxxxxxx> wrote: > Matt Magoffin wrote: >> >> We have 100+ postgres processes running, so for an individual process, >> could the 1024 file limit be doing anything to this query? Or would I see >> an explicit error message regarding this condition? >> >> > > with 100 concurrent postgres connections, if they all did something > requiring large amounts of work_mem, you could allocate 100 * 125MB (I > believe thats what you said it was set to?) which is like 12GB :-O > > in fact a single query thats doing multiple sorts of large datasets for a > messy join (or other similar activity) can involve several instances of > workmem. multiply that by 100 queries, and ouch. > > have you considered using a connection pool to reduce the postgres process > count? No matter what I am pretty conservative with work_mem for these reasons. Plus, I tested most of our queries and raising work_mem above 16Meg had no real positive effect on most queries. If I have a single reporting query that can use work_mem over that I set it and run that query by itself (from things like cron jobs) rather than just leaving work_mem really high. High work_mem is a bit of a foot gun. -- Sent via pgsql-general mailing list (pgsql-general@xxxxxxxxxxxxxx) To make changes to your subscription: http://www.postgresql.org/mailpref/pgsql-general