Sven Willenberger <sven@xxxxxxx> writes: > I have found the answer/problem. On a hunch I increased maxdsiz to 1.5G > in the loader.conf file and rebooted. I ran vacuumdb and watched top as > the process proceeded. What I saw was SIZE sitting at 603MB (which was > 512MB plus another 91MB which corresponded nicely to the value of RES > for the process. A bit into the process I saw SIZE jump to 1115 -- i.e. > another 512 MB of RAM was requested and this time allocated. At one > point SIZE dropped back to 603 and then back up to 1115. I suspect the > same type of issue was occuring in regular vacuum from the psql client > connecting to the backend, for some reason not as frequently. I am > gathering that maintenance work mem is either not being recognized as > having already been allocated and another malloc is made or the process > is thinking the memory was released and tried to grab a chunk of memory > again. Hmm. It's probably a fragmentation issue. VACUUM will allocate a maintenance work mem-sized chunk during command startup, but that's likely not all that gets allocated, and if any stuff allocated after it is not freed at the same time, the process size won't go back down. Which wouldn't be a killer in itself, but unless the next iteration is able to fit that array in the same space, you'd see the above behavior. BTW, do you have any evidence that it's actually useful to set maintenance work mem that high for VACUUM? A quick and dirty solution would be to bound the dead-tuples array size at something more sane... regards, tom lane ---------------------------(end of broadcast)--------------------------- TIP 7: don't forget to increase your free space map settings