=?UTF-8?q?Mart=C3=ADn_Fern=C3=A1ndez?= <fmartin91@xxxxxxxxx> writes: > We basically started a VACUUM on a given table, waited for one index to process (captured cleaned rows count) and cancel the VACUUM. When we run another VACUUM on the same table the dead rows removed from the first index was a number slightly higher than the value logged on the first VACUUM. This behaviour made us feel that the work done to clean dead tuples on the first index was performed again. The unit of work that doesn't have to be repeated if VACUUM is canceled is: 1. Scan a bunch of heap pages to identify dead tuples; 2. Scan *all* the table's indexes to remove the corresponding index entries; 3. Rescan those heap pages to actually remove the tuples. It sounds like you canceled partway through phase 2. The actual size of this unit of work is the number of dead-tuple TIDs that will fit in maintenance_work_mem (at six or eight bytes apiece, I forget whether it's aligned...). Normally, people make maintenance_work_mem big so that they can reduce the number of index scan cycles needed to complete vacuuming a table. But if you're concerned about reducing the amount of work lost to a cancel, you might try *reducing* maintenance_work_mem. This will make vacuum slower overall (more index scans), but you have a better chance that it will manage to actually remove some tuples before getting canceled. Or you could look at fixing the access patterns that are causing so many autovacuum cancels. regards, tom lane