Manfred Koizar <mkoi-pg@aon.at> writes: > The first step, however, (acquire_sample_rows() in analyze.c) has to > read more rows than finally end up in the sample. It visits less than > O(nblocks) pages but certainly more than O(1). > A vague feeling tries to tell me that the number of page reads is > somehow related to the harmonic numbers 1 + 1/2 + 1/3 + ... + 1/n, which > grow like O(ln(n)). Good guess. Vitter's paper says the expected time to sample n rows from a table of size N is O(n * (1 + log(N/n))). > I have an idea how this could be done with O(1) page reads. The hard part is getting a genuinely random sample when we don't know N in advance. We do however know the table size in blocks, so if you're willing to make assumptions about constant tuple density you could do something different. (But the tuple density assumption is exactly the weak spot of what we've got, so I'm unconvinced that would be a big step forward.) regards, tom lane ---------------------------(end of broadcast)--------------------------- TIP 9: the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match