Yeb Havinga wrote:
Michael Stone wrote:
On Mon, Jul 26, 2010 at 03:23:20PM -0600, Greg Spiegelberg wrote:
I know I'm talking development now but is there a case for a pg_xlog
block
device to remove the file system overhead and guaranteeing your data is
written sequentially every time?
If you dedicate a partition to xlog, you already get that in practice
with no extra devlopment.
Due to the LBA remapping of the SSD, I'm not sure of putting files
that are sequentially written in a different partition (together with
e.g. tables) would make a difference: in the end the SSD will have a
set new blocks in it's buffer and somehow arrange them into sets of
128KB of 256KB writes for the flash chips. See also
http://www.anandtech.com/show/2899/2
But I ran out of ideas to test, so I'm going to test it anyway.
Same machine config as mentioned before, with data and xlog on separate
partitions, ext3 with barrier off (save on this SSD).
pgbench -c 10 -M prepared -T 3600 -l test
starting vacuum...end.
transaction type: TPC-B (sort of)
scaling factor: 300
query mode: prepared
number of clients: 10
duration: 3600 s
number of transactions actually processed: 10856359
tps = 3015.560252 (including connections establishing)
tps = 3015.575739 (excluding connections establishing)
This is about 25% faster than data and xlog combined on the same filesystem.
Below is output from iostat -xk 1 -p /dev/sda, which shows each second
per partition statistics.
sda2 is data, sda3 is xlog In the third second a checkpoint seems to start.
avg-cpu: %user %nice %system %iowait %steal %idle
63.50 0.00 30.50 2.50 0.00 3.50
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s
avgrq-sz avgqu-sz await svctm %util
sda 0.00 6518.00 36.00 2211.00 148.00 35524.00
31.75 0.28 0.12 0.11 25.00
sda1 0.00 2.00 0.00 5.00 0.00 636.00
254.40 0.03 6.00 2.00 1.00
sda2 0.00 218.00 36.00 40.00 148.00 1032.00
31.05 0.00 0.00 0.00 0.00
sda3 0.00 6298.00 0.00 2166.00 0.00 33856.00
31.26 0.25 0.12 0.12 25.00
avg-cpu: %user %nice %system %iowait %steal %idle
60.50 0.00 37.50 0.50 0.00 1.50
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s
avgrq-sz avgqu-sz await svctm %util
sda 0.00 6514.00 33.00 2283.00 140.00 35188.00
30.51 0.32 0.14 0.13 29.00
sda1 0.00 0.00 0.00 3.00 0.00 12.00
8.00 0.00 0.00 0.00 0.00
sda2 0.00 0.00 33.00 2.00 140.00 8.00
8.46 0.03 0.86 0.29 1.00
sda3 0.00 6514.00 0.00 2278.00 0.00 35168.00
30.88 0.29 0.13 0.13 29.00
avg-cpu: %user %nice %system %iowait %steal %idle
33.00 0.00 34.00 18.00 0.00 15.00
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s
avgrq-sz avgqu-sz await svctm %util
sda 0.00 3782.00 7.00 7235.00 28.00 44068.00
12.18 69.52 9.46 0.09 62.00
sda1 0.00 0.00 0.00 1.00 0.00 4.00
8.00 0.00 0.00 0.00 0.00
sda2 0.00 322.00 7.00 6018.00 28.00 25360.00
8.43 69.22 11.33 0.08 47.00
sda3 0.00 3460.00 0.00 1222.00 0.00 18728.00
30.65 0.30 0.25 0.25 30.00
avg-cpu: %user %nice %system %iowait %steal %idle
9.00 0.00 36.00 22.50 0.00 32.50
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s
avgrq-sz avgqu-sz await svctm %util
sda 0.00 1079.00 3.00 11110.00 12.00 49060.00
8.83 120.64 10.95 0.08 86.00
sda1 0.00 2.00 0.00 2.00 0.00 320.00
320.00 0.12 60.00 35.00 7.00
sda2 0.00 30.00 3.00 10739.00 12.00 43076.00
8.02 120.49 11.30 0.08 83.00
sda3 0.00 1047.00 0.00 363.00 0.00 5640.00
31.07 0.03 0.08 0.08 3.00
avg-cpu: %user %nice %system %iowait %steal %idle
62.00 0.00 31.00 2.00 0.00 5.00
Device: rrqm/s wrqm/s r/s w/s rkB/s wkB/s
avgrq-sz avgqu-sz await svctm %util
sda 0.00 6267.00 51.00 2493.00 208.00 35040.00
27.71 1.80 0.71 0.12 31.00
sda1 0.00 0.00 0.00 3.00 0.00 12.00
8.00 0.00 0.00 0.00 0.00
sda2 0.00 123.00 51.00 344.00 208.00 1868.00
10.51 1.50 3.80 0.10 4.00
sda3 0.00 6144.00 0.00 2146.00 0.00 33160.00
30.90 0.30 0.14 0.14 30.00
--
Sent via pgsql-performance mailing list (pgsql-performance@xxxxxxxxxxxxxx)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-performance