On 14 Listopad 2011, 22:58, Cody Caughlan wrote: >> Seems reasonable, although I'd bump up the checkpoint_timeout (the 5m is >> usually too low). > > Ok, will do. Yes, but find out what that means and think about the possible impact first. It usually improves the checkpoint behaviour but increases the recovery time and you may need more checkpoint segments. And I doubt this will fix the issue you've described. > I ran bonnie++ on a slave node, doing active streaming replication but > otherwise idle: > http://batch-files-test.s3.amazonaws.com/sql03.prod.html > > bonnie++ on the master node: > http://batch-files-test.s3.amazonaws.com/sql01.prod.html > > If I am reading this right, this is my first time using it, the > numbers dont look too good. Are those instances equal, i.e. use tha same RAID10 config etc.? It surprises me a bit that the slave performs much better than the master, for example the sequential reads are much faster (210MB/s vs. 60MB/s) and it handles about twice the number of seeks (345 vs. 170). But this may be skewed because of the workload. > Heres a gist of running "iostat -x 3" for about a few minutes: > > https://gist.github.com/f94d98f2ef498a522ac2 > > Indeed, the %iowat and await values can spike up drastically. OK, so xvdb-xvde are individual drives and dm-0 is the RAID10 device, right? According to the log_checkpoint info, you're writing about 15000 (120MB) buffers in 270s, i.e. about 440kB/s. But according to the iostat you're writing up to 4MB/s, so it's not just about the checkpoints. What else is going on there? How much WAL do you write? Do you have iotop installed? That might give you a hint what processes are writing data etc. I'm a bit confused that the w/s don't add up while r/s do. Tomas -- Sent via pgsql-performance mailing list (pgsql-performance@xxxxxxxxxxxxxx) To make changes to your subscription: http://www.postgresql.org/mailpref/pgsql-performance