Search Postgresql Archives

figuring out a streaming replication failure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The standby log:

->  2010-11-14 17:40:16 MST - 887 -LOG:  database system was shut down in recovery at 2010-11-14 17:40:10 MST
->  2010-11-14 17:40:16 MST - 887 -LOG:  entering standby mode
->  2010-11-14 17:40:16 MST - 887 -LOG:  consistent recovery state reached at 3/3988FF8
->  2010-11-14 17:40:16 MST - 887 -LOG:  redo starts at 3/3988F68
->  2010-11-14 17:40:16 MST - 887 -LOG:  invalid record length at 3/3988FF8
->  2010-11-14 17:40:16 MST - 885 -LOG:  database system is ready to accept read only connections
->  2010-11-14 17:40:16 MST - 890 -LOG:  streaming replication successfully connected to primary
->  2010-11-15 02:24:26 MST - 890 -FATAL:  could not receive data from WAL stream: FATAL:  requested WAL segment 000000010000000300000004 has already been removed
	
->  2010-11-15 02:24:26 MST - 887 -LOG:  unexpected pageaddr 2/B9BF2000 in log file 3, segment 4, offset 12525568
->  2010-11-15 02:24:27 MST - 2790 -LOG:  streaming replication successfully connected to primary
->  2010-11-15 02:24:27 MST - 2790 -FATAL:  could not receive data from WAL stream: FATAL:  requested WAL segment 000000010000000300000004 has already been removed
	
->  2010-11-15 02:24:32 MST - 2791 -LOG:  streaming replication successfully connected to primary
->  2010-11-15 02:24:32 MST - 2791 -FATAL:  could not receive data from WAL stream: FATAL:  requested WAL segment 000000010000000300000004 has already been removed

...

Now, the standby is geographically isolated from the master, so it's over an internet connection, so it's not a shock that with a large enough update and wal_keep_segments not large enough, speed of disk would outrun speed of network sufficiently for this to happen.

But as far as I know there was almost no write activity at 2am, no active users at all, no batch processing. There is a pg_dumpall that kicks off at 2am and these errors start about the same time that it finished. I also did the original synch and standby launch immediately after a mass update before autovacuum had a chance to run, so at some point there would be a lot of tuples marked dead.

wal_keep_segments was at 64, the first segment still around was 000000010000000300000010, checkpoint_segments was 16. In the midst of the pg_dumpall the master logs do show messages about checkpoint flushing too often. The 70ish log segments still around show mod times right around 2:23, progressing a second or so each, whereas they were created over a much longer period going back to the day before.

1 question: what happened here? Why were log files created the day before updated?

1 suggestion: would it be possible to not delete wal segments that are needed by a currently attached standby?

-- 
Scott Ribe
scott_ribe@xxxxxxxxxxxxxxxx
http://www.elevated-dev.com/
(303) 722-0567 voice





-- 
Sent via pgsql-general mailing list (pgsql-general@xxxxxxxxxxxxxx)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general



[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Postgresql Jobs]     [Postgresql Admin]     [Postgresql Performance]     [Linux Clusters]     [PHP Home]     [PHP on Windows]     [Kernel Newbies]     [PHP Classes]     [PHP Books]     [PHP Databases]     [Postgresql & PHP]     [Yosemite]
  Powered by Linux