Hello,
I have a huge table with 141456059 records on a PostgreSQL 10.18 database.
When I try to do a pg_dump on that table, postgresql gives a segfault, displaying this message:
2021-12-22 14:08:03.437 UTC [15267] LOG: server process (PID 25854) was terminated by signal 11: Segmentation fault
2021-12-22 14:08:03.437 UTC [15267] DETAIL: Failed process was running: COPY ********** TO stdout;
2021-12-22 14:08:03.437 UTC [15267] DETAIL: Failed process was running: COPY ********** TO stdout;
2021-12-22 14:08:03.437 UTC [15267] LOG: terminating any other active server processes
2021-12-22 14:08:03.438 UTC [15267] LOG: archiver process (PID 16034) exited with exit code 2
2021-12-22 14:08:04.196 UTC [15267] LOG: all server processes terminated; reinitializing
2021-12-22 14:08:05.785 UTC [25867] LOG: database system was interrupted while in recovery at log time 2021-12-22 14:02:29 UTC
2021-12-22 14:08:05.785 UTC [25867] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target.
2021-12-22 14:08:03.438 UTC [15267] LOG: archiver process (PID 16034) exited with exit code 2
2021-12-22 14:08:04.196 UTC [15267] LOG: all server processes terminated; reinitializing
2021-12-22 14:08:05.785 UTC [25867] LOG: database system was interrupted while in recovery at log time 2021-12-22 14:02:29 UTC
2021-12-22 14:08:05.785 UTC [25867] HINT: If this has occurred more than once some data might be corrupted and you might need to choose an earlier recovery target.
On the linux log I only see this:
Dec 22 14:08:03 kernel: postmaster[25854]: segfault at 14be000 ip 00007f828fabb5f9 sp 00007fffe43538b8 error 6 in libc-2.17.so[7f828f96d000+1c2000]
I'm guessing I'm hitting some (memory?) limit, is there anything I can do to prevent this? Shouldn't PostgreSQL have some different behavior instead of crashing the server?
--
Paulo Silva <paulojjs@xxxxxxxxx>