Hi Jeff,
On Tue, 19 Dec 2006, Arnau wrote:
I've got a DB in production that is bigger than 2GB that dumping it
takes more than 12 hours. I have a new server to replace this old one
where I have restore the DB's dump. The problem is I can't afford to
have the server out of business for so long, so I need your advice about
how you'd do this dump/restore. The big amount of data is placed in two
tables (statistics data), so I was thinking in dump/restore all except
this two tables and once the server is running again I'd dump/restore
this data. The problem is I don't know how exactly do this.
Arnau,
2GB and it takes 12 hours? What sort of server is this running on?
Does your postgresql.conf have all default values perhaps? I routinely
dump DBs that are 4-8GB in size and it takes about 10-15 minutes.
It's a dual Xeon with 4 GB of ram and with a RAID 5. Probably it has the
default values. Any suggestion about what parameters I should change to
speed it up?
--
Arnau