Have a good day. This friday i'm going to migrate an entire database of a government institution in my country. Those are like 4 database of 2GB each one. So, i was preparing about what i'm going to do tomorrow. They currently have PostgreSQL 9.1 installed from source i think, and they will update RHEL version to 5.7, it will be a reinstall of operative system, so before that i will backup the databases, copy contents of pg_hba.conf and so on, and after the reinstallation of RHEL i'm going to install PostgreSQL 9.3 from official repositories alongside contrib and plpython.
I've already tested that on a Centos 5.7 VM and it works fine. But well, yesterday i was practicing creation of backups and restoring backups in my laptop with the next specs:
Core i7 2.2 Ghz
8GB RAM
30GB empty space
Windows 8.1
In my laptop i also have PostgreSQL 9.3. First i used the sample database pgdellstore, what i did is that i created a database, and ran the SQL file of the database. After that, i created a backup of that database pgdellstore with pg_dump, a custom format, without compression:
pg_dump -h localhost -p 5432 -U postgres -W -Fc -Z0 -C -d pgdellstore > pgdellstore.backup
And everything as normal. After that i created a second database and restored my custom backup with pg_restore trying to speed up using -j option:
pg_restore -h localhost -p 5432 -U postgres -W -d pgdellstore -j4 pgdellstore.backup
Also, before that, i backed up my postgresql.conf and edited the current one with the next parameters, according to some research that i did in google about performance tuning of pg_restore:
shared_buffers=2GB
maintenance_work_mem=1GB
work_mem=128MB
wal_buffers=16MB
checkpoint_segments=8
autovacuum=off
archive_mode=off
fsync=off
full_page_writes=off
checkpoint_timeout=15min
checkpoint_completion_target=0.9
track_counts=off
synchronous_commit=off
bgwriter_delay=50ms
And it restored it in 2 seconds i think, for me it was extremely fast (is a database of 25MB aproximately) and i had doubts, so i have some queries to check database tables size and i compared the results of the first database (the one that i created with original sql script) and the second one (the one that i created using a custom backup of the first one) and i got this:
DB created with original SQL script:
schemaname | tablename | reltuples | tamanio | tamanioord
------------+--------------+-----------+------------+------------
public | customers | 20000 | 5016 kB | 5136384
public | orderlines | 60350 | 4440 kB | 4546560
public | cust_hist | 60350 | 3976 kB | 4071424
public | products | 10000 | 1552 kB | 1589248
public | orders | 12000 | 1384 kB | 1417216
public | inventory | 10000 | 704 kB | 720896
public | categories | 16 | 24 kB | 24576
public | afiliado | 4 | 24 kB | 24576
public | pruebafechas | 0 | 8192 bytes | 8192
public | reorder | 0 | 0 bytes | 0
(10 filas)
DB created with custom backup based on first db
schemaname | tablename | reltuples | tamanio | tamanioord
------------+--------------+-----------+------------+------------
public | customers | 20000 | 4992 kB | 5111808
public | orderlines | 60350 | 4416 kB | 4521984
public | cust_hist | 60350 | 3952 kB | 4046848
public | products | 10000 | 1528 kB | 1564672
public | orders | 12000 | 1360 kB | 1392640
public | inventory | 10000 | 680 kB | 696320
public | afiliado | 4 | 24 kB | 24576
public | categories | 16 | 24 kB | 24576
public | pruebafechas | 0 | 8192 bytes | 8192
public | reorder | 0 | 0 bytes | 0
(10 filas)
This is the query that i used to get those results:
SELECT tbl.schemaname, tbl.tablename, obj.reltuples
, pg_size_pretty(pg_total_relation_size(text('"' || tbl.schemaname || '"."' || tbl.tablename || '"'))) tamanio, pg_total_relation_size(text('"' || tbl.schemaname || '"."' || tbl.tablename || '"')) tamanioord
FROM pg_tables tbl, pg_class obj WHERE tbl.tablename = obj.relname
AND tbl.schemaname NOT IN ('pg_catalog', 'information_schema') ORDER BY tamanioord DESC;
As you can see, there are certain variations in size, the number of tuples remains, but the fields tamanio (formatted size) and tamanioord (size without format) have a difference, for example the table customers has 5016 kB versus 4992 kB . I'm afraid that in someway i lost data or something like that, or according to your experience that can be normal and doesn't mean that information is corrupted or something, because tomorrow when i will do that with productions databases it will be a chaos if information gets corrupted.
Also, my second concern is the next: 4 months ago, i did a routine maintenance of those production database, and first as usual i backup all the databases. I took one of the backups of a development database for testing purposes. The backup file weights 279MB and the database is like 1.8 GB . So, yesterday i also practiced restoring this backup with pg_restore in a new database, with the same configuration from above. The command used was this:
pg_restore -h localhost -p 5432 -U postgres -W --create -d debdbrest -j4 devdb.backup
And it took only 1:30 mins aproximately. For me is suspiciously fast becuse when i was reading about pg_restore and how to enhace performance during backup restoring, i found various cases in previous postgresql versions about pg_restore is too slow, and also, i tried the same restore of the same backups 2 days ago but without postgresql.conf parameters modified and without using -j flag, and it started and after 2 hours it hadn't finished, so i cancelled; also in the first try i noted that it printed a lot of output in the console and in this second try it didn't print anything.
But that time reduction from more than 2 hours to 1:30 mins is too suspicious for me, but i don't know if, as the previous question that i have, you consider that this is normal (a backup of 279MB of a database of 1840MB restored in just 1.30 min).
Regards.
***************************
Oscar Calderon
Analista de Sistemas
Soluciones Aplicativas S.A. de C.V.
www.solucionesaplicativas.com
Cel. (503) 7741 7850 Tel. (503) 2522-2834