On 01/06/2015 10:09 AM, Michael Heaney
wrote:
I'm
fairly new to Postgres, and have a design issue for which an array
of integers might be a good solution. But I'd like to hear from
the experts before proceeding down this path.
Essentially, I'm trying to model the relationship between a group
of biological samples and their genes. Each sample (there are ~10K
of them now, with more coming) will have about 30,000 genes.
Conversely, a particular gene may be present in almost all
samples.
So I've created the following table to handle the many-to-many
relationship:
create table sample_gene (id serial, sample_id int, gene_id int);
which looks like this when populated:
sample_id | gene_id
---------------------------------------
1 | 1
1 | 2
...
1 | 30475
2 | 1
2 | 2
...
2 | 29973
3 | 1
etc.
The table now contains hundreds of millions of rows (with many,
many more to come). Join performance between samples and genes is
quite slow, even with indexes on sample_id and gene_id.
So it occurred to me: why not eliminate all the duplicate
sample_id values by storing the gene_id's in an array, like so:
create table sample_gene_array (id serial, sample_id int, gene_id
int [] );
So now the table data looks like this:
sample_id | gene_id []
---------------------------------------
1 | [1:30475]
2 | [1:29973]
etc.
The new table is significantly smaller, and performance (using
ANY[] ) is quite good. Nevertheless, I'm uneasy. I come from a
Sybase ASE background, and so have no experience with arrays as
datatypes. Is it okay to store 30K+ gene values in an array in the
linking table (or maybe even in the sample table itself, thus
eliminating the linking table)? Should I unnest the gene_id's
first, before using them to join to the gene table?
TIA for any guidance you can provide. Again, I'm a Postgres
neophyte - but I'm in awe of the power and flexibility of this
database, and wish that I'd started using it sooner.
------
Michael Heaney
JCVI
Think
genotype calls (sample, marker, call) and this just explodes in
terms of number of rows but I'm trying to not let that bother
me. Partion by chromosome, or chromosome arm (even if partition
is just separate tables) and the magnitude becomes manageable.
At least as manageable as multi-gigabyte gvcf files per sample
from GATK.
But I'm a little confused: all samples for a given species will
have the same genes (roughly). Are you storing gene variants
(sequence or otherwise) per sample?
|