On Mon, 25 Jan 2010, nair rajiv wrote:
I am working on a project that will take out structured content from
wikipedia and put it in our database...
there is a table which will approximately have 5 crore entries after data
harvesting.
Have you asked the Wikimedia Foundation if they mind you consuming that
much of their bandwidth, or even if there are copyright issues involved in
grabbing that much of their data?
(The other problem with using the word "crore" is that although it may
mean 10000000 in a few countries, it could also mean 500000.)
Matthew
--
Of course it's your fault. Everything here's your fault - it says so in your
contract. - Quark
--
Sent via pgsql-performance mailing list (pgsql-performance@xxxxxxxxxxxxxx)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-performance