Re: splitting data into multiple tables

From: nair rajiv <nair331(at)gmail(dot)com>
To: Matthew Wakeling <matthew(at)flymine(dot)org>, pgsql-performance(at)postgresql(dot)org
Subject: Re: splitting data into multiple tables
Date: 2010-01-26 15:17:09
Message-ID: d67ff5e61001260717h6b049a3ch529b9894fb8e123a@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

On Tue, Jan 26, 2010 at 5:15 PM, Matthew Wakeling <matthew(at)flymine(dot)org>wrote:

> On Mon, 25 Jan 2010, nair rajiv wrote:
>
>> I am working on a project that will take out structured content from
>> wikipedia and put it in our database...
>>
>> there is a table which will approximately have 5 crore entries after data
>> harvesting.
>>
>
> Have you asked the Wikimedia Foundation if they mind you consuming that
> much of their bandwidth, or even if there are copyright issues involved in
> grabbing that much of their data?
>

We are downloading the nt and owl files kept for download at
http://wiki.dbpedia.org/Downloads34

> (The other problem with using the word "crore" is that although it may mean
> 10000000 in a few countries, it could also mean 500000.)
>
> Matthew
>
> --
> Of course it's your fault. Everything here's your fault - it says so in
> your
> contract. - Quark
>

In response to

Browse pgsql-performance by date

  From Date Subject
Next Message Mark Hills 2010-01-26 16:00:40 Poor query plan across OR operator
Previous Message Matthew Wakeling 2010-01-26 11:45:41 Re: splitting data into multiple tables