Re: multi terabyte fulltext searching

From: Teodor Sigaev <teodor(at)sigaev(dot)ru>
To: Benjamin Arai <benjamin(at)araisoft(dot)com>
Cc: Oleg Bartunov <oleg(at)sai(dot)msu(dot)su>, Postgresql <pgsql-general(at)postgresql(dot)org>
Subject: Re: multi terabyte fulltext searching
Date: 2007-03-21 16:16:00
Message-ID: 46015A40.7090606@sigaev.ru
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

> I am currently using GIST indexes because I receive about 10GB of new
> data a week (then again I am not deleting any information). The do not
> expect to be able to stop receiving text for about 5 years, so the data
> is not going to become static any time soon. The reason I am concerned
> with performance is that I am providing a search system for several
> newspapers since essentially the beginning of time. Many bibliographer
> etc would like to use this utility but if each search takes too long I
> am not going to be able to support many concurrent users.

Use GiST and GIN indexes together: any data older than one month (which doesn't
change) with GIN index and new data with GiST. And one time per month moves data
from GiST to GIN.

--
Teodor Sigaev E-mail: teodor(at)sigaev(dot)ru
WWW: http://www.sigaev.ru/

In response to

Browse pgsql-general by date

  From Date Subject
Next Message Benjamin Arai 2007-03-21 16:16:26 Re: multi terabyte fulltext searching
Previous Message Benjamin Arai 2007-03-21 16:15:11 Re: multi terabyte fulltext searching