On Tue, Aug 17, 2010 at 8:21 PM, Jose Ildefonso Camargo Tolosa
> I'm analyzing the possibility of using PostgreSQL to store a huge
> amount of data (around 1000M records, or so....), and these, even
> though are short (each record just have a timestamp, and a string that
> is less than 128 characters in length), the strings will be matched
> against POSIX Regular Expressions (different regexps, and maybe
> Because I don't have a system large enough to test this here, I have
> to ask you (I may borrow a medium-size server, but it would take a
> week or more, so I decided to ask here first). How is the performance
> of Regexp matching in PostgreSQL? Can it use indexes? My guess is:
> no, because I don't see a way of generally indexing to match regexp :(
> , so, tablescans for this huge dataset.....
> What do you think of this?
Yes it can index such things, but it has to index them in a fixed way.
i.e. you can create functional indexes with pre-built regexes. But
for ones where the values change each time, you're correct, no indexes
will be used.
Could full text searching be used instead?
In response to
pgsql-sql by date
|Next:||From: Jose Ildefonso Camargo Tolosa||Date: 2010-08-18 02:30:01|
|Subject: Re: Question about POSIX Regular Expressions performance on large dataset.|
|Previous:||From: Jose Ildefonso Camargo Tolosa||Date: 2010-08-18 02:21:25|
|Subject: Question about POSIX Regular Expressions performance on large dataset.|