Re: Optimal database table optimization method

From: Adrian von Bidder <avbidder(at)fortytwo(dot)ch>
To: pgsql-general(at)postgresql(dot)org
Subject: Re: Optimal database table optimization method
Date: 2010-03-04 20:57:25
Message-ID: 201003042157.33401@fortytwo.ch
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

Heyho!

Disclaimer: I don't have much experience with big databases, so this is
based on my understanding of the theory.

On Thursday 04 March 2010 20.32:46 Roger Tannous wrote:
> I have a database table that is growing too big (few hundred million
> rows) that needs to be optimized, but before I get into partitioning it,
> I thought I'd ask about suggestions.
>
> Here is the usage:
>
> 0 . Table contains about 10 columns of length about 20 bytes each.
>
> 1. INSERTS are performed at a rate of hundreds of times per second.
> 2. SELECT statements are performed based on column 'a' (where a='xxxx'
> 3. DELETE statements are performed based on a DATE column. (delete
> where date older than 1 year) usually once per day.

>
> The key requirement is to speed up INSERT and SELECT statements, and be
> able to keep history data of 1 year back without locking the whole table
> down while deleting.

How are your INSERTs coming in? In batches, from multiple sources?

> I would guess that I must have two indexes, one for column 'a', and the
> other for the date field. or is it possible to optimize both ?

Sounds reasonable, but see below.

> Will there be a necessary trade-off between speed on select and speed of
> delete?
>
> Is partitioning the only solution ?

How far did you already optimize your set up? WAL on a separate disk
(perhaps even solid state?), index and data separated?

> What are good strategies for
> partitioning such table?

With partitioning by date, you might be able to drop the index on the date
(if not needed for queries) because to just delete the oldest partition (and
creating a new one) each day the "delete" part doesn't involve any queries,
you just drop the table.

If you can make the INSERT logic smart so that it inserts into the partition
directly (instead of selecting the partition via trigger/rule), INSERT speed
presumably isn't slower than without partitioning.

I have no idea how good pg is with select queries over 365 partitions (1
year). OTOH: perhaps, with a date based partitioning, you can cache the
result of the select queries and add just the results of the newest data, so
the SELECT would onnly need to access the recent data? (Obviously depends
on the nature of these queries...)

cheers
-- vbi

--
Verbing weirds language.
-- Calvin & Hobbes

In response to

Responses

Browse pgsql-general by date

  From Date Subject
Next Message Lee Hachadoorian 2010-03-04 21:30:11 Connection timeouts from pgAdmin
Previous Message Niranjan Maturi (nmaturi) 2010-03-04 20:43:15 Re: Failed to run initdb: 128