Re: Getting even more insert performance (250m+rows/day)

From: Mark Lewis <mark(dot)lewis(at)mir3(dot)com>
To: "Daniel J(dot) Luke" <dluke(at)geeklair(dot)net>
Cc: Dave Dutcher <dave(at)tridecap(dot)com>, pgsql-performance(at)postgresql(dot)org
Subject: Re: Getting even more insert performance (250m+rows/day)
Date: 2006-05-24 20:18:32
Message-ID: 1148501913.28095.70.camel@archimedes
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

> The data gets inserted in batches every 5 minutes and I potentially
> have people querying it constantly, so I can't remove and re-create
> the index.

How live does your data need to be? One possibility would be to use a
separate table for each batch instead of a separate table per day,
create the indexes after the import and only after the indexes have been
created make the table available for user queries.

You'd be trading latency for throughput in that case.

Also, you mentioned that you're CPU-bound, but that you have multiple
CPU's. In that case, performing N concurrent imports (where N is the
number of processor cores available) might be a win over a single-
threaded import.

-- Mark Lewis

In response to

Browse pgsql-performance by date

  From Date Subject
Next Message Daniel J. Luke 2006-05-24 20:20:20 Re: Getting even more insert performance (250m+rows/day)
Previous Message Daniel J. Luke 2006-05-24 20:18:06 Re: Getting even more insert performance (250m+rows/day)