Optimizing data layout for reporting in postgres

From: Doug El <doug_el(at)yahoo(dot)com>
To: pgsql-general(at)postgresql(dot)org
Subject: Optimizing data layout for reporting in postgres
Date: 2009-12-24 20:01:27
Message-ID: 245275.51880.qm@web114001.mail.gq1.yahoo.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

Hi,

I have to summarize some data to be queried and there are about 5 million raw records a day I need to summarize. In a nutshell I don't think I'm laying it out in an optimal fashion, or not taking advantage of some postgres features perhaps, I'm looking for feedback.

The raw incoming data is in the form of

ip string uint uint uint uint

So for any given record say:

8.8.8.8 helloworld 1 2 3 4

First, I need to be able to query how many total and how many unique requests there were (unique by ip), over given time frame.

So for the below data on the same day that's total two, but one unique

8.8.8.8 helloworld 1 2 3 4
8.8.8.8 helloworld 1 2 3 4

Further for all fields (but ip which is not stored) I need to be able to query and get total/unique counts based off any combination of criteria.

So if I refer to them as columns A-E

A B C D E
string uint uint uint uint

I need to be able and say how many where col A = 'helloworld' and say col C = 4.
Or perhaps col E = 4 and col c < 3 etc, any combination.

The only way I could see to do this was to take the 5 million daily raw records, sort them, then summarize that list with total and unique counts as so:

A B C D E F G H
date stringid uint uint uint uint total unique

Primary key is A-F (date stringid uint uint uint uint)

This gives me a summary of about 900k records a day from the 4 million raw.

I have things organized with monthly tables and yearly schemas. The string column also has its own monthly lookup table, so there's just a string id that's looked up.

The database however is still quite huge and grows very fast, even simple daily queries are fairly slow even on a fast server. I have a few indexes on what I know are common columns queried against but again, any combination of data can be queried, and indexes do increase db size of course.

I feel like there's got to be some better way to organize this data and make it searchable. Overall speed is more important than disk space usage for this application.

Perhaps there are some native features in postgres I'm not taking advantage of here, that would tip the scales in my favor. I've done a fair amount of research on the configuration file settings and feel like I have a fairly optimized config for it as far as that goes, and have done the things mentioned here: http://wiki.postgresql.org/wiki/SlowQueryQuestions

Very much appreciate any suggestions, thank you in advance.

Doug

Responses

Browse pgsql-general by date

  From Date Subject
Next Message Alexander Solovkin 2009-12-24 20:14:38 Installation problem
Previous Message Mark Morgan Lloyd 2009-12-24 19:20:35 Updating from 8.2 to 8.4