Skip site navigation (1) Skip section navigation (2)

Re: how to handle a big table for data log

From: Greg Spiegelberg <gspiegelberg(at)gmail(dot)com>
To: kuopo <spkuo(at)cs(dot)nctu(dot)edu(dot)tw>
Cc: Jorge Montero <jorge_montero(at)homedecorators(dot)com>, pgsql-performance(at)postgresql(dot)org
Subject: Re: how to handle a big table for data log
Date: 2010-07-27 20:02:08
Message-ID: (view raw, whole thread or download thread mbox)
Lists: pgsql-performance
On Tue, Jul 20, 2010 at 9:51 PM, kuopo <spkuo(at)cs(dot)nctu(dot)edu(dot)tw> wrote:

> Let me make my problem clearer. Here is a requirement to log data from a
> set of objects consistently. For example, the object maybe a mobile phone
> and it will report its location every 30s. To record its historical trace, I
> create a table like
> *CREATE TABLE log_table
> (
>   id integer NOT NULL,
>  data_type integer NOT NULL,
>  data_value double precision,
>  ts timestamp with time zone NOT NULL,
>  CONSTRAINT log_table_pkey PRIMARY KEY (id, data_type, ts)
> )*;
> In my location log example, the field data_type could be longitude or
> latitude.
I witnessed GridSQL in action many moons ago that managed a massive database
log table.  From memory, the configuration was 4 database servers with a
cumulative 500M+ records and queries were running under 5ms.  May be worth a


In response to

pgsql-performance by date

Next:From: Yeb HavingaDate: 2010-07-27 20:06:16
Subject: Re: Slow query using the Cube contrib module.
Previous:From: Kevin GrittnerDate: 2010-07-27 18:32:18
Subject: Re: Linux Filesystems again - Ubuntu this time

Privacy Policy | About PostgreSQL
Copyright © 1996-2017 The PostgreSQL Global Development Group