Re: "Healing" a table after massive updates

From: "Scott Marlowe" <scott(dot)marlowe(at)gmail(dot)com>
To: "Bill Moran" <wmoran(at)collaborativefusion(dot)com>
Cc: "Alvaro Herrera" <alvherre(at)commandprompt(dot)com>, "Gauthier, Dave" <dave(dot)gauthier(at)intel(dot)com>, pgsql-general(at)postgresql(dot)org
Subject: Re: "Healing" a table after massive updates
Date: 2008-09-11 16:22:42
Message-ID: dcc563d10809110922i3ec99047p5f90262e3383b71a@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-general

On Thu, Sep 11, 2008 at 8:56 AM, Bill Moran
<wmoran(at)collaborativefusion(dot)com> wrote:
> In response to Alvaro Herrera <alvherre(at)commandprompt(dot)com>:
>
>> Bill Moran wrote:
>> > In response to "Gauthier, Dave" <dave(dot)gauthier(at)intel(dot)com>:
>> >
>> > > I might be able to answer my own question...
>> > >
>> > > vacuum FULL (analyze is optional)
>> >
>> > CLUSTER _may_ be a better choice, but carefully read the docs regarding
>> > it's drawbacks first. You may want to do some benchmarks to see if it's
>> > really needed before you commit to it as a scheduled operation.
>>
>> What drawbacks?
>
> There's the whole "there will be two copies of the table on-disk" thing
> that could be an issue if it's a large table.

I've also found cluster to be pretty slow, even on 8.3. On a server
that hits 30-40Megs a second write speed for random access during
pgbench, it's writing out at 1 to 2 megabytes a second when it runs,
and takes the better part of a day on our biggest table. vacuumdb -fz
+ reindexdb ran in about 6 hours which means we could fit it into our
maintenance window. vacuum moves a lot more data per second than
cluster.

In response to

Responses

Browse pgsql-general by date

  From Date Subject
Next Message johnf 2008-09-11 16:24:40 Re: keep alive losing connections
Previous Message Dave Page 2008-09-11 16:21:18 Re: Windows ODBC Driver