On 26/02/07, Pallav Kalva <pkalva(at)livedatagroup(dot)com> wrote:
> I am in the process of cleaning up one of our big table, this table
> has 187 million records and we need to delete around 100 million of them.
> I am deleting around 4-5 million of them daily in order to catchup
> with vacuum and also with the archive logs space. So far I have deleted
> around 15million in past few days.
> max_fsm_pages value is set to 1200000. Vacuumdb runs once daily,
> here is the output from last night's vacuum job
> INFO: free space map: 999 relations, 798572 pages stored; 755424
> total pages needed
> DETAIL: Allocated FSM size: 1000 relations + 1200000 pages = 7096
> kB shared memory.
> From the output it says 755424 total pages needed , this number
> keeps growing daily even after vacuums are done daily. This was around
> 350K pages before the delete process started.
> I am afraid that this number will reach the max_fsm_pages limit
> soon and vacuums thereafter will never catch up .
> Can anyone please explain this behavior ? What should I do to catch
> up with vacuumdb daily ?
Vacuum adds to free pages to the fsm so that they can be reused. If
you don't fill up those free pages the fsm will fill up. Once the fsm
is full no more pages can be added to the fsm. If you start writing to
the free pages via inserts when vacuum next runs more free pages will
be added that did not fit previously in the free space map due to it
If you are really deleting that many records you may be better coping
those you want to a new table and dropping the old one. To actually
recover space you need to either run vacuum full or cluster.
This ought to be in the manual somewhere as this question gets asked
about once a week.
In response to
pgsql-performance by date
|Next:||From: Gene||Date: 2007-02-26 16:12:53|
|Subject: does prepareThreshold work? forced to use old driver|
|Previous:||From: Pallav Kalva||Date: 2007-02-26 14:44:02|
|Subject: Vacuumdb - Max_FSM_Pages Problem.|