Re: pg_dump additional options for performance

From: Simon Riggs <simon(at)2ndquadrant(dot)com>
To: Alvaro Herrera <alvherre(at)commandprompt(dot)com>
Cc: Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>, "Joshua D(dot) Drake" <jd(at)commandprompt(dot)com>, Jeff Davis <pgsql(at)j-davis(dot)com>, pgsql-hackers(at)postgresql(dot)org
Subject: Re: pg_dump additional options for performance
Date: 2008-02-26 19:50:35
Message-ID: 1204055435.4252.430.camel@ebony.site
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-hackers

On Tue, 2008-02-26 at 16:18 -0300, Alvaro Herrera wrote:
> Tom Lane wrote:
> > "Joshua D. Drake" <jd(at)commandprompt(dot)com> writes:
> > > IMO the place to start is COPY which is per my tests, slow. Multi
> > > worker connection restore is great and I have proven that with some
> > > work it can provide o.k. results but it is certainly not acceptable.
> >
> > It was already pointed out to you that we can hope for only incremental
> > speedups in COPY per se. Don't be too quick to dismiss the discussion
> > of large-grain parallelism, because I don't see anything else within
> > reach that might give integer multiples rather than percentage points.
>
> Well, one idea would be dividing the input file in similarly-sized parts
> and giving each one to a different COPY process. This would help in
> cases where you have a single very large table to restore.
>
> Another thing we could do is selective binary output/input for bytea
> columns, to avoid the escaping step.

This is exactly what Dimitri is working on.

--
Simon Riggs
2ndQuadrant http://www.2ndQuadrant.com

In response to

Browse pgsql-hackers by date

  From Date Subject
Next Message Simon Riggs 2008-02-26 19:55:29 Re: pg_dump additional options for performance
Previous Message Joshua D. Drake 2008-02-26 19:47:55 Re: pg_dump additional options for performance