On Wednesday 22 May 2002 18:40, Ron Johnson wrote:
> On Wed, 2002-05-22 at 08:45, John Taylor wrote:
> > Hi all,
> > I'm (still) working on an application to regularly populate my database with some data provided from
> > a third party DB.
> > I'm still having really bad performance problems.
> > There are 300,000 records to be inserted, but I'm only getting 10,000/hour.
> > 30 hours to populate a single table is just not on. There must be something seriously bad going on.
> From what I see of your INSERT statement, you are joining orderheader
> to orderlines and then inserting those records into orderlines.
> Is that correct?
Yes, unfortunately I get incomplete data, so I need to select some details before inserting.
> I think I'd make a view from the SELECT portion. Then, nightly:
> 1. COPY out the view.
> 2. Drop the index(es) on orderlines
> 3. COPY into orderlines
> 4. recreate indexes on orderlines.
So, how would I merge the data from the view with the data to copy in ?
I was thinking of generating the copy file by performing a select for each row,
but a selecting each row into a file, and then copy in might be more expensive than
just inserting it like I do now.
> Do the same for orderlineupdates. The thing is is that
> modifying indexes is a _very_ expensive operation, no matter
> what RDBMS or hardware configuration you have...
Yes, I was thinking of that for my next step.
Once I figure how to get copy working, I'll also try dropping indexes.
pgsql-novice by date
|Next:||From: Marc Spitzer||Date: 2002-05-22 22:07:43|
|Subject: Re: Better way to bulk-load millions of CSV records into|
|Previous:||From: John Taylor||Date: 2002-05-22 21:05:01|
|Subject: Re: optimising data load|