Re: performance while importing a very large data set in to database

From: Scott Marlowe <scott(dot)marlowe(at)gmail(dot)com>
To: Ashish Kumar Singh <ashishkumar(dot)singh(at)altair(dot)com>
Cc: pgsql-performance(at)postgresql(dot)org
Subject: Re: performance while importing a very large data set in to database
Date: 2009-12-05 20:42:17
Message-ID: dcc563d10912051242r328bd125xc2eae3a20fbe8a46@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

On Wed, Dec 2, 2009 at 4:31 PM, Ashish Kumar Singh
<ashishkumar(dot)singh(at)altair(dot)com> wrote:
> Hello Everyone,
>
> I have a very bit big database around 15 million in size, and the dump file
> is around 12 GB.
>
> While importing this dump in to database I have noticed that initially query
> response time is very slow but it does improves with time.
>
> Any suggestions to improve performance after dump in imported in to database
> will be highly appreciated!

This is pretty normal. When the db first starts up or right after a
load it has nothing in its buffers or the kernel cache. As you access
more and more data the db and OS learned what is most commonly
accessed and start holding onto those data and throw the less used
stuff away to make room for it. Our production dbs run at a load
factor of about 4 to 6, but when first started and put in the loop
they'll hit 25 or 30 and have slow queries for a minute or so.

Having a fast IO subsystem will help offset some of this, and
sometimes "select * from bigtable" might too.

In response to

Responses

Browse pgsql-performance by date

  From Date Subject
Next Message Scott Marlowe 2009-12-05 20:45:07 Re: query cost too high, anyway to reduce it
Previous Message Scott Marlowe 2009-12-05 20:39:11 Re: Large DB, several tuning questions: Index sizes, VACUUM, REINDEX, Autovacuum