Take a look at:
I am not sure if RedHat provides different default values than what
PostgreSQL normally comes with.
I would also run some typical test queries after changing some of these
values, especially work_mem to make sure your not impacting performance
I'm not sure what else is running on the system, or what type of usage
you are expecting, but 2GB of ram might turn out to be enough for
Also take a look at this:
I found it to be useful, although I believe some of the suggestions the
author makes are assuming that you have a large PostgreSQL installation.
He has an interesting note about random_page_cost which might be one of
the numbers you might want to change.
Mag Gam wrote:
> Hello All,
> Running 8.3.4. My situation is a little unique. I am running on a 1
> core with 2GB of memory on Redhat Linux 5.2. My entire installation of
> pgsql is about 8GB (compressed) from pgdump. I have 6 databases The
> data is keep growing since I plan to add more field to my database and
> it will increase dramatically.
> My goal is I don't want to use a lot of memory! My storage is faily
> fast, I can do about 250Mb/sec (sustained). I would like to leverage
> my I/O instead of memory, eventhough I will suffer performance
> problems. Also, is it possible to make the database data logs files
> (the binary files) large? I am thinking of making them the size of 1G
> each instead of very small files? My FS does better performance for
> larger files...
> Any ideas?
In response to
pgsql-novice by date
|Next:||From: Brad Nicholson||Date: 2008-12-12 20:43:15|
|Subject: Re: tuning question|
|Previous:||From: Bruce Hyatt||Date: 2008-12-12 03:09:29|
|Subject: Re: Connecting via postgresql jdbc|