Connection pooling - Number of connections

From: Brett Wooldridge <brett(dot)wooldridge(at)gmail(dot)com>
To: pgsql-performance(at)postgresql(dot)org
Subject: Connection pooling - Number of connections
Date: 2014-03-24 13:27:10
Message-ID: CAPg6oRMd6YzgKv9ZBdK806Rux4EjJ4u046F1j=96NWU+urmYEw@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

Hi, Brett Wooldridge here, one of the principals of HikariCP. I thought
I'd wade into the conversation pool a little myself if you guys don't mind.

Speaking to David's point...
>> Reaching the maxPoolSize from the minPoolSize means creating the
>> connections at the crucial moment where the client application is in the
>> desperate need of completing an important query/transaction which the
>> primary responsibility since it cannot hold the data collected.

This was one of the reasons I was proposing the fixed pool design. In my
experience, even in pools that maintain a minimum number of idle
connections, responding to spike demands is problematic. If you have a
pool with say 30 max. connections, and a 10 minimum idle connection goal, a
sudden spike demand for 20 connections means the pool can satisfy 10
instantly but then is left to [try to] establish 10 connections before the
application's connectionTimeout (read acquisition timeout from the pool) is
reached. This in turn generates a spike demand on the database slowing
down not only the connection establishments themselves but also slowing
down the completion of transactions that might actually return connections
to the pool.

As I think Tom noted is a slidestack I read somewhere, there is a "knee" in
the performance curve beyond which additional connections cause a drop in
TPS. While users often think it is a good idea to have maxPoolSize of 100,
the reality is they can retire/reuse connections faster with a much smaller
pool. I didn't see a pool of a 2 or 3 dozen connections actually impacting
performance much when half of them are idle and half are executing
transactions (ie. the idle ones don't impact the overall performance much).

Finally, one of my contentions was, either your database server has
resources or it doesn't. Either it has enough memory and processing power
for N connections or it doesn't. If the pool is set below, near, or at
that capacity what is the purpose of releasing connections in that case?
Yes, it frees up memory, but that memory is not really available for other
use given that at any instant the maximum capacity of the pool may be
demanded.
Instead releasing resources only to try to reallocate them during a demand
peak seems counter-productive.

I'd appreciate any shared thoughts on my presuppositions.

-Brett

Responses

Browse pgsql-performance by date

  From Date Subject
Next Message Matthew Spilich 2014-03-24 14:37:57 semaphore waits and performance stall
Previous Message Heikki Linnakangas 2014-03-24 07:08:46 Re: Suboptimal query plan when using expensive BCRYPT functions