I know it's the same to count 1 o 20M rows, but it takes much longer
to count 20M. I really don't know much about the internal of the
driver, but maybe some kind of time out?
On Fri, Feb 22, 2008 at 2:33 PM, Kris Jurka <books(at)ejurka(dot)com> wrote:
> On Fri, 22 Feb 2008, Andres Olarte wrote:
> > Just did a quick test with synthetic data (20M rows 3 columns: serial,
> > timestamp, and text). And it worked ok both on Java (with driver
> > postgresql-8.1-407.jdbc3.jar) and PgAdmin.
> > Perhaps you could try to build a test case to expose the problem?
> From a client perspective, select count(*) should be the same no matter
> how many rows there are and what types of columns are in the table, so a
> test case won't be helpful. If the JVM is just up and dying, that's the
> JVM's fault not a pg issue.
> Kris Jurka
In response to
pgsql-jdbc by date
|Next:||From: Dave Cramer||Date: 2008-02-22 21:13:18|
|Subject: Re: JVM crash when select count(*) on large table through JDBC|
|Previous:||From: Kris Jurka||Date: 2008-02-22 20:33:44|
|Subject: Re: JVM crash when select count(*) on large table through