Re: Bulkloading using COPY - ignore duplicates?

From: Daniel Kalchev <daniel(at)digsys(dot)bg>
To: Bruce Momjian <pgman(at)candle(dot)pha(dot)pa(dot)us>
Cc: "Mikheev, Vadim" <vmikheev(at)sectorbase(dot)com>, Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>, Lee Kindness <lkindness(at)csl(dot)co(dot)uk>, Peter Eisentraut <peter_e(at)gmx(dot)net>, Jim Buttafuoco <jim(at)buttafuoco(dot)net>, PostgreSQL Development <pgsql-hackers(at)postgresql(dot)org>
Subject: Re: Bulkloading using COPY - ignore duplicates?
Date: 2002-01-04 07:36:01
Message-ID: 200201040736.JAA29349@dcave.digsys.bg
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-hackers

>>>Bruce Momjian said:
> Mikheev, Vadim wrote:
> > > > Bruce Momjian <pgman(at)candle(dot)pha(dot)pa(dot)us> writes:
> > > > > Seems nested transactions are not required if we load
> > > > > each COPY line in its own transaction, like we do with
> > > > > INSERT from pg_dump.
> > > >
> > > > I don't think that's an acceptable answer. Consider
> > >
> > > Oh, very good point. "Requires nested transactions" added to TODO.
> >
> > Also add performance issue with per-line-commit...
> >
> > Also-II - there is more common name for required feature - savepoints.
>
> OK, updated TODO to prefer savepoints term.

Now, how about the same functionality for

INSERT into table1 SELECT * from table2 ... WITH ERRORS;

Should allow the insert to complete, even if table1 has unique indexes and we
try to insert duplicate rows. Might save LOTS of time in bulkloading scripts
not having to do single INSERTs.

Guess all this will be available in 7.3?

Daniel

Responses

Browse pgsql-hackers by date

  From Date Subject
Next Message Vadim Mikheev 2002-01-04 07:47:36 Re: Bulkloading using COPY - ignore duplicates?
Previous Message Tom Lane 2002-01-04 07:13:01 Re: freebsd/alpha probs