| From: | "Ian Cass" <ian(dot)cass(at)mblox(dot)com> |
|---|---|
| To: | pgsql-sql(at)postgresql(dot)org |
| Subject: | Quickest way to insert unique records? |
| Date: | 2002-06-26 10:12:49 |
| Message-ID: | 04e601c21cfa$1f853320$6602a8c0@salamander |
| Views: | Whole Thread | Raw Message | Download mbox | Resend email |
| Thread: | |
| Lists: | pgsql-sql |
Hi,
I've got a number of files containing generic log data & some of the lines
may or may not be duplicated across files that I'm feeding into a database
using Perl DBI. I'm just ignoring any duplicate record errors. This is fine
for day to day running when the data feeds in at a sensible rate, however,
if I wanted to feed in a load of old data in a short space of time, this
solution simply is not quick enough.
I can modify the feeder script to generate formated CSV files that I can
then COPY into the database into a temporary table. However, I'll then need
to select each record from the temporary table and insert into the main
table, omitting duplicates.
I guess I'd need something like this....
INSERT INTO messages (host, messageid, body, and, loads, more)
SELECT host, messageid, body, and, loads, more
FROM messages_tmp ;
However, when that hit a duplicate, it would fail wouldn't it?
Also, would this actually be any quicker than direct insertion from Perl
DBI?
--
Ian Cass
| From | Date | Subject | |
|---|---|---|---|
| Next Message | Subhashini Karthikeyan | 2002-06-26 11:47:24 | sequence chages after firing update |
| Previous Message | Christopher Kings-Lynne | 2002-06-26 06:46:32 | Re: what is the difference between default 0 vs default '0' |