Re: [DESIGN] ParallelAppend

From: Thom Brown <thom(at)linux(dot)com>
To: Robert Haas <robertmhaas(at)gmail(dot)com>
Cc: Kouhei Kaigai <kaigai(at)ak(dot)jp(dot)nec(dot)com>, "pgsql-hackers(at)postgresql(dot)org" <pgsql-hackers(at)postgresql(dot)org>, Amit Kapila <amit(dot)kapila16(at)gmail(dot)com>, Kyotaro HORIGUCHI <horiguchi(dot)kyotaro(at)lab(dot)ntt(dot)co(dot)jp>
Subject: Re: [DESIGN] ParallelAppend
Date: 2015-11-17 21:59:21
Message-ID: CAA-aLv7=4045zc_s7M77BQUR31k0inYHHAnk0Z2+HSBGwKPKZg@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-hackers

On 17 November 2015 at 20:08, Robert Haas <robertmhaas(at)gmail(dot)com> wrote:
> On Tue, Nov 17, 2015 at 4:26 AM, Thom Brown <thom(at)linux(dot)com> wrote:
>
>> However, the first parallel seq scan shows it getting 170314 rows.
>> Another run shows it getting 194165 rows. The final result is
>> correct, but as you can see from the rows on the Append node (59094295
>> rows), it doesn't match the number of rows on the Gather node
>> (30000000).
>
> Is this the same issue reported in
> http://www.postgresql.org/message-id/CAFj8pRBF-i=qDg9b5nZrXYfChzBEZWmthxYPhidQvwoMOjHtzg@mail.gmail.com
> and not yet fixed? I am inclined to think it probably is.

Yes, that seems to be the same issue.

>> And also, for some reason, I can no longer get this using more than 2
>> workers, even with max_worker_processes = 16 and max_parallel_degree =
>> 12. I don't know if that's anything to do with this patch though.
>
> The number of workers is limited based on the size of the largest
> table involved in the Append. That probably needs considerable
> improvement, of course, but this patch is still a step forward over
> not-this-patch.

Ah, okay. I wasn't aware of this. I'll bear that in mind in future.

Thom

In response to

Responses

Browse pgsql-hackers by date

  From Date Subject
Next Message Andrew Dunstan 2015-11-17 22:51:12 Re: proposal: multiple psql option -c
Previous Message Bert 2015-11-17 21:51:22 Re: Parallel Seq Scan