PostgreSQL Weekly News - February 21, 2021

Posted on 2021-02-22 by PWN

PostgreSQL Weekly News - February 21, 2021

Person of the week:

PostgreSQL Product News

JDBC 42.2.19 released

Pgpool-II 4.2.2, 4.1.6, 4.0.13, 3.7.18 and 3.6.25, a connection pooler and statement replication system for PostgreSQL, released.

pgBadger v11.5, a PostgreSQL log analyzer and graph tool written in Perl, released.

pg_probackup 2.4.10, a utility to manage backup and recovery of PostgreSQL database clusters, released.

pgFormatter 5.0, a formatter/beautifier for SQL code, released.

PostgreSQL Jobs for February

PostgreSQL in the News

Planet PostgreSQL:

PostgreSQL Weekly News is brought to you this week by David Fetter

Submit news and announcements by Sunday at 3:00pm PST8PDT to

Applied Patches

Thomas Munro pushed:

Tom Lane pushed:

  • Minor fixes to improve regex debugging code. When REG_DEBUG is defined, ensure that an un-filled "struct cnfa" is all-zeroes, not just that it has nstates ==
  • This is mainly so that looking at "struct subre" structs in gdb doesn't distract one with a lot of garbage fields during regex compilation. Adjust some places that print debug output to have suitable fflush calls afterwards. In passing, correct an erroneous ancient comment: the concatenation subre-s created by parsebranch() have op == '.' not ','. Noted while fooling around with some regex performance improvements.

  • Simplify loop logic in nodeIncrementalSort.c. The inner loop in switchToPresortedPrefixMode() can be implemented as a conventional integer-counter for() loop, removing a couple of redundant boolean state variables. The old logic here was a remnant of earlier development, but as things now stand there's no reason for extra complexity. Also, annotate the test case added by 82e0e2930 to explain why it manages to hit the corner case fixed in that commit, and add an EXPLAIN to verify that it's creating an incremental-sort plan. Back-patch to v13, like the previous patch. James Coleman and Tom Lane Discussion:

  • Convert tsginidx.c's GIN indexing logic to fully ternary operation. Commit 2f2007fbb did this partially, but there were two remaining warts. checkcondition_gin handled some uncertain cases by setting the out-of-band recheck flag, some by returning TS_MAYBE, and some by doing both. Meanwhile, TS_execute arbitrarily converted a TS_MAYBE result to TS_YES. Thus, if checkcondition_gin chose to only return TS_MAYBE, the outcome would be TS_YES with no recheck flag, potentially resulting in wrong query outputs. The case where this'd happen is if there were GIN_MAYBE entries in the indexscan results passed to gin_tsquery_[tri]consistent, which so far as I can see would only happen if the tidbitmap used to accumulate indexscan results grew large enough to become lossy. I initially thought of fixing this by ensuring we always set the recheck flag as well as returning TS_MAYBE in uncertain cases. But that errs in the other direction, potentially forcing rechecks of rows that provably match the query (since the recheck flag remains set even if TS_execute later finds that the answer must be TS_YES). Instead, let's get rid of the out-of-band recheck flag altogether and rely on returning TS_MAYBE. This requires exporting a version of TS_execute that will actually return the full ternary result of the evaluation ... but we likely should have done that to start with. Unfortunately it doesn't seem practical to add a regression test case that covers this: the amount of data needed to cause the GIN bitmap to become lossy results in a longer runtime than I think we want to have in the tests. (I'm wondering about allowing smaller work_mem settings to ameliorate that, but it'd be a matter for a separate patch.) Per bug #16865 from Dimitri Nüscheler. Back-patch to v13 where the faulty commit came in. Discussion:

  • Make some minor improvements in the regex code. Push some hopefully-uncontroversial bits extracted from an upcoming patch series, to remove non-relevant clutter from the main patches. In compact(), return immediately after setting REG_ASSERT error; continuing the loop would just lead to assertion failure below. (Ask me how I know.) In parseqatom(), remove assertion that moresubs() did its job. When moresubs actually did its job, this is redundant with that function's final assert; but when it failed on OOM, this is an assertion crash. We could avoid the crash by adding a NOERR() check before the assertion, but it seems better to subtract code than add it. (Note that there's a NOERR exit a few lines further down, and nothing else between here and there requires moresubs to have succeeded. So we don't really need an extra error exit.) This is a live bug in assert-enabled builds, but given the very low likelihood of OOM in moresub's tiny allocation, I don't think it's worth back-patching. On the other hand, it seems worthwhile to add an assertion that our intended v->subs[subno] target is still null by the time we are ready to insert into it, since there's a recursion in between. In pg_regexec, ensure we fflush any debug output on the way out, and try to make MDEBUG messages more uniform and helpful. (In particular, ensure that all of them are prefixed with the subre's id number, so one can match up entry and exit reports.) Add some test cases in test_regex to improve coverage of lookahead and lookbehind constraints. Adding these now is mainly to establish that this is indeed the existing behavior. Discussion:

  • Fix another ancient bug in parsing of BRE-mode regular expressions. While poking at the regex code, I happened to notice that the bug squashed in commit afcc8772e had a sibling: next() failed to return a specific value associated with the '}' token for a "{m,n}" quantifier when parsing in basic RE mode. Again, this could result in treating the quantifier as non-greedy, which it never should be in basic mode. For that to happen, the last character before "}" that sets "nextvalue" would have to set it to zero, or it'd have to have accidentally been zero from the start. The failure can be provoked repeatably with, for example, a bound ending in digit "0". Like the previous patch, back-patch all the way.

  • Invent "rainbow" arcs within the regex engine. Some regular expression constructs, most notably the "." match-anything metacharacter, produce a sheaf of parallel NFA arcs covering all possible colors (that is, character equivalence classes). We can make a noticeable improvement in the space and time needed to process large regexes by replacing such cases with a single arc bearing the special color code "RAINBOW". This requires only minor additional complication in places such as pull() and push(). Callers of pg_reg_getoutarcs() must now be prepared for the possibility of seeing a RAINBOW arc. For the one known user, contrib/pg_trgm, that's a net benefit since it cuts the number of arcs to be dealt with, and the handling isn't any different than for other colors that contain too many characters to be dealt with individually. This is part of a patch series that in total reduces the regex engine's runtime by about a factor of four on a large corpus of real-world regexes. Patch by me, reviewed by Joel Jacobson Discussion:

  • Recognize "match-all" NFAs within the regex engine. This builds on the previous "rainbow" patch to detect NFAs that will match any string, though possibly with constraints on the string length. This definition is chosen to match constructs such as ".*", ".+", and ".{1,100}". Recognizing such an NFA after the optimization pass is fairly cheap, since we basically just have to verify that all arcs are RAINBOW arcs and count the number of steps to the end state. (Well, there's a bit of complication with pseudo-color arcs for string boundary conditions, but not much.) Once we have these markings, the regex executor functions longest(), shortest(), and matchuntil() don't have to expend per-character work to determine whether a given substring satisfies such an NFA; they just need to check its length against the bounds. Since some matching problems require O(N) invocations of these functions, we've reduced the runtime for an N-character string from O(N^2) to O(N). Of course, this is no help for non-matchall sub-patterns, but those usually have constraints that allow us to avoid needing O(N) substring checks in the first place. It's precisely the unconstrained "match-all" cases that cause the most headaches. This is part of a patch series that in total reduces the regex engine's runtime by about a factor of four on a large corpus of real-world regexes. Patch by me, reviewed by Joel Jacobson Discussion:

  • Fix regex engine to suppress useless concatenation sub-REs. The comment for parsebranch() claims that it avoids generating unnecessary concatenation nodes in the "subre" tree, but it missed some significant cases. Once we've decided that a given atom is "messy" and can't be bundled with the preceding atom(s) of the current regex branch, parseqatom() always generated two new concat nodes, one to concat the messy atom to what follows it in the branch, and an upper node to concatenate the preceding part of the branch to that one. But one or both of these could be unnecessary, if the messy atom is the first, last, or only one in the branch. Improve the code to suppress such useless concat nodes, along with the no-op child nodes representing empty chunks of a branch. Reducing the number of subre tree nodes offers significant savings not only at execution but during compilation, because each subre node has its own NFA that has to be separately optimized. (Maybe someday we'll figure out how to share the optimization work across multiple tree nodes, but it doesn't look easy.) Eliminating upper tree nodes is especially useful because they tend to have larger NFAs. This is part of a patch series that in total reduces the regex engine's runtime by about a factor of four on a large corpus of real-world regexes. Patch by me, reviewed by Joel Jacobson Discussion:

  • Convert regex engine's subre tree from binary to N-ary style. Instead of having left and right child links in subre structs, have a single child link plus a sibling link. Multiple children of a tree node are now reached by chasing the sibling chain. The beneficiary of this is alternation tree nodes. A regular expression with N (>1) branches is now represented by one alternation node with N children, rather than a tree that includes N alternation nodes as well as N children. While the old representation didn't really cost anything extra at execution time, it was pretty horrid for compilation purposes, because each of the alternation nodes had its own NFA, which we were too stupid not to separately optimize. (To make matters worse, all of those NFAs described the entire alternation pattern, not just the portion of it that one might expect from the tree structure.) We continue to require concatenation nodes to have exactly two children. This data structure is now prepared to support more, but the executor's logic would need some careful redesign, and it's not clear that a lot of benefit could be had. This is part of a patch series that in total reduces the regex engine's runtime by about a factor of four on a large corpus of real-world regexes. Patch by me, reviewed by Joel Jacobson Discussion:

  • Avoid generating extra subre tree nodes for capturing parentheses. Previously, each pair of capturing parentheses gave rise to a separate subre tree node, whose only function was to identify that we ought to capture the match details for this particular sub-expression. In most cases we don't really need that, since we can perfectly well put a "capture this" annotation on the child node that does the real matching work. As with the two preceding commits, the main value of this is to avoid generating and optimizing an NFA for a tree node that's not really pulling its weight. The chosen data representation only allows one capture annotation per subre node. In the legal-per-spec, but seemingly not very useful, case where there are multiple capturing parens around the exact same bit of the regex (i.e. "((xyz))"), wrap the child node in N-1 capture nodes that act the same as before. We could work harder at that but I'll refrain, pending some evidence that such cases are worth troubling over. In passing, improve the comments in regex.h to say what all the different re_info bits mean. Some of them were pretty obvious but others not so much, so reverse-engineer some documentation. This is part of a patch series that in total reduces the regex engine's runtime by about a factor of four on a large corpus of real-world regexes. Patch by me, reviewed by Joel Jacobson Discussion:

Michaël Paquier pushed:

Amit Kapila pushed:

Peter Geoghegan pushed:

  • Adjust lazy_scan_heap() accounting comments. Explain which particular LP_DEAD line pointers get accounted for by the tups_vacuumed variable.

  • Avoid misinterpreting GiST pages in pageinspect. GistPageSetDeleted() sets pd_lower when deleting a page, and sets the page contents to a GISTDeletedPageContents. Avoid treating deleted GiST pages as regular slotted pages within pageinspect. Oversight in commit 756ab291. Author: Andrey Borodin

  • Add "LP_DEAD item?" column to GiST pageinspect functions. This brings gist_page_items() and gist_page_items_bytea() in line with nbtree's bt_page_items() function. Minor follow-up to commit 756ab291, which added the GiST functions. Author: Andrey Borodin Discussion:

  • nbtree README: move VACUUM linear scan section. Discuss VACUUM's linear scan after discussion of tuple deletion by VACUUM, but before discussion of page deletion by VACUUM. This progression is a lot more natural. Also tweak the wording a little. It seems unnecessary to talk about how it worked prior to PostgreSQL 8.2.

  • Add nbtree README section on page recycling. Consolidate discussion of how VACUUM places pages in the FSM for recycling by adding a new section that comes after discussion of page deletion. This structure reflects the fact that page recycling is explicitly decoupled from page deletion in Lanin & Shasha's paper. Page recycling in nbtree is an implementation of what the paper calls "the drain technique". This decoupling is an important concept for nbtree VACUUM. Searchers have to detect and recover from concurrent page deletions, but they will never have to reason about concurrent page recycling. Recycling can almost always be thought of as a low level garbage collection operation that asynchronously frees the physical space that backs a logical tree node. Almost all code need only concern itself with logical tree nodes. (Note that "logical tree node" is not currently a term of art in the nbtree code -- this all works implicitly.) This is preparation for an upcoming patch that teaches nbtree VACUUM to remember the details of pages that it deletes on the fly, in local memory. This enables the same VACUUM operation to consider placing its own deleted pages in the FSM later on, when it reaches the end of btvacuumscan().

Fujii Masao pushed:

  • Display the time when the process started waiting for the lock, in pg_locks, take 2. This commit adds new column "waitstart" into pg_locks view. This column reports the time when the server process started waiting for the lock if the lock is not held. This information is useful, for example, when examining the amount of time to wait on a lock by subtracting "waitstart" in pg_locks from the current time, and identify the lock that the processes are waiting for very long. This feature uses the current time obtained for the deadlock timeout timer as "waitstart" (i.e., the time when this process started waiting for the lock). Since getting the current time newly can cause overhead, we reuse the already-obtained time to avoid that overhead. Note that "waitstart" is updated without holding the lock table's partition lock, to avoid the overhead by additional lock acquisition. This can cause "waitstart" in pg_locks to become NULL for a very short period of time after the wait started even though "granted" is false. This is OK in practice because we can assume that users are likely to look at "waitstart" when waiting for the lock for a long time. The first attempt of this patch (commit 3b733fcd04) caused the buildfarm member "rorqual" (built with --disable-atomics --disable-spinlocks) to report the failure of the regression test. It was reverted by commit 890d2182a2. The cause of this failure was that the atomic variable for "waitstart" in the dummy process entry created at the end of prepare transaction was not initialized. This second attempt fixes that issue. Bump catalog version. Author: Atsushi Torikoshi Reviewed-by: Ian Lawrence Barwick, Robert Haas, Justin Pryzby, Fujii Masao Discussion:

  • Fix "invalid spinlock number: 0" error in pg_stat_wal_receiver. Commit 2c8dd05d6c added the atomic variable writtenUpto into walreceiver's shared memory information. It's initialized only when walreceiver started up but could be read via pg_stat_wal_receiver view anytime, i.e., even before it's initialized. In the server built with --disable-atomics and --disable-spinlocks, this uninitialized atomic variable read could cause "invalid spinlock number: 0" error. This commit changed writtenUpto so that it's initialized at the postmaster startup, to avoid the uninitialized variable read via pg_stat_wal_receiver and fix the error. Also this commit moved the read of writtenUpto after the release of spinlock protecting walreceiver's shared variables. This is necessary to prevent new spinlock from being taken by atomic variable read while holding another spinlock, and to shorten the spinlock duration. This change leads writtenUpto not to be consistent with the other walreceiver's shared variables protected by a spinlock. But this is OK because writtenUpto should not be used for data integrity checks. Back-patch to v13 where commit 2c8dd05d6c introduced the bug. Author: Fujii Masao Reviewed-by: Michael Paquier, Thomas Munro, Andres Freund Discussion:

  • Fix bug in COMMIT AND CHAIN command. This commit fixes COMMIT AND CHAIN command so that it starts new transaction immediately even if savepoints are defined within the transaction to commit. Previously COMMIT AND CHAIN command did not in that case because commit 280a408b48 forgot to make CommitTransactionCommand() handle a transaction chaining when the transaction state was TBLOCK_SUBCOMMIT. Also this commit adds the regression test for COMMIT AND CHAIN command when savepoints are defined. Back-patch to v12 where transaction chaining was added. Reported-by: Arthur Nascimento Author: Fujii Masao Reviewed-by: Arthur Nascimento, Vik Fearing Discussion:

  • Fix psql's ON_ERROR_ROLLBACK so that it handles COMMIT AND CHAIN. When ON_ERROR_ROLLBACK is enabled, psql releases a temporary savepoint if it's idle in a valid transaction block after executing a query. But psql doesn't do that after RELEASE or ROLLBACK is executed because a temporary savepoint has already been destroyed in that case. This commit changes psql's ON_ERROR_ROLLBACK so that it doesn't release a temporary savepoint also when COMMIT AND CHAIN is executed. A temporary savepoint doesn't need to be released in that case because COMMIT AND CHAIN also destroys any savepoints defined within the transaction to commit. Otherwise psql tries to release the savepoint that COMMIT AND CHAIN has already destroyed and cause an error "ERROR: savepoint "pg_psql_temporary_savepoint" does not exist". Back-patch to v12 where transaction chaining was added. Reported-by: Arthur Nascimento Author: Arthur Nascimento Reviewed-by: Fujii Masao, Vik Fearing Discussion:

Heikki Linnakangas pushed:

  • Make ExecGetInsertedCols() and friends more robust and improve comments. If ExecGetInsertedCols(), ExecGetUpdatedCols() or ExecGetExtraUpdatedCols() were called with a ResultRelInfo that's not in the range table and isn't a partition routing target, the functions would dereference a NULL pointer, relinfo->ri_RootResultRelInfo. Such ResultRelInfos are created when firing RI triggers in tables that are not modified directly. None of the current callers of these functions pass such relations, so this isn't a live bug, but let's make them more robust. Also update comment in ResultRelInfo; after commit 6214e2b228, ri_RangeTableIndex is zero for ResultRelInfos created for partition tuple routing. Noted by Coverity. Backpatch down to v11, like commit 6214e2b228. Reviewed-by: Tom Lane, Amit Langote

Andres Freund pushed:

Peter Eisentraut pushed:

Magnus Hagander pushed:

Tomáš Vondra pushed:

Pending Patches

Justin Pryzby sent in a patch to remove some dead code in the REINDEX (CONCURRENTLY, TABLESPACE ..) implementation.

Masahiro Ikeda sent in another revision of a patch to add statistics related to write/sync wal record, and make the wal receiver report WAL statistics.

Michaël Paquier sent in another revision of a patch to refactor the HMAC implementations.

Daniel Gustafsson sent in another revision of a patch to support checksum enable/disable in a running cluster.

Tomáš Vondra sent in another revision of a patch to implement BRIN multi-range indexes.

Peter Eisentraut sent in another revision of a patch to improve new hash partition bound check error messages by enhancing the error message "every hash partition modulus must be a factor of the next larger modulus" with a detail message that shows the particular numbers involved.

Andrey Borodin sent in another revision of a patch to make all SLRU buffer sizes configurable.

Justin Pryzby sent in another revision of a patch to allow CREATE INDEX CONCURRENTLY on partitioned tables.

Takayuki Tsunakawa sent in two more revisions of a patch to speed up COPY FROM into tables with remote partitions in part by adding and using COPY routines to the FDW API.

Peter Eisentraut sent in a patch to set SNI for SSL connections from the client.

Mark Rofail and Joel Jacobson traded patches to implement foreign key arrays.

Bharath Rupireddy sent in two more revisions of a patch to implement ALTER SUBSCRIPTION ... DROP PUBLICATION.

Peter Smith sent in three more revisions of a patch to implement logical decoding of two-phase transactions.

Peter Geoghegan sent in two more revisions of a patch to use full a 64-bit XID for nbtree page deletion, recycle pages deleted during the same VACUUM, and show "pages newly deleted" in VACUUM VERBOSE output.

Amit Langote sent in two more revisions of a patch to ensure that foreign key triggers are create on partitioned tables, and enforce foreign key constraints correctly during cross-partition updates.

Greg Nancarrow and Amit Langote traded patches to make it possible to execute INSERT ... SELECT ... in parallel.

Seamus Abshere and Amit Langote traded patches to allow setting parallel_workers on partitioned tables.

Michaël Paquier sent in a patch to add no_storage, fallback table AM for relations without storage.

Tomáš Vondra and Matthias van de Meent traded patches to extend COPY progress reporting.

Amit Langote sent in another revision of a patch to make updates and deletes on inheritance trees scale better.

Pavel Stěhule sent in another revision of a patch to implement schema variables.

Pavel Stěhule sent in another revision of a patch to make it possible to read pg_dump options from a file.

Paul Martinez sent in a patch to document the effect of max_replication_slots on the subscriber side.

Amit Langote sent in three more revisions of a patch to fix tuple routing to initialize batching only for inserts.

Andy Fan sent in two more revisions of a patch to Introduce notnullattrs field in RelOptInfo to indicate which attrs are not null in current query.

Yugo Nagata sent in another revision of a patch to implement incremental maintenance of materialized views.

David Rowley sent in two more revisions of a patch to add TID Range Scans to support efficient scanning ranges of TIDs.

Kyotaro HORIGUCHI sent in two revisions of a patch to intended to fix a bug that manifested as an error in some invocations of ALTER TABLE by correcting the set of relation types that are actually permitted.

Justin Pryzby and Michaël Paquier traded patches to implement progress reporting for CREATE INDEX on partitioned tables.

Amit Kapila sent in three more revisions of a patch to distinguishing between physical and logical replication connections when there is an authentication failure.

John Naylor sent in four more revisions of a patch to validate UTF-8 using SIMD instructions.

Vik Fearing sent in three revisions of a patch to implement the SQL standard TRIM_ARRAY, which removes the last elements from an array.

David Rowley and Andy Fan traded patches to Allow estimate_num_groups() to pass back further details about the estimation, allow users of simplehash.h to perform direct deletions, add a Result Cache executor node, and remove some code duplication in nodeResultCache.c.

Justin Pryzby and Dilip Kumar traded patches to add custom table compression methods.

Takashi Menjo sent in another revision of a patch to make it possible to use persistent memory for WAL operations.

Mark Dilger sent in another revision of a patch to add a heapcheck contrib extension.

Thomas Munro sent in two more revisions of a patch to fix a bug that manifested as pg_collation_actual_version() ERROR: cache lookup failed for collation 123 by fixing the way collation providers are handled.

Kyotaro HORIGUCHI sent in another revision of a patch to make it possible to specify a CRL directory in addition to a CRL file.

Daniel Gustafsson sent in another revision of a patch to make it possible to use NSS as a TLS backend for libpq.

Álvaro Herrera sent in another revision of a patch to implement batch/pipelining support in libpq.

Bharath Rupireddy sent in another revision of a patch to add table AMs for multi- and single inserts, then use same for CTAS, REFRESH MATERIALIZED VIEW, and COPY.

Tomáš Vondra sent in another revision of a patch to implement extended statistics on expressions.

Hou Zhijie sent in another revision of a patch to add a new GUC, enable_parallel_dml (boolean default false), and a new table option, parallel_dml_enabled (boolean default true). Each do what they say.

Thomas Munro sent in a patch to add sort_template.h for making fast sort functions.

Amit Langote sent in a patch to allow batching of inserts during cross-partition updates.

Daniel Gustafsson sent in a patch to disallow SSL compression on the grounds that it makes cryptanalytic attacks easier to mount.

Peter Eisentraut sent in another revision of a patch to Simplify printing of LSNs by adding a LSN_FORMAT_ARGS macro for use in printf-style printing of LSNs.

Konstantin Knizhnik sent in another revision of a patch to fix an infelicity between stored procedures and accesses to TOAST data.

Jan Wieck sent in a patch to add a wire protocol hook and a sample implementation, telnet.

Guillaume Lelarge sent in another revision of a patch to add an --extension (-e) option to pg_dump, which specifies an extension to be dumped. It can be used 0 or more times.

Nathan Bossart sent in another revision of a patch to avoid creating archive status ".ready" files too early, and keep track of notified-ready-for-archive position through crashes.

Nathan Bossart sent in another revision of a patch to clarify the documentation of the behavior of RESET ROLE when "role" is set per-user, per-database, or via command-line options, and add configuration parameter entries for "role" and "session_authorization".

Paul Guo sent in another revision of a patch to speed up pg_rewind by fsync()ing only the affected files/directories, and using copy_file_range() for file copying.

Michaël Paquier sent in another revision of a patch to implement the missing locking functions for OpenSSL <= 1.0.2.

Iwata Aya sent in another revision of a patch to add tracing to libpq.

Andres Freund sent in another revision of a patch to implement wal prohibit state using global barriers, Error or Assert before START_CRIT_SECTION for WAL write, and document the effects of these for transam and page via READMEs.

Georgios Kokolatos sent in another revision of a patch to make dbsize a bit more consistent.

Denis Smirnov sent in two more revisions of a patch to make the analyze AM more flexible by allowing for variable-sized blocks.

Paul Guo sent in two revisions of a patch to freeze tuples during CTAS, which can avert unneeded vacuums.

Andy Fan sent in two more revisions of a patch to make some static functions extern and extend ChangeVarNodes so it can change var->attno, and build some implied pruning quals to extend the usecase of planning time partition pruning and init partition pruning.

Marcus Wanner sent in two revisions of a patch to present empty prepares to the output plugin.

Ajin Cherian and Marcus Wanner traded patches to preserve the ordering of PREPAREs vs COMMITs in logical decoding.

Andrew Dunstan sent in another revision of a patch to help track down cfbot failures by having the make files and the msvc build system create a well-known file with the location(s) to search for log files if there's an error.

Andy Fan sent in a patch to introduce notnullattrs field in RelOptInfo to indicate which attrs are not null in current query, and add UniqueKey with EquivalenceClass for single rel only.

Thomas Munro sent in a patch per Andrew Gierth to remove an outdated reference to spindles in the description of effective_io_concurrency.

Bharath Rupireddy sent in a patch to update multiple progress params at once.

Vik Fearing sent in two revisions of a patch to implement GROUP BY DISTINCT per the standard.

Vik Fearing sent in a patch to add catalog version to pg_config along with a new catalog_version guc, accessible from SQL.