== PostgreSQL Weekly News - December 04 2016 ==
== PostgreSQL Product News ==
PostgreSQL Automatic Failover (PAF) v2.1_beta1 released.
Pyrseas 0.7.3, a toolkit for PostgreSQL version control, released.
PostGIS 2.3.1, the industry standard geographic information
system package for PostgreSQL, released.
PGroonga 1.1.9, a full text search platform for all languages, released.
== PostgreSQL Jobs for December ==
== PostgreSQL Local ==
CHAR(16) will take place in New York, December 6, 2016.
PGDay.IT 2016 will take place in Prato on December the 13th 2016.
PGConf India 2017 will be on March 2-3, 2017 in Bengaluru, Karnataka.
Proposals are due via <papers AT pgconf DOT in> by 31st December, 2016.
PostgreSQL(at)SCaLE will take place on March 2-3, 2017, at Pasadena Convention
Center, as part of SCaLE 15X.
== PostgreSQL in the News ==
Planet PostgreSQL: http://planet.postgresql.org/
PostgreSQL Weekly News is brought to you this week by David Fetter
Submit news and announcements by Sunday at 3:00pm Pacific time.
Please send English language ones to david(at)fetter(dot)org, German language
to pwn(at)pgug(dot)de, Italian language to pwn(at)itpug(dot)org(dot)
== Applied Patches ==
Tom Lane pushed:
- Code review for early drop of orphaned temp relations in autovacuum. Commit
a734fd5d1 exposed some race conditions that existed previously in the autovac
code, but were basically harmless because autovac would not try to delete
orphaned relations immediately. Specifically, the test for orphaned-ness was
made on a pg_class tuple that might be dead by now, allowing autovac to try to
remove a table that the owning backend had just finished deleting. This
resulted in a hard crash due to inadequate caution about accessing the table's
catalog entries without any lock. We must take a relation lock and then
recheck whether the table is still present and still looks deletable before we
do anything. Also, it seemed to me that deleting multiple tables per
transaction, and trying to continue after errors, represented unjustifiable
complexity. We do not expect this code path to be taken often in the field,
nor even during testing, which means that prioritizing performance over
correctness is a bad tradeoff. Rip all that out in favor of just starting a
new transaction after each successful temp table deletion. If we're unlucky
enough to get an error, which shouldn't happen anyway now that we're being
more cautious, let the autovacuum worker fail as it normally would. In
passing, improve the order of operations in the initial scan loop. Now that
we don't care about whether a temp table is a wraparound hazard, there's no
need to perform extract_autovac_opts, get_pgstat_tabentry_relid, or
relation_needs_vacanalyze for temp tables. Also, if GetTempNamespaceBackendId
returns InvalidBackendId (indicating it doesn't recognize the schema as temp),
treat that as meaning it's NOT an orphaned temp table, not that it IS one,
which is what happened before because BackendIdGetProc necessarily failed.
The case really shouldn't come up for a table that has RELPERSISTENCE_TEMP,
but the consequences if it did seem undesirable. (This might represent a
back-patchable bug fix; not sure if it's worth the trouble.) Discussion:
- Fix busted tab-completion pattern for ALTER TABLE t ALTER c DROP ...
Evidently a thinko in commit 9b181b036. Kyotaro Horiguchi
- Fix estimate_expression_value to constant-fold SQLValueFunction nodes.
Oversight in my commit 0bb51aa96. Noted while poking at a recent bug report
--- HEAD's estimates for a query using CURRENT_DATE were unexpectedly much
worse than 9.6's.
- Fix incorrect variable type in set_rel_consider_parallel(). func_parallel()
returns char not Oid. Harmless, but still wrong. Amit Langote
- Add uuid to the set of types supported by contrib/btree_gist. Paul Jungwirth,
reviewed and hacked on by Teodor Sigaev, Ildus Kurbangaliev, Adam Brusselback,
Chris Bandy, and myself. Discussion:
- Test all contrib-created operator classes with amvalidate. I'd supposed that
people would do this manually when creating new operator classes, but the
folly of that was exposed today. The tests seem fast enough that we can just
apply them during the normal regression tests. contrib/isn fails the checks
for lack of complete sets of cross-type operators. That's a nice-to-have
policy rather than a functional requirement, so leave it as-is, but insert
ORDER BY in the query to ensure consistent cross-platform output. Discussion:
- Improve eqjoinsel_semi's behavior for small inner relations with no stats. If
we don't have any MCV statistics for the inner relation, and we don't trust
its numdistinct estimate either, eqjoinsel_semi falls back to a very
conservative estimate (that 50% of the outer rows have matches). This is
particularly problematic if the inner relation is completely empty, since then
even an explicit ANALYZE won't produce any pg_statistic entries, so there's no
way to budge the planner off the bad estimate. We'd produce a better estimate
in such cases if we used the nd2/nd1 selectivity heuristic, so an easy fix is
to treat the nd2 estimate as non-default if we derive it from clamping to the
inner rel's rowcount estimate. This won't fix every related case (mainly
because the rowcount estimate might be larger than DEFAULT_NUM_DISTINCT), but
it seems like a sane extension of the existing logic, so let's apply the
change in HEAD and see if anyone complains. Per bug #14438 from Nikolay
- Fix bogus handling of JOIN_UNIQUE_OUTER/INNER cases for parallel joins.
consider_parallel_nestloop passed the wrong jointype down to its subroutines
for JOIN_UNIQUE_INNER cases (it should pass JOIN_INNER), and it thought that
it could pass paths other than innerrel->cheapest_total_path to
create_unique_path, which create_unique_path is not on board with. These bugs
would lead to assertion failures or other errors, suggesting that this code
path hasn't been tested much. hash_inner_and_outer's code for parallel join
effectively treated both JOIN_UNIQUE_OUTER and JOIN_UNIQUE_INNER the same as
JOIN_INNER (for different reasons :-(), leading to incorrect plans that
treated a semijoin as if it were a plain join. Michael Day submitted a test
case demonstrating that hash_inner_and_outer failed for JOIN_UNIQUE_OUTER, and
I found the other cases through code review. Report:
- Doc: improve description of trim() and related functions. Per bug #14441 from
Mark Pether, the documentation could be misread, mainly because some of the
examples failed to show what happens with a multicharacter "characters to
trim" string. Also, while the text description in most of these entries was
fairly clear that the "characters" argument is a set of characters not a
substring to match, some of them used variant wording that was a bit less
clear. trim() itself suffered from both deficiencies and was thus pretty
misinterpretable. Also fix failure to explain which of LEADING/TRAILING/BOTH
is the default. Discussion:
- Delete deleteWhatDependsOn() in favor of more performDeletion() flag bits.
deleteWhatDependsOn() had grown an uncomfortably large number of assumptions
about what it's used for. There are actually only two minor differences
between what it does and what a regular performDeletion() call can do, so
let's invent additional bits in performDeletion's existing flags argument that
specify those behaviors, and get rid of deleteWhatDependsOn() as such. (We'd
probably have done it this way from the start, except that performDeletion
didn't originally have a flags argument, IIRC.) Also, add a SKIP_EXTENSIONS
flag bit that prevents ever recursing to an extension, and use that when
dropping temporary objects at session end. This provides a more general
solution to the problem addressed in a hacky way in commit 08dd23cec: if an
extension script creates temp objects and forgets to remove them again, the
whole extension went away when its contained temp objects were deleted. The
previous solution only covered temp relations, but this solves it for all
object types. These changes require minor additions in dependency.c to pass
the flags to subroutines that previously didn't get them, but it's still a net
savings of code, and it seems cleaner than before. Having done this, revert
the special-case code added in 08dd23cec that prevented addition of pg_depend
records for temp table extension membership, because that caused its own
oddities: dropping an extension that had created such a table didn't
automatically remove the table, leading to a failure if the table had another
dependency on the extension (such as use of an extension data type), or to a
duplicate-name failure if you then tried to recreate the extension. But we
keep the part that prevents the pg_temp_nnn schema from becoming an extension
member; we never want that to happen. Add a regression test case covering
these behaviors. Although this fixes some arguable bugs, we've heard few
field complaints, and any such problems are easily worked around by explicitly
dropping temp objects at the end of extension scripts (which seems like good
practice anyway). So I won't risk a back-patch. Discussion:
- Fix broken wait-for-previous-process-to-exit loop in regression test. Must do
pg_stat_clear_snapshot() inside test's loop, or our snapshot of
pg_stat_activity will never change :-(. Thinko in b3427dade -- evidently my
workstation never really iterated the loop in testing. Per buildfarm.
- Don't mess up pstate->p_next_resno in transformOnConflictClause().
transformOnConflictClause incremented p_next_resno while generating the phony
targetlist for the EXCLUDED pseudo-rel. Then that field got incremented some
more during transformTargetList, possibly leading to free_parsestate
concluding that we'd overrun the allowed length of a tlist, as reported by
Justin Pryzby. We could fix this by resetting p_next_resno to 1 after using
it for the EXCLUDED pseudo-rel tlist, but it seems easier and less coupled to
other places if we just don't use that field at all in this loop. (Note that
this doesn't change anything about the resnos that end up appearing in the
main target list, because those are all replaced with target-column numbers by
updateTargetListEntry.) In passing, fix incorrect type OID assigned to the
whole-row Var for "EXCLUDED.*" (somehow this escaped having any bad
consequences so far, but it's certainly wrong); remove useless assignment to
var->location; pstrdup the column names in case of a relcache flush; and
improve nearby comments. Back-patch to 9.5 where ON CONFLICT was introduced.
Álvaro Herrera pushed:
- Fix get_relation_info name typo'ed in a comment. Plus add a missing comment
about this in get_relation_info itself. Author: Amit Langote. Discussion:
- Permit dump/reload of not-too-large >1GB tuples. Our documentation states
that our maximum field size is 1 GB, and that our maximum row size of 1.6 TB.
However, while this might be attainable in theory with enough contortions, it
is not workable in practice; for starters, pg_dump fails to dump tables
containing rows larger than 1 GB, even if individual columns are well below
the limit; and even if one does manage to manufacture a dump file containing a
row that large, the server refuses to load it anyway. This commit enables
dumping and reloading of such tuples, provided two conditions are met: 1. no
single column is larger than 1 GB (in output size -- for bytea this includes
the formatting overhead) 2. the whole row is not larger than 2 GB There are
three related changes to enable this: a. StringInfo's API now has two
additional functions that allow creating a string that grows beyond the
typical 1GB limit (and "long" string). ABI compatibility is maintained. We
still limit these strings to 2 GB, though, for reasons explained below. b.
COPY now uses long StringInfos, so that pg_dump doesn't choke trying to emit
rows longer than 1GB. c. heap_form_tuple now uses the MCXT_ALLOW_HUGE flag in
its allocation for the input tuple, which means that large tuples are accepted
on input. Note that at this point we do not apply any further limit to the
input tuple size. The main reason to limit to 2 GB is that the FE/BE protocol
uses 32 bit length words to describe each row; and because the documentation
is ambiguous on its signedness and libpq does consider it signed, we cannot
use the highest-order bit. Additionally, the StringInfo API uses "int" (which
is 4 bytes wide in most platforms) in many places, so we'd need to change that
API too in order to improve, which has lots of fallout. Backpatch to 9.5,
which is the oldest that has MemoryContextAllocExtended, a necessary piece of
infrastructure. We could apply to 9.4 with very minimal additional effort,
but any further than that would require backpatching "huge" allocations too.
This is the largest set of changes we could find that can be back-patched
without breaking compatibility with existing systems. Fixing a bigger set of
problems (for example, dumping tuples bigger than 2GB, or dumping fields
bigger than 1GB) would require changing the FE/BE protocol and/or changing the
StringInfo API in an ABI-incompatible way, neither of which would be
back-patchable. Authors: Daniel Vérité, Álvaro Herrera. Reviewed by: Tomas
- Fix Windows build for 78c8c814390f. Author: Petr Jelínek
- Fix outdated comments Commit 597a87ccc9a6f neglected to update some comments;
fix. Report and patch by Thomas Munro. Reviewed by Petr Jelínek.
Stephen Frost pushed:
- Correct psql documentation example. An example in the psql documentation had
an incorrect field name from what the command actually produced. Pointed out
by Fabien COELHO. Back-patch to 9.6 where the example was added. Discussion:
- Clarify pg_dump -b documentation. The documentation around the -b/--blobs
option to pg_dump seemed to imply that it might be possible to add blobs to a
"schema-only" dump or similar. Clarify that blobs are data and therefore will
only be included in dumps where data is being included, even when -b is used
to request blobs be included. The -b option has been around since before 9.2,
so back-patch to all supported branches. Discussion:
- Add --no-blobs option to pg_dump. Add an option to exclude blobs when running
pg_dump. By default, blobs are included but this option can be used to
exclude them while keeping the rest of the dump. Commment updates and
regression tests from me. Author: Guillaume Lelarge. Reviewed-by: Amul Sul.
Robert Haas pushed:
- libpq: Add target_session_attrs parameter. Commit
274bb2b3857cc987cfa21d14775cae9b0dababa5 made it possible to specify multiple
IPs in a connection string, but that's not good enough for the case where you
have a read-write master and a bunch of read-only standbys and want to connect
to whichever server is the master at the current time. This commit allows
that, by making it possible to specify target_session_attrs=read-write as a
connection parameter. There was extensive discussion of the best name for the
connection parameter and its values as well as the best way to distinguish
master and standbys. For now, adopt the same solution as JDBC: if the user
wants a read-write connection, issue 'show transaction_read_only' and
rejection the connection if the result is 'on'. In the future, we could add
additional values of this new target_session_attrs parameter that issue
different queries; or we might have some way of distinguishing the server type
without resorting to an SQL query; but right now, we have this, and that's
(hopefully) a good start. Victor Wagner and Mithun Cy. Design review by
Álvaro Herrera, Catalin Iacob, Takayuki Tsunakawa, and Craig Ringer; code
review by me. I changed Mithun's patch to skip all remaining IPs for a host
if we reject a connection based on this new parameter, rewrote the
documentation, and did some other cosmetic cleanup. Discussion:
- Improve hash index bucket split behavior. Previously, the right to split a
bucket was represented by a heavyweight lock on the page number of the primary
bucket page. Unfortunately, this meant that every scan needed to take a
heavyweight lock on that bucket also, which was bad for concurrency. Instead,
use a cleanup lock on the primary bucket page to indicate the right to begin a
split, so that scans only need to retain a pin on that page, which is they
would have to acquire anyway, and which is also much cheaper. In addition to
reducing the locking cost, this also avoids locking out scans and inserts for
the entire lifetime of the split: while the new bucket is being populated with
copies of the appropriate tuples from the old bucket, scans and inserts can
happen in parallel. There are minor concurrency improvements for vacuum
operations as well, though the situation there is still far from ideal. This
patch also removes the unworldly assumption that a split will never be
interrupted. With the new code, a split is done in a series of small steps
and the system can pick up where it left off if it is interrupted prior to
completion. While this patch does not itself add write-ahead logging for hash
indexes, it is clearly a necessary first step, since one of the things that
could interrupt a split is the removal of electrical power from the machine
performing it. Amit Kapila. I wrote the original design on which this patch
is based, and did a good bit of work on the comments and README through
multiple rounds of review, but all of the code is Amit's. Also reviewed by
Jesper Pedersen, Jeff Janes, and others. Discussion:
- libpq: Fix inadvertent change in PQhost() behavior. Commit
274bb2b3857cc987cfa21d14775cae9b0dababa5 caused PQhost() to return the value
of the hostaddr parameter rather than the relevant host when the latter
parameter was specified. That's wrong. Commit
9a1d0af4ad2cbd419115b453d811c141b80d872b then amplified the damage by using
PQhost() in more places, so that the SSL test suite started failing. Report
by Andreas Karlsson; patch by me.
- Add max_parallel_workers GUC. Increase the default value of the existing
max_worker_processes GUC from 8 to 16, and add a new max_parallel_workers GUC
with a maximum of 8. This way, even if the maximum amount of parallel query
is happening, there is still room for background workers that do other things,
as originally envisioned when max_worker_processes was added. Julien Rouhaud,
reviewed by Amit Kapila and by revised by me.
- Clarify that pg_stat_activity.query has a length limit. There was always
documentation of the GUC that controlled what the limit actually was, but
previously the documentation of the field itself made no mention of that
limit. Ian Barwick
- Add a crude facility for dealing with relative pointers. C doesn't have any
sort of built-in understanding of a pointer relative to some arbitrary base
address, but dynamic shared memory segments can be mapped at different
addresses in different processes, so any sort of shared data structure stored
within a dynamic shared memory segment can't use absolute pointers. We could
use something like Size to represent a relative pointer, but then the compiler
provides no type-checking. Use stupid macro tricks to get some type-checking.
Patch originally by me. Concept suggested by Andres Freund. Recently
resubmitted as part of Thomas Munro's work on dynamic shared memory
allocation. Discussion: 20131205144434(dot)GG12398(at)alap2(dot)anarazel(dot)de Discussion:
- Management of free memory pages. This is intended as infrastructure for a
full-fledged allocator for dynamic shared memory. The interface looks a bit
like a real allocator, but only supports allocating and freeing memory in
multiples of the 4kB page size. Further, to free memory, you must know the
size of the span you wish to free, in pages. While these are make it
unsuitable as an allocator in and of itself, it still serves as very useful
scaffolding for a full-fledged allocator. Robert Haas and Thomas Munro. This
code is mostly the same as my 2014 submission, but Thomas fixed quite a few
bugs and made some changes to the interface. Discussion:
- Introduce dynamic shared memory areas. Programmers discovered decades ago
that it was useful to have a simple interface for allocating and freeing
memory, which is why malloc() and free() were invented. Unfortunately, those
handy tools don't work with dynamic shared memory segments because those are
specific to PostgreSQL and are not necessarily mapped at the same address in
every cooperating process. So invent our own allocator instead. This makes
it possible for processes cooperating as part of parallel query execution to
allocate and free chunks of memory without having to reserve them prior to the
start of execution. It could also be used for longer lived objects; for
example, we could consider storing data for pg_stat_statements or the stats
collector in shared memory using these interfaces, rather than writing them to
files. Basically, anything that needs shared memory but can't predict in
advance how much it's going to need might find this useful. Thomas Munro and
Robert Haas. The original code (of mine) on which Thomas based his work was
actually designed to be a new backend-local memory allocator for PostgreSQL,
but that hasn't gone anywhere - or not yet, anyway. Thomas took that work and
performed major refactoring and extensive modifications to make it work with
dynamic shared memory, including the addition of appropriate locking.
- Fix thinko in b3427dade14cc31eb48740bc9ea98b5954470b24.
Peter Eisentraut pushed:
- Straighten out some whitespace
- doc: Fix typo. From: Tsunakawa, Takayuki <tsunakawa(dot)takay(at)jp(dot)fujitsu(dot)com>
- doc: Remove claim about large shared_buffers on Windows. Testing has shown
that it is no longer correct. From: Tsunakawa, Takayuki
<tsunakawa(dot)takay(at)jp(dot)fujitsu(dot)com> Reviewed-by: amul sul <sulamul(at)gmail(dot)com>
- Use grammar symbol function_with_argtypes consistently. Instead of sometimes
referring to a function signature like func_name func_args, use the existing
function_with_argtypes symbol, which combines the two. Reviewed-by: Alvaro
Herrera <alvherre(at)2ndquadrant(dot)com> Reviewed-by: Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>
- Move function_with_argtypes to a better location. It was apparently added for
use by GRANT/REVOKE, but move it closer to where other function signature
related things are kept. Reviewed-by: Alvaro Herrera
<alvherre(at)2ndquadrant(dot)com> Reviewed-by: Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us>
- Add aggregate_with_argtypes and use it consistently. This works like
function_with_argtypes, but aggregates allow slightly different arguments.
Reviewed-by: Alvaro Herrera <alvherre(at)2ndquadrant(dot)com> Reviewed-by: Tom Lane
- Use latch instead of select() in walreceiver. Replace use of poll()/select()
by WaitLatchOrSocket(), which is more portable and flexible. Also change
walreceiver to use its procLatch instead of a custom latch. From: Petr
- Refactor libpqwalreceiver. The whole walreceiver API is now wrapped into a
struct, like most of our other loadable module APIs. The libpq connection is
no longer a global variable in libpqwalreceiver. Instead, it is encapsulated
into a struct that is passed around the functions. This allows multiple
walreceivers to run at the same time. Add some rudimentary support for
logical replication connections to libpqwalreceiver. These changes are mostly
cosmetic and are going to be useful for the future logical replication
patches. From: Petr Jelinek <petr(at)2ndquadrant(dot)com>
Heikki Linnakangas pushed:
- Remove dead stuff from pgcrypto. pgp-pubkey-DISABLED test has been unused
since 2006, when support for built-in bignum math was added (commit 1abf76e8).
pgp-encrypt-DISABLED has been unused forever, AFAICS. Also remove a couple of
unused error codes.
- Remove code points < 0x80 from character conversion tables. PostgreSQL treats
characters with < 0x80 leading byte as plain ASCII, and they are not even
passed to the conversion routines. There is no point in having them in the
conversion tables. Everything in the tables were direct ASCII-ASCII mappings,
except for two: * SHIFT_JIS_2004 code point 0x5C (backslash in ASCII) was
mapped to Unicode YEN SIGN character. * Unicode 0x5C (backslash again) was
mapped to "REVERSE SOLIDUS" in SHIFT_JIS_2004 These mappings never had any
effect, so there's no functional change from removing them. Discussion:
- Rewrite the perl scripts to produce our Unicode conversion tables. Generate
EUC_CN mappings from gb-18030-2000.xml, because GB2312.TXT is no longer
available. Get UHC from windows-949-2000.xml, it's more up-to-date. Plus
tons more small changes. With these changes, the perl scripts faithfully
produce the *.map files we have in the repository, from the external source
files. In the passing, fix the Makefile to also download CP932.TXT and
CP950.TXT. Based on patches by Kyotaro Horiguchi, reviewed by Daniel
- Remove leading zeros, for consistency with other map files. The common style
is to pad to 4 digits. Running the current perl scripts to generate these map
files would override this change, but the next commit will rewrite the perl
scripts to produce this style. I'm doing this as a separate commit, to make it
more clear what non-cosmetic changes the next commit makes to the map files.
- Make all unicode perl scripts to use strict, rearrange logic for clarity. The
loops were a bit difficult to understand, due to breaking out of them early.
Also fix things that perlcritic complained about. Daniel Gustafsson
Andres Freund pushed:
- Perform one only projection to compute agg arguments. Previously we did a
ExecProject() for each individual aggregate argument. That turned out to be a
performance bottleneck in queries with multiple aggregates. Doing all the
argument computations in one ExecProject() is quite a bit cheaper because
ExecProject's fastpath can do the work at once in a relatively tight loop, and
because it can get all the required columns with a single slot_getsomeattr and
save some other redundant setup costs. Author: Andres Freund. Reviewed-By:
Heikki Linnakangas. Discussion:
- User narrower representative tuples in the hash-agg hashtable. So far the
hashtable stored representative tuples in the form of its input slot, with all
columns in the hashtable that are not needed (i.e. not grouped upon or
functionally dependent) set to NULL. Thats good for saving memory, but it
turns out that having tuples full of NULL isn't free. slot_deform_tuple is
faster if there's no NULL bitmap even if no NULLs are encountered, and
skipping over leading NULLs isn't free. So compute a separate tuple
descriptor that only contains the needed columns. As columns have already been
moved in/out the slot for the hashtable that does not imply additional per-row
overhead. Author: Andres Freund. Reviewed-By: Heikki Linnakangas.
Michael Meskes pushed:
- Added missing "EXEC SQL" to statement.
Noah Misch pushed:
- Remove wrong CloseHandle() call. In accordance with its own documentation,
invoke CloseHandle() only when directed in the documentation for the function
that furnished the handle. GetModuleHandle() does not so direct. We have
been issuing this call only in the rare event that a CRT DLL contains no
"_putenv" symbol, so lack of bug reports is uninformative. Back-patch to 9.2
(all supported versions). Christian Ullrich, reviewed by Michael Paquier.
- Refine win32env.c cosmetics. Replace use of plain 0 as a null pointer
constant. In comments, update terminology and lessen redundancy. Back-patch
to 9.2 (all supported versions) for the convenience of back-patching the next
two commits. Christian Ullrich and Noah Misch, reviewed (in earlier versions)
by Michael Paquier.
- Make pgwin32_putenv() visit debug CRTs. This has no effect in the most
conventional case, where no relevant DLL uses a debug build. For an example
where it does matter, given a debug build of MIT Kerberos, the
krb_server_keyfile parameter usually had no effect. Since nobody wants a
Heisenbug, back-patch to 9.2 (all supported versions). Christian Ullrich,
reviewed by Michael Paquier.
- Make pgwin32_putenv() follow DLL loading and unloading. Until now, the first
putenv() call of a given postgres.exe process would cache the set of loaded
CRTs. If a CRT unloaded after that call, the next putenv() would crash. That
risk was largely theoretical, because the first putenv() precedes all
PostgreSQL-initiated module loading. However, this might explain bad
interactions with antivirus and other software that injects threads
asynchronously. If an additional CRT loaded after the first putenv(),
pgwin32_putenv() would not discover it. That CRT would have all environment
changes predating its load, but it would not receive later
PostgreSQL-initiated changes. An additional CRT loading concurrently with the
first putenv() might miss that change in addition to missing later changes.
Fix all those problems. This removes the cache mechanism from
pgwin32_putenv(); the cost, less than 100 μs per backend startup, is
negligible. No resulting misbehavior was known to be user-visible given the
core distribution alone, but one can readily construct an affected extension
module. No back-patch given the lack of complaints and the potential for
behavior changes in non-PostgreSQL code running in the backend. Christian
Ullrich, reviewed by Michael Paquier.
- Make pgwin32_putenv() probe every known CRT, regardless of compiler. This
extends to MinGW builds the provision for MSVC-built libraries to see putenv()
effects. Doing so repairs, for example, the handling of the
krb_server_keyfile parameter when linked with MSVC-built MIT Kerberos. Like
the previous commit, no back-patch.
- Document recipe for testing compatibility with old Perl. Craig Ringer,
reviewed by Kyotaro HORIGUCHI and Michael Paquier.
== Pending Patches ==
Pavel Stěhule sent in another revision of a patch to add session server side
Kyotaro HORIGUCHI sent in another revision of a patch to refactor psql's tab
completion code for more sanity and add some new tab completions based on this.
Etsuro Fujita sent in another revision of a patch to push down more full joins
in the postgres_fdw.
Matheus de Oliveira sent in another revision of a patch to add GRANT/REVOKE ON
SCHEMAS to ALTER DEFAULT PRIVILEGES.
Masahiko Sawada sent in another revision of a patch to add quorum commit for
multiple synchronous replicas.
Julian Markwort sent in another revision of a patch to add a PGPASSFILE
connection option to libpq.
Mithun Cy sent in another revision of a patch to add pg_autoprewarm.
Andrew Borodin sent in two revisions of a patch to make a more
concurrency-friendly vacuum of GIN indexes.
Tom Lane sent in a patch to fix the order of tests in findDependentObjects().
Dilip Kumar sent in two more revisions of a patch to push scan keys down to the
Petr Jelínek and Peter Eisentraut traded patches to add logical replication.
Haribabu Kommi sent in another revision of a patch to add a 64-bit (EUI-64) mac
Stephen Frost sent in a patch to fix some compiler warnings.
Emre Hasegeli sent in another revision of a patch to add the customer
contains/is contained by operators for the inet data type.
Dilip Kumar sent in another revision of a patch to add parallel bitmap heap
Peter Eisentraut sent in a patch to move collation import to the backend.
Amit Langote sent in a doc patch to clarify a point in ALTER TABLE.
Simon Riggs sent in a patch to make EvalPlanQual() follow the LockTuple()
Heikki Linnakangas and Michaël Paquier traded patches to add secure random
Pavel Stěhule and Álvaro Herrera traded patches to add xmltable().
Etsuro Fujita sent in another revision of a patch to ensure that altering a
foreign table invalidats prepared statement execution plans in which it is
Amit Langote sent in another revision of a patch to implement declarative
Amit Langote sent in two revisions of a patch to clarify the behavior of ALTER
TABLE with parallelism enabled.
Peter Eisentraut sent in another revision of a patch to implement a pg_sequence
Tomas Vondra sent in another revision of a patch to add two slab-like memory
Ian Barwick sent in a doc patch to clarify the length of query text displayed in
Stephen Frost sent in another revision of a patch to add support for restrictive
Ashutosh Bapat sent in another revision of a patch to implement partition-wise
join for join between (declaratively) partitioned tables.
Thomas Munro sent in another revision of a patch to create a DSA area to provide
work space for parallel execution.
Peter Eisentraut sent in another revision of a patch intended to support making
DROP FUNCTION able to drop multiple functions per statement.
Mithun Cy sent in a patch to fix a bug in failover at the libpq connect level.
Michaël Paquier sent in another revision of a patch to fix checkpoint skip logic
on idle systems by tracking LSN progress.
Amit Kapila sent in two more revisions of a patch to fix an infelicity between
parallel execution and prepared statements.
Andreas Karlsson sent in another revision of a patch to reload SSL certificates
Pavan Deolasee sent in another revision of a patch to implement WARM.
Thomas Munro sent in a patch to provide the right format string for dsa_pointer
to printf-like functions.
Amit Kapila sent in a patch to fix a failed assertion in _hash_splitbucket_guts.
Fabien COELHO sent in another revision of a patch to add backslash-return (well
newline really) continuations to all pgbench backslash-commands.
Peter Geoghegan sent in another revision of a patch to add parallel tuplesort.
pgsql-announce by date
|Next:||From: D'Arcy Cain||Date: 2016-12-11 11:31:05|
|Subject: PyGreSQL 5.0.3 released|
|Previous:||From: Jehan-Guillaume de Rorthais||Date: 2016-12-04 14:42:59|
|Subject: PostgreSQL Automatic Failover (PAF) v2.1 beta1|