From: | Joe <svn(at)freedomcircle(dot)net> |
---|---|
To: | Tom Lane <tgl(at)sss(dot)pgh(dot)pa(dot)us> |
Cc: | Dawid Kuroczko <qnex42(at)gmail(dot)com>, pgsql-general <pgsql-general(at)postgresql(dot)org> |
Subject: | Re: Converting MySQL tinyint to PostgreSQL |
Date: | 2005-07-12 21:37:32 |
Message-ID: | 42D4381C.6090208@freedomcircle.net |
Views: | Raw Message | Whole Thread | Download mbox | Resend email |
Thread: | |
Lists: | pgsql-general |
Tom Lane wrote:
> Because the length specification is in *characters*, which is not by any
> means the same as *bytes*.
>
> We could possibly put enough intelligence into the low-level tuple
> manipulation routines to count characters in whatever encoding we happen
> to be using, but it's a lot faster and more robust to insist on a count
> word for every variable-width field.
I guess what you're saying is that PostgreSQL stores characters in
varying-length encodings. If it stored character data in Unicode (UCS-16) it
would always take up two-bytes per character. Have you considered supporting
NCHAR/NVARCHAR, aka NATIONAL character data? Wouldn't UCS-16 be needed to
support multi-locale clusters (as someone as inquiring about recently)?
Joe
From | Date | Subject | |
---|---|---|---|
Next Message | Tom Lane | 2005-07-12 21:45:04 | Re: Checkpoints are occurring too frequently... |
Previous Message | Joshua D. Drake | 2005-07-12 21:36:35 | Re: Checkpoints are occurring too frequently... |