Re: Reducing the overhead of NUMERIC data
От | Tom Lane |
---|---|
Тема | Re: Reducing the overhead of NUMERIC data |
Дата | |
Msg-id | 8698.1131070663@sss.pgh.pa.us обсуждение исходный текст |
Ответ на | Re: Reducing the overhead of NUMERIC data (Gregory Maxwell <gmaxwell@gmail.com>) |
Ответы |
Re: Reducing the overhead of NUMERIC data
|
Список | pgsql-hackers |
Gregory Maxwell <gmaxwell@gmail.com> writes: > Another way to look at this is in the context of compression: With > unicode, characters are really 32bit values... But only a small range > of these values is common. So we store and work with them in a > compressed format, UTF-8. > As such it might be more interesting to ask some other questions like: > are we using the best compression algorithm for the application, and, > why do we sometimes stack two compression algorithms? Actually, the real reason we use UTF-8 and not any of the sorta-fixed-size representations of Unicode is that the backend is by and large an ASCII, null-terminated-string engine. *All* of the supported backend encodings are ASCII-superset codes. Making everything null-safe in order to allow use of UCS2 or UCS4 would be a huge amount of work, and the benefit is at best questionable. regards, tom lane
В списке pgsql-hackers по дате отправления: