Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/wbolster/mysql-latin1-codec
Python string codec for MySQL's latin1 encoding
https://github.com/wbolster/mysql-latin1-codec
Last synced: about 2 months ago
JSON representation
Python string codec for MySQL's latin1 encoding
- Host: GitHub
- URL: https://github.com/wbolster/mysql-latin1-codec
- Owner: wbolster
- License: other
- Created: 2013-12-06T12:05:06.000Z (about 11 years ago)
- Default Branch: master
- Last Pushed: 2014-01-31T20:28:12.000Z (almost 11 years ago)
- Last Synced: 2024-10-30T00:26:57.252Z (about 2 months ago)
- Language: Python
- Homepage:
- Size: 168 KB
- Stars: 6
- Watchers: 3
- Forks: 4
- Open Issues: 1
-
Metadata Files:
- Readme: README.rst
- License: LICENSE.rst
Awesome Lists containing this project
README
=================================================
Python string codec for MySQL's *latin1* encoding
=================================================:Author: Wouter Bolsterlee
:License: 3-clause BSD
:URL: https://github.com/wbolster/mysql-latin1-codecOverview
========This project provides a Python string codec for MySQL's *latin1* encoding, and
an accompanying *iconv*-like command line script for use in shell pipes.Rationale
=========* MySQL defaults to using the *latin1* encoding for all its textual data, but
its *latin1* encoding is not actually *latin1* but a MySQL-specific variant.* Due to improperly written applications or wrongly configured databases, many
existing databases keep data in MySQL *latin1* columns, even if that data is
not actually *latin1* data.* MySQL will *not* complain about this, so this often goes unnoticed. The
problems only appear when you try to access the database from another
application, or when you try to ``grep`` through database dumps produced by
``mysqldump``.* Many libraries do not support this encoding properly, and using the real
*latin1* encoding leads to corruption when processing data, especially when
the database contains text that is not in a West-European language.* This string codec makes it possible to reconstruct the exact bytes as they
were stored in MySQL. This is a valuable tool for fixing text encoding issues
with such databases. For convenience, a command line interface that behaves
like *iconv* is also included.Version history
===============* Version 2.0 (2014-01-31)
* Added explicit ``register()`` function instead of having side-effects upon
module import.* Version 1.0 (2013-12-04)
* Initial releaseInstallation
============::
$ pip install mysql-latin1-codec
The package supports both Python 2 and Python 3.
Usage
=====You can use this project in two ways: as a stand-alone command line tool and as
a Python module.Command line tool
-----------------The command line tool behaves like *iconv*::
$ python -m mysql_latin1_codec --help
usage: mysql_latin1_codec.py [-h] [-f encoding] [-t encoding] [-o filename]
[-c]
[inputs [inputs ...]]iconv-like tool to encode or decode data using MySQL's "latin1" dialect,
which, despite the name, is a mix of cp1252 and ISO-8859-1.positional arguments:
inputs Input file(s) (defaults to stdin)optional arguments:
-h, --help show this help message and exit
-f encoding, --from encoding
Source encoding (uses MySQL's "latin1" if omitted)
-t encoding, --to encoding
Target encoding (uses MySQL's "latin1" if omitted)
-o filename, --output filename
Output file (defaults to stdout)
-c, --skip-invalid Omit invalid characters from output (not the default)Python API
----------In Python code, simply import the module named ``mysql_latin1_codec`` and call
the ``register()`` function. A string codec named ``mysql_latin1`` will be
registered in Python's codec registry::import mysql_latin1_codec
mysql_latin1_codec.register()
You can use it using the normal ``.decode()`` and ``.encode()`` methods on
(byte)strings, and you can also specify it as the ``encoding`` argument to
various I/O functions like ``io.open()``. Example::# String encoding/decoding round-trip
s1 = u'foobar'
s2 == text.encode('mysql_latin1').decode('mysql_latin1')
assert s1 == s2# Reading files
import io
with io.open('/path/to/file', 'r', encoding='mysql_latin1') as fp:
for line in fp:
passPractical examples
==================The example below ‘fixes’ dumps that contain doubly encoded UTF-8 data, i.e.
real UTF-8 data stored in a MySQL *latin1* table. By default ``mysqldump`` makes
UTF-8 dumps, but if MySQL thinks the data is *latin1*, it will convert it again,
resulting in double encoded data. ::$ cat backup-of-broken-database-produced-by-mysqldump.sql \
| python -m mysql_latin1_codec -f UTF-8 \
| iconv -c -f UTF-8 -t UTF-8 \
> legible-text-in-utf8.sqlThe *iconv* pipe in this example removes invalid UTF-8 sequences, while keeping
the valid parts as-is. MySQL truncates values whose size exceeds the column's
maximum size, but if MySQL doesn't know that it is handling UTF-8 data (because
the database schema and the broken application did not tell it to do so) it
truncates the byte sequence, not the character sequence. This may result in
incomplete UTF-8 sequences when a multi-byte sequence is truncated somewhere in
the middle. Since those characters cannot be recovered anyway, removing them is
the right solution in this case.In code you can do something similar to the example above::
original = b'...' # byte string containing doubly-encoded UTF-8 data
s = original.decode('UTF-8').encode('mysql_latin1').decode('UTF-8', 'replace')Another example to ‘fix’ a dump that contains GB2312 (Simplified Chinese) data
stored in a MySQL *latin1* column, again misinterpreted and encoded to UTF-8 by
``mysqldump``::$ cat mojibake-crap.sql \
| python -m mysql_latin1_codec -f UTF-8 \
| iconv -f GB2312 -t UTF-8 \
> legible-text-in-utf8.sqlTechnical background
====================How MySQL defines *latin1*
--------------------------The character set that MySQL uses when *latin1* is specified, is not actually
the well-known *latin1* character set, officially known as ISO-8859-1. What
MySQL calls *latin1* is actually a custom encoding based on *cp-1252* (also
known as *windows-1252*).The MySQL documentation on `West European Character Sets 9§ 10.1.14.2)
`_ contains:``latin1`` is the default character set. MySQL's ``latin1`` is the same as
the Windows ``cp1252`` character set. THis means it is the same as official
``ISO 8859-1`` or IANA (Internet Assigned Numbers Authority) ``latin1``,
except that IANA ``latin1`` treats the code points between ``0x80`` and
``0x9f`` as “undefined”, whereas ``cp1252``, and therefore MySQL's
``latin``, assign characters for those positions. For example, ``0x80`` is
the Euro sign. For the “undefined” entries in ``cp1252``, MySQL translates
``0x81`` to Unicode ``0x0081``, ``0x8d`` to ``0x008d``, ``0x8ff`` to
``0x008f``, ``0x90`` to ``0x0090``, and ``0x9d`` to ``0x009d``.Some more details can be found in the MySQL source code in the file
``strings/ctype-latin1.c``::WL#1494 notes:
We'll use cp1252 instead of iso-8859-1.
cp1252 contains printable characters in the range 0x80-0x9F.
In ISO 8859-1, these code points have no associated printable
characters. Therefore, by converting from CP1252 to ISO 8859-1,
one would lose the euro (for instance). Since most people are
unaware of the difference, and since we don't really want a
"Windows ANSI" to differ from a "Unix ANSI", we will:- continue to pretend the latin1 character set is ISO 8859-1
- actually allow the storage of euro etc. so it's actually cp1252Also we'll map these five undefined cp1252 character:
0x81, 0x8D, 0x8F, 0x90, 0x9D
into corresponding control characters:
U+0081, U+008D, U+008F, U+0090, U+009D.
like ISO-8859-1 does. Otherwise, loading "mysqldump"
output doesn't reproduce these undefined characters.As you can see, this encoding is significantly different from ISO-8859-1 (the
real *latin1*), but MySQL misleadingly labels it as *latin* anyway.Why this can be a problem
-------------------------MySQL's *latin1* encoding allows for arbitrary data to be stored in database
columns, without any validation. This means *latin1* text columns can store any
byte sequence, for example UTF-8 encoded text (which uses a variable number of
bytes per character) or even JPEG images (which is not text at all).This is of course not the proper use of *latin1* columns. Even in this modern
Unicode-aware world, in which all properly written software that handles text
should use UTF-8 (or another Unicode encoding), it is quite common to stumble
upon wrongly configured databases or badly written software. Most applications
use the same (incorrect) assumptions for both storing and retrieving data, so in
many setups this will still ‘just work’, and the problem can go unnoticed for a
long time.What makes this problem worse, is that MySQL defaults to using the *latin1*
character encoding, mostly for historical and backward-compatibility reasons.
This means many databases in the real world are (perhaps mistakingly) configured
to store data in columns that use MySQL's *latin1* encoding, even though the
actual data stores in those columns is not encoding using *latin1* at all.This can lead to a variety of problems, such as encoding or decoding errors,
double encoded text, malfunctioning string operations, or incorrect truncation
which can lead to data corruption. In many cases this manifests itself as
`mojibake `_ text. This may be caused by
a misinterpretation of the characters that the bytes represent, or by double
encodings, e.g. UTF-8 in a *latin1* column that was converted to UTF-8 again by
a backup script.Many tools, like Python's built-in text codecs and the *iconv* (both the command
line tool and the C library) cannot convert data encoding using this custom
MySQL encoding. This makes it quite hard to ‘recover’ e.g. UTF-8 data that was
stored in a *latin1* column, and subsequently dumped using *mysqldump*, even if
you know what you're doing and which actual encoding was used.When invoked on the command line, this script converts the dump file(s)
specified on the command line (or standard input if no files were given). The
data is interpreted as UTF-8 and encoded as MySQL's *latin1* and written to the
standard output. The output is the raw data, which likely needs further
processing, e.g. using iconv to "reinterpret" the data correctly (e.g. as
UTF-8).I have no idea what you are talking about!
==========================================No worries, that's okay.