Strings in HDF5¶
Note
The rules around reading & writing string data were redesigned for h5py 3.0. Refer to the h5py 2.10 docs for how to store strings in older versions.
Reading strings¶
String data in HDF5 datasets is read as bytes by default: bytes
objects
for variable-length strings, or numpy bytes arrays ('S'
dtypes) for
fixed-length strings. Use Dataset.asstr()
to retrieve str
objects.
Variable-length strings in attributes are read as str
objects. These are
decoded as UTF-8 with surrogate escaping for unrecognised bytes.
Storing strings¶
When creating a new dataset or attribute, Python str
or bytes
objects
will be treated as variable-length strings, marked as UTF-8 and ASCII respectively.
Numpy bytes arrays ('S'
dtypes) make fixed-length strings.
You can use string_dtype()
to explicitly specify any HDF5 string datatype.
When writing data to an existing dataset or attribute, data passed as bytes is
written without checking the encoding. Data passed as Python str
objects
is encoded as either ASCII or UTF-8, based on the HDF5 datatype.
In either case, null bytes ('\x00'
) in the data will cause an error.
Warning
Fixed-length string datasets will silently truncate longer strings which are written to them. Numpy byte string arrays do the same thing.
Fixed-length strings in HDF5 hold a set number of bytes. It may take multiple bytes to store one character.
What about NumPy’s U
type?¶
NumPy also has a Unicode type, a UTF-32 fixed-width format (4-byte characters). HDF5 has no support for wide characters. Rather than trying to hack around this and “pretend” to support it, h5py will raise an error if you try to store data of this type.
How to store raw binary data¶
If you have a non-text blob in a Python byte string (as opposed to ASCII or
UTF-8 encoded text, which is fine), you should wrap it in a void
type for
storage. This will map to the HDF5 OPAQUE datatype, and will prevent your
blob from getting mangled by the string machinery.
Here’s an example of how to store binary data in an attribute, and then recover it:
>>> binary_blob = b"Hello\x00Hello\x00"
>>> dset.attrs["attribute_name"] = np.void(binary_blob)
>>> out = dset.attrs["attribute_name"]
>>> binary_blob = out.tobytes()
Object names¶
Unicode strings are used exclusively for object names in the file:
>>> f.name
'/'
You can supply either byte or unicode strings when creating or retrieving objects. If a byte string is supplied, it will be used as-is; Unicode strings will be encoded as UTF-8.
In the file, h5py uses the most-compatible representation; H5T_CSET_ASCII for characters in the ASCII range; H5T_CSET_UTF8 otherwise.
>>> grp = f.create_dataset(b"name")
>>> grp2 = f.create_dataset("name2")
Encodings¶
HDF5 supports two string encodings: ASCII and UTF-8.
We recommend using UTF-8 when creating HDF5 files, and this is what h5py does
by default with Python str
objects.
If you need to write ASCII for compatibility reasons, you should ensure you only
write pure ASCII characters (this can be done by
your_string.encode("ascii")
), as otherwise your text may turn into
mojibake.
You can use string_dtype()
to specify the encoding for string data.
See also
- Joel Spolsky’s introduction to Unicode & character sets
If this section looks like gibberish, try this.
For reading, as long as the encoding metadata is correct, the defaults for
Dataset.asstr()
will always work.
However, HDF5 does not enforce the string encoding, and there are files where
the encoding metadata doesn’t match what’s really stored.
Most commonly, data marked as ASCII may be in one of the many “Extended ASCII”
encodings such as Latin-1. If you know what encoding your data is in,
you can specify this using Dataset.asstr()
. If you have data
in an unknown encoding, you can also use any of the builtin python error
handlers.
Variable-length strings in attributes are read as str
objects, decoded as
UTF-8 with the 'surrogateescape'
error handler. If an attribute is
incorrectly encoded, you’ll see ‘surrogate’ characters such as '\udcb1'
when reading it:
>>> s = "2.0±0.1"
>>> f.attrs["string_good"] = s # Good - h5py uses UTF-8
>>> f.attrs["string_bad"] = s.encode("latin-1") # Bad!
>>> f.attrs["string_bad"]
'2.0\udcb10.1'
To recover the original string, you’ll need to encode it with UTF-8, and then decode it with the correct encoding:
>>> f.attrs["string_bad"].encode('utf-8', 'surrogateescape').decode('latin-1')
'2.0±0.1'
Fixed length strings are different; h5py doesn’t try to decode them:
>>> s = "2.0±0.1"
>>> utf8_type = h5py.string_dtype('utf-8', 30)
>>> ascii_type = h5py.string_dtype('ascii', 30)
>>> f.attrs["fixed_good"] = np.array(s.encode("utf-8"), dtype=utf8_type)
>>> f.attrs["fixed_bad"] = np.array(s.encode("latin-1"), dtype=ascii_type)
>>> f.attrs["fixed_bad"]
b'2.0\xb10.1'
>>> f.attrs["fixed_bad"].decode("utf-8")
Traceback (most recent call last):
File "<input>", line 1, in <module>
f.attrs["fixed_bad"].decode("utf-8")
UnicodeDecodeError: 'utf-8' codec can't decode byte 0xb1 in position 3: invalid start byte
>>> f.attrs["fixed_bad"].decode("latin-1")
'2.0±0.1'
As we get bytes back, we only need to decode them with the correct encoding.