-
Notifications
You must be signed in to change notification settings - Fork 90
Trying out additional example files #9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@ajelenak is the person who knows must about this part! |
I believe that many of the possible compression types are indeed supported by zarr, but the script only tries with gzip here. I don't know what szip and lzf compressions are, but they are would presumably need to be implemented in numcodecs to be readable if they are not already under a different name. It would be worth your while finding out which of these cases applies. |
I can confirm that compiling and pointing to the SZIP decompressor is very doable, could be made into a conda package if really needed. The API is not documented, but looks guessable. |
Better to use https://gitlab.dkrz.de/k202009/libaec if you have not already (https://portal.hdfgroup.org/display/HDF5/Szip+Compression+in+HDF+Products). And speaking of conda... https://github.com/conda-forge/libaec-feedstock |
Yes, Fletcher32 is a checksum HDF5 filter. It is used to catch any read errors from HDF5 dataset chunks. When using the Fletcher32 filter, a checksum is calculated on every chunk write operation and stored with the chunk. |
Just wanted to point people here towards an optimized read-only approach to work with the icesat2 HDF5 data described in this issue http://icesat2sliderule.org/h5coro , would be interesting to compare against the fsspec-reference-maker |
This is really neat, and I'm excited to try things out with some additional HDF files!
I realize the goal is to flesh out the specification and this is not a general conversion tool yet, but it seems like working with more HDF files out in the wild might bring things to light.
Some initial questions/suggestions:
lggr.debug(f'_ARRAY_CHUNKS = {h5obj.chunks}')
What isn't supported?
https://github.com/intake/fsspec-reference-maker/blob/bf41138add53b0201e583aa40840cd4fa5fb907b/fsspec_reference_maker/hdf.py#L103-L106
The first file I tried to generate .zchunkstore with ran into the above, code and traceback below:
The text was updated successfully, but these errors were encountered: