IOError Traceback (most recent call last)
<ipython-input-45-509cebb66565> in <module>()
1 print geno["matrix"].shape
2 print geno["matrix"].dtype
----> 3 geno["matrix"][0]
/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_hl/dataset.pyc in __getitem__(self, args)
443 mspace = h5s.create_simple(mshape)
444 fspace = selection._id
--> 445 self.id.read(mspace, fspace, arr, mtype)
446
447 # Patch up the output for NumPy
/home/sarah/anaconda/lib/python2.7/site-packages/h5py/h5d.so in h5py.h5d.DatasetID.read (h5py/h5d.c:2782)()
/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_proxy.so in h5py._proxy.dset_rw (h5py/_proxy.c:1709)()
/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_proxy.so in h5py._proxy.H5PY_H5Dread (h5py/_proxy.c:1379)()
IOError: Can't read data (Can't open directory)
I've not seen this error before, but from what I gather from Google
it's an error related to the HDF5 plug-in mechanism. The dataset you
have presumably uses a third-party "filter" which isn't available.
Your best bet is to ask the person you got the file from for an
"un-filtered" version, or for a copy of the filter library.
Hmmm... gzip should work everywhere, since it's distributed with HDF5 itself.
Could I ask you to (1) let me know your anaconda version ("conda -V"),
and (2) run the attached script, and post the output?
Enter codeSummary of the h5py configuration
---------------------------------
h5py 2.3.1
HDF5 1.8.13
Python 2.7.8 |Anaconda 2.1.0 (32-bit)| (default, Aug 21 2014, 18:22:40)
[GCC 4.4.7 20120313 (Red Hat 4.4.7-1)]
sys.platform linux2
sys.maxsize 2147483647
numpy 1.9.0
encode 1 decode 2
shuffle
True
3
f32
True
3
szip
False
gzip
False
so
True
3
lzf
True
3
Traceback (most recent call last):
File "t.py", line 25, in <module>
dset = f.create_dataset('x', (10,), compression='gzip')
File "/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_hl/group.py", line 94, in create_dataset
dsid = dataset.make_new_dset(self, shape, dtype, data, **kwds)
File "/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_hl/dataset.py", line 99, in make_new_dset
shuffle, fletcher32, maxshape, scaleoffset)
File "/home/sarah/anaconda/lib/python2.7/site-packages/h5py/_hl/filters.py", line 103, in generate_dcpl
raise ValueError('Compression filter "%s" is unavailable' % compression)
ValueError: Compression filter "gzip" is unavailable here...--
You received this message because you are subscribed to the Google Groups "h5py" group.
To unsubscribe from this group and stop receiving emails from it, send an email to h5py+uns...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
There was definitely a problem with the anaconda hdf5 recipe. Zlib is required for gzip support. The pull request was approved, and I expect you should now be able to do a "conda update hdf5". You don't have to wait for the next release of anaconda, they push updates for individual packages continually.
--
Summary of the h5py configuration
---------------------------------
h5py 2.3.1
HDF5 1.8.12
Python 2.7.6 (default, Nov 18 2013, 15:12:51)
[GCC 4.2.1 Compatible Apple LLVM 5.0 (clang-500.2.79)]
sys.platform darwin
sys.maxsize 9223372036854775807
numpy 1.8.0
encode 1 decode 2
shuffle True 3
f32 True 3
szip False
Traceback (most recent call last):
File "testh5py.py", line 19, in <module>
print name, h5z.filter_avail(filter), h5z.get_filter_info(filter)
File "h5z.pyx", line 91, in h5py.h5z.get_filter_info (h5py/h5z.c:855)