> Now the question is: is this an efficient way of doing things? In
> particular, is the frequent write access to the dataset buffered, and
> do
> I need to keep the number of resize events small? Or do I need to
> manually buffer the simulation output in an in-memory numpy.array and
> write it in chunks myself?
To be honest, I'm not sure. I think you will have to benchmark this
to be certain of the result. My impression from working with HDF5 is
that you want to keep the number of resize events small, but I'm not
sure what effect this has on performance. It may be that it doesn't
matter.
Another approach would be to create a large array and then trim it
when you're done. However, you would want to benchmark this to make
sure HDF5 doesn't end up wasting file space, which can happen
sometimes.
I am at a conference this week but will get to your other email this
weekend. I think there are some ways to make your tree traversal
faster.
Andrew
> Now the question is: is this an efficient way of doing things? InTo be honest, I'm not sure. I think you will have to benchmark this
> particular, is the frequent write access to the dataset buffered, and
> do
> I need to keep the number of resize events small? Or do I need to
> manually buffer the simulation output in an in-memory numpy.array and
> write it in chunks myself?
to be certain of the result. My impression from working with HDF5 is
that you want to keep the number of resize events small, but I'm not
sure what effect this has on performance. It may be that it doesn't
matter.