Hi Sasha,
RNNs are definitely useful and something we are looking at investing
in.
Yes, computationally efficient confidence intervals is an area I
have done research in. The bootstrap is great because it's a
general purpose tool, but can still be a pain computationally
becuase no matter what, you have to train many models, even if its
only on a subsample of the data (for example, as in Bag of Little
Bootstraps (BLB)). The efficient confidence interval (CI) methods I
work on are based on asymptotic theory (influence function/curves)
rather than resampling theory -- they require no retraining of
models, which is why they are the most computationally efficient.
The only drawback I see in influence-curve based methods is that a
different algo/implementation must be created for each estimator
that you are trying to create a CI for.
I worked a lot with imbalanced binary response data problems (where
I used AUC to evaluate models) in my applied research, so I was
particularly interested in generating CIs for cross-validated AUC
estimates. I have implemented that in the cvAUC R package
(
https://github.com/ledell/cvAUC) and yes, I do plan to integrate
this work into H2O in the future (for AUC and other model
performance metrics). You can read more about this in the last
chapter of my dissertation, if you are interested:
http://www.stat.berkeley.edu/~ledell/papers/ledell-phd-thesis.pdf,
or in the article here:
http://projecteuclid.org/euclid.ejs/1437742107
-Erin